GithubHelp home page GithubHelp logo

hannesstark / protein-localization Goto Github PK

View Code? Open in Web Editor NEW
53.0 3.0 9.0 12.63 MB

Using Transformer protein embeddings with a linear attention mechanism to make SOTA de-novo predictions for the subcellular location of proteins :microscope:

License: MIT License

Python 25.70% Jupyter Notebook 74.30%
bio-embeddings pytorch protein-localization machine-learning attention-mechanism attention

protein-localization's Introduction

๐Ÿ‘‹ Feel free to reach out to me about any project!

I am happy to chat about research or anything else. Find my email and other ways to reach me on my website!

protein-localization's People

Contributors

hannesstark avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar

protein-localization's Issues

a minor question about the input dimensions in LA module

Hello HannesStark,
LightAttention is a really well-designed module for solving signal-peptides classitfition task.
Recently, inspired by your design, I am test it on predicting targeting peptide target-efficiency task. However, the input of LA module is
self.feature_convolution = nn.Conv1d(embeddings_dim, embeddings_dim, kernel_size, stride=1,
padding=kernel_size // 2)
where the input channel of conv1d is the same dimension as the embeddings_dim. I have read through your paper but not found a clearly explaination on this. I guess you may use masked dimensions or something else to make sure the input length is consistent. But here is the problem, my own dataset contains fixed-size peptides, which is 12 amino acids long. Shall I make it masked as u did or just define the first embeddings_dim which is conv1d(input_channel) as 12?

And I would greatly appreciate it if you could offer any insights or suggestions on my prediction tasks which turned classification task(the location prediction which this work focused on) into regression tasks( where my dataset is a bunch of scalar value).

thanks!

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.