GithubHelp home page GithubHelp logo

sunnydreamrain / indrnn_pytorch Goto Github PK

View Code? Open in Web Editor NEW
131.0 131.0 30.0 3.13 MB

Independently Recurrent Neural Networks (IndRNN) implemented in pytorch.

Python 95.77% Shell 4.23%
action indrnn language-modeling rnn skeleton

indrnn_pytorch's People

Contributors

anirudh257 avatar sunnydreamrain avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar

indrnn_pytorch's Issues

Not able to prepare the dataset shape correctly.

Hi ... Thanks alot for this fabulous work. I'm still trying to understand the code and and trying to run this in my machine.
The main issue I'm having is, I'm not able to prepare the dataset as I required.

I'm trying to understand the following code

datasets=train_datasets
dataname=datasets+'.npy'
labelname=datasets+'_label.npy'
lenname=datasets+'_len.npy'
data_handle=np.load(dataname)
label_handle=np.load(labelname)
len_handle=np.load(lenname)

What is this train_datasets should be? What's it shape need to be? Can you elaborate more on this. Because I need to arrange it to feed to the network

cuda version of IndRNNCell?

Hi, is there any chance to implement a cuda version of IndRNNCell? the purpose is to speed up processing variable length sequences.

issue on grad

Hi, thanks for your great work! But I have encountered something wrong while reproducing the code.
During "train", the grad of the most layers is "None". Except layers "classify_weight"、"classify_bias"、"RNN5_weight"、"RNN5_bias" have not-None grad, others have grad which are None. As a result, error happens when running to "grad_climp", as showed in the following figure.
Maybe something goes wrong with layer "RNN5_weight_hh" during loss.backward() I think.
lossNone_issue

I wonder how to address this problem. Looking forward to the reply, thank you!

Not able to reproduce the Results

Hi,

Thanks for the wonderful work. I tried to reproduce the results for Action Recognition task on NTU RGB+D dataset on the subject split, by running the provided command.

  • python -u Indrnn_action_train.py --dropout 0.25 --use_weightdecay_nohiddenW
    The maximum Accuracy I am able to reach is 72.48%.
    Is there something needed to be changed to get the desired results?

Settings to reproduce resIndRNN results on PTB

First of all, thanks a lot for this great work!

I am trying to reproduce results with resIndRNN on word level PTB data, however following the recommended settings in the paper I was only able to get around 60 perplexity (in the paper it is around 59). Would it be possible for you to share also the configuration similar to the denseIndRNN case? Thanks a lot in advance!

By the way, to have deterministic behavior, I would also add the following three lines in train_language.py, see PyTorch note on reproducibility.

np.random.seed(seed)
torch.backends.cudnn.deterministic = True
torch.backends.cudnn.benchmark = False

Also, please correct me if I am wrong, but while reading your Theano implementation, it seems that the resIndRNN implemented there uses the original resnet configuration (dense then activation layer) while this version uses the new pre-activation configuration (activation then dense layer). Would this be the reason for the different results?

Input sequence padding

Hello,
Thank you verymuch for sharing your work.
I have a question about input shape
[48000, 300,50,3] input shape --> As per I understood ,300 represents sequence length of one .skeleton file . When we have .skeletion files such as 154 ,155 sequence lengths ,did you fill the rest of the length with zero padding to make it as 300 sequence length ?
Will it affects the accuracy of action recognition?

Thank you and waiting for your reply.

Question about action recognition experiment on this pytorch implementation!!!!

hello, thanks for your excellent work. I notice that you implementation on action recognition is different from the paper formulated. I want to know why you have those changes in your codebase?
ps: your code is return F.relu(input + hx * self.weight_hh.unsqueeze(0).expand(hx.size(0), len(self.weight_hh))) can you explain this code ? hope for your reply

Word level PTB repo?

I noticed that there are word-level PTB results in the paper but I only find character level in the repo. Is there any folder for the word level PTB?

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.