GithubHelp home page GithubHelp logo

arrasl / lrp_for_lstm Goto Github PK

View Code? Open in Web Editor NEW
213.0 13.0 60.0 12.69 MB

Layer-wise Relevance Propagation (LRP) for LSTMs.

License: Other

Python 56.50% Jupyter Notebook 43.50%
recurrent-neural-networks lstms interpretability xai

lrp_for_lstm's People

Contributors

arrasl avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

lrp_for_lstm's Issues

Using LRP for clickstream data

I've trained a bidirectional LSTM model on clickstream data sequences, and I want to explain the predictions using LRP. Would it be possible to adjust this implementation, so it can be used on sequential data?
A sequence, for example, looks like: [5,4,3,4,4,3,3,4,5]

How to propropagate individual hidden layer relevance scores of attention through LSTM?

My model consists of an encoder LSTM, an attention layer and a Linear decoder layer for the task of binary classification. So far I have propagated LRP all the way till the hidden layer inputs to the attention layer and am not sure how to propagate each hidden layer relevance to the input layer through the encoder LSTM.

This repo only assumes that the model is a simple encoder LSTM, and a linear decoder which takes the final hidden state as input to produce the output class, if I am right.

How can I propagate these individual hidden layer scores throught the LSTM using this approach? If I only try and propagate the last hidden layer scores through the LSTM using this code it 1. doesn't take the other hidden state scores into account 2. assumes that the attention layer only takes the last hidden state as the input.

I understand that this may be an open question, any help/advice on how to proceed will be greatly appreciated.

A question about single LSTM

Hello, I try to write a LRP for single LSTM, I found that the relevance focus on the first sequence. Just like this,
image
I don't think that the first sequence is more important. Could you please expain it?
Thank you~

Relevance sensitive sentence representations

Hello @ArrasL,
Thank you for open sourcing your implementation of LRP.
Your work here and the comparative analysis in the paper is really interesting.
I tried to add LIME, LIMSSE to the list of methods and conduct similar experiments as suggested in the latter paper.
I rechecked the scores of my implementation of gradient based methods with yours in run_example.ipynb.

I used relevance as weights for creating sentence representations, I used them in 2 settings,
a. as raw scores
b. normalized values (softmax)
I have attached the plot for your reference. I wanted to check with you how you had used them because my plot for Gradient Input is not close to the one that we see in the paper.

Any guidelines or directions would be extremely helpful as i wish to reproduce the results.

sentence_reps_3

error for run for first line


ModuleNotFoundError Traceback (most recent call last)
in ()
----> 1 from code.LSTM.LSTM_bidi import *
2 from code.util.heatmap import html_heatmap
3
4 import codecs
5 import numpy as np

ModuleNotFoundError: No module named 'code.LSTM'; 'code' is not a package

Trained LSTM model adaptation

Hello! May I assume the trained LSTM in the example consists of 1 LSTM layer? Can it be adapted into stacked LSTM scenario? Thx.

regression

Is it possible to adapt this code in order to deal with a regression problem? Let's say that each word of a sentence is a timestep of a timeseries and the number of words actually represents the history feed to the LSTM.

Architecture differences: Keras LSTM architecture to LSTM_bidi.py

I've been reviewing the bidirectional LSTM used in this example because I'd like to apply the LRP technique to my own model. However, I noticed that the weights and biases in the Keras LSTM are different from here and I would like to verify my observations are correct.

  1. In this example, there are weights on the LSTM outputs from the Left and Right LSTM. Is that the implementation of the Dense layer?

  2. In this example, there are separate biases for the h_Left/h_Right arrays and the x_Left/x_Right arrays? Keras provides only a single bias array for the two and I was wondering whether there are different LSTM architectures being followed. I also noticed that alewarne (at https://github.com/alewarne/Layerwise-Relevance-Propagation-for-LSTMs/blob/master/lstm_network.py) implemented this code with a single bias array.

Thanks for providing this reference code - I appreciate it.

Stacked/multilayer LSTM with Attention

My model consists of a multi-layer LSTM, an attention layer and a linear layer for binary classification. I need the stacked LSTMs and attention layer to achieve decent performance for a relatively complex task. I'm chugging through your detailed explanation for issue #8 (thank you!) for attention layer.

How would I propagate back the scores through the stacked LSTMs? Any guidance and advice would be much appreciated.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.