GithubHelp home page GithubHelp logo

lmartak / distill-nn-tree Goto Github PK

View Code? Open in Web Editor NEW
64.0 64.0 27.0 45.33 MB

Distillation of Neural Network Into a Soft Decision Tree

Home Page: https://vgg.fiit.stuba.sk/people/martak/distill-nn-tree

License: MIT License

Shell 0.17% Jupyter Notebook 95.44% Python 4.38%

distill-nn-tree's People

Contributors

lmartak avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar

distill-nn-tree's Issues

the loss proposed in the paper

Hi thank you for sharing this code but I had a question :
what do you think about the first log used inside the loss function that you used in line 199 what is the aim of this log

Also why did you move the minus near the second log not the first log as stated in the paper.?

Thank you

self.eps

Hi , thanks for sharing this useful code, I was wondering what is the goal of self.eps attribute for the Soft Tree ?

Thank you

Results are not reproducible

Hi, I ran your code on my own server, the results are:
[No distill]
10000/10000 [==============================] - 8s 783us/sample - loss: 7.5134 - acc: 0.9085accuracy: 90.85% | loss: 7.513414146804809
10000/10000 [==============================] - 8s 785us/sample - loss: 7.5161 - acc: 0.9032
accuracy: 90.32% | loss: 7.516079863357544

[distill with soft target]
Saving trained model to assets/distilled/tree-model.
10000/10000 [==============================] - 7s 711us/sample - loss: 7.7522 - acc: 0.8254
accuracy: 82.54% | loss: 7.7521795679092405
10000/10000 [==============================] - 8s 758us/sample - loss: 7.7434 - acc: 0.8189
accuracy: 81.89% | loss: 7.7433844789505

General Question

Thanks for sharing! How exactly do you label the paths and especially leafs? Are you storing the probabilities to look for likeliest labels at each node and look in which leafs data points ended up from your training set?

Thanks

hello,about bigger sizes input?

First of all, thank you for such a good code. I want to ask, when I input a larger size, such as 224 * 224 * 3, I find that the training has no effect, is it necessary to change some parts of the code?

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.