GithubHelp home page GithubHelp logo

allensmile / foundations_for_deep_learning Goto Github PK

View Code? Open in Web Editor NEW

This project forked from threefoldo/foundations_for_deep_learning

0.0 2.0 0.0 67.99 MB

Building a scalable foundation for deep learning

foundations_for_deep_learning's Introduction

foundations for deep learning:

  1. I emphasize mathematical/conceptual foundations because implementations of ideas(ex. Torch, Tensorflow) will keep evolving but the underlying theory must be sound. Anybody with an interest in deep learning can and should try to understand why things work.
  2. I include neuroscience as a useful conceptual foundation for two reasons. First, it may provide inspiration for future models and algorithms. Second, the success of deep learning can contribute to useful hypotheses and models for computational neuroscience.
  3. Information Theory is also a very useful foundation as there's a strong connection between data compression and statistical prediction. In fact, data compressors and machine learning models approximate Kolmogorov Complexity which is the ultimate data compressor.

You might notice that I haven't emphasized the latest benchmark-beating paper. My reason for this is that a good theory ought to be scalable which means that it should be capable of explaining why deep models generalise and we should be able to bootstrap these explanations for more complex models(ex. sequences of deep models(aka RNNs)). This is how all good science is done.

For an excellent historical overview of deep learning, I would recommend reading Deep Learning in Neural Networks.

Classics:

  1. Learning Deep Generative Models(Salakhutdinov 2015. Annual Review of Statistics and Its Application. )
  2. Uncertainty in Deep Learning(Yarin Gal 2017. University of Cambridge.)
  3. Markov Chain Monte Carlo and Variational Inference: Bridging the Gap (Salimans 2015. ICML. )
  4. Weight Normalization (Salimans 2016. NIPS.)
  5. Mixture Density Networks (Bishop 1994)
  6. Dropout as a Bayesian Approximation(Yarin Gal 2016. ICML. )
  7. Why does unsupervised pre-training help deep learning?(Erhan et al. 2010. Journal of Machine Learning Research.)
  8. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift(S. Ioffe 2015. ICML.)
  9. Bayesian Back-Propagation (W. Buntine & A. Weigend 1991. )
  10. Deep Learning in Neural Networks: An Overview (J. Schmidhuber. 2015. Neural Networks.)

Mathematical papers:

  1. Dropout Rademacher Complexity of Deep Neural Networks(Wei Gao 2015. Science China Information Sciences.)
  2. Distribution-Specific Hardness of Learning Neural Networks(Shamir 2017. CoRR.)
  3. Lessons from the Rademacher Complexity for Deep Learning(Sokolic 2016.ICLR.)
  4. A mathematical theory of Deep Convolutional Neural Networks for Feature Extraction(Wiatowski 2016. CoRR.)
  5. Spectral Representations for Convolutional Neural Networks(Rippl 2015. NIPS.)
  6. Electron-Proton dynamics in deep learning(Zhang 2017. CoRR.)
  7. Empirical Risk Minimization for Learning Theory(Vapnik 1991. NIPS.)
  8. The Loss Surfaces of Multilayer Networks(Y LeCun et al. 2015. AISTATS.)
  9. Understanding Synthetic Gradients and Decoupled Neural Interfaces(W. Czarnecki 2017. CoRR.)
  10. Dataset Shift(Storkey 2013)
  11. The loss surface of deep and wide neural networks(Q. Nguyen 2017)
  12. A Differential Equation for Modeling Nesterov’s Accelerated Gradient Method(W. Su 2016. Journal of Machine Learning Research.)
  13. The Physical Systems behind Optimization (L. Yang et al. 2017.)
  14. Simple Explanation of the No-Free-Lunch Theorem and Its Implications (Y. Ho. 2002. Journal of optimization theory and applications.)
  15. Adam: A method for Stochastic Optimization (D. Kingma 2015. ICLR.)
  16. On the ability of neural nets to express distributions (H. Lee, R. Ge, T. Ma, A. Risteski & S. Arora, 2017)

Information Theory papers:

  1. Shannon Information and Kolmogorov Complexity (Grunwald 2010)
  2. Discovering Neural Nets with Low Kolmogorov Complexity(Schmidhuber 1997. Neural Networks.)
  3. Opening the black box of Deep Neural Networks via Information (Schwartz-Ziv 2017.)

Neuroscience papers:

  1. Towards an integration of deep learning and neuroscience(Marblestone 2016. Frontiers in Computational Neuroscience.)
  2. Equilibrium Propagation(Scellier 2016. Frontiers in Computational Neuroscience.)
  3. Towards Biologically plausible deep learning(Bengio 2015. CoRR.)
  4. Random synaptic feedback weights support error backpropagation for deep learning(Lillicrap 2016. Nature communications.)
  5. Towards deep learning with spiking neurons(Mesnard 2016. NIPS.)
  6. Towards deep learning with spiking dendrites(Guergiuev 2017)
  7. Variational learning for recurrent spiking networks(Rezende 2011. NIPS.)
  8. A view of Neural Networks as dynamical systems(Cessac 2009. I. J. Bifurcation and Chaos)

Note: This is a work in progress. I have a lot more papers to add.

foundations_for_deep_learning's People

Contributors

aidanrocke avatar

Watchers

 avatar  avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.