Go to file
2018-06-02 15:21:47 -05:00
LICENSE Initial commit 2017-12-21 12:33:57 -06:00
README.md Update README.md 2018-06-02 15:21:47 -05:00

Papers on Understanding and Explaining Neural Networks

This is an on-going attempt to consolidate all interesting efforts in the area of understanding / interpreting / explaining / visualizing neural networks.


1. GUI tools

  • Deep Visualization Toolbox code | pdf

2. Feature Visualization / Activation Maximization

  • Synthesizing the preferred inputs for neurons in neural networks via deep generator networks code | pdf
  • Plug and Play Generative Networks pdf | code

3. Heatmap / Attribution

  • Learning how to explain neural networks: PatternNet and PatternAttribution (pdf)
  • A Theoretical Explanation for Perplexing Behaviors of Backpropagation-based Visualizations pdf

Layer-wise Backpropagation

  • Beyond saliency: understanding convolutional neural networks from saliency prediction on layer-wise relevance propagation pdf
  • Explaining NonLinear Classification Decisions With Deep Tayor Decomposition pdf

4. Inverting Neural Networks

  • Understanding Deep Image Representations by Inverting Them pdf
  • Inverting Visual Representations with Convolutional Networks pdf
  • Neural network inversion beyond gradient descent pdf

5. Bayesian approaches

  • Yang, S. C. H., & Shafto, P. Explainable Artificial Intelligence via Bayesian Teaching. NIPS 2017 pdf

6. Distilling DNNs into more interpretable models

  • Interpreting CNNs via Decision Trees pdf
  • Distilling a Neural Network Into a Soft Decision Tree pdf

7. DNNs that learn to explain

  • Deep Learning for Case-Based Reasoning through Prototypes pdf

8. Others

  • Understanding Deep Architectures by Interpretable Visual Summaries pdf