LICENSE | ||
README.md |
Papers on Understanding and Explaining Neural Networks
This is an on-going attempt to consolidate all interesting efforts in the area of understanding / interpreting / explaining / visualizing neural networks.
1. GUI tools
2. Feature Visualization / Activation Maximization
- Synthesizing the preferred inputs for neurons in neural networks via deep generator networks code | pdf
- Plug and Play Generative Networks pdf | code
3. Heatmap / Attribution
- Learning how to explain neural networks: PatternNet and PatternAttribution (pdf)
- A Theoretical Explanation for Perplexing Behaviors of Backpropagation-based Visualizations pdf
Layer-wise Backpropagation
- Beyond saliency: understanding convolutional neural networks from saliency prediction on layer-wise relevance propagation pdf
- Explaining NonLinear Classification Decisions With Deep Tayor Decomposition pdf
4. Inverting Neural Networks
- Understanding Deep Image Representations by Inverting Them pdf
- Inverting Visual Representations with Convolutional Networks pdf
- Neural network inversion beyond gradient descent pdf
5. Bayesian approaches
- Yang, S. C. H., & Shafto, P. Explainable Artificial Intelligence via Bayesian Teaching. NIPS 2017 pdf
6. Distilling DNNs into more interpretable models
7. DNNs that learn to explain
- Deep Learning for Case-Based Reasoning through Prototypes pdf
- Unsupervised Learning of Neural Networks to Explain Neural Networks pdf
8. Others
- Understanding Deep Architectures by Interpretable Visual Summaries pdf