Convolutional Dynamic Alignment Networks for Interpretable Classifications (2104.00032v2)
Abstract: We introduce a new family of neural network models called Convolutional Dynamic Alignment Networks (CoDA-Nets), which are performant classifiers with a high degree of inherent interpretability. Their core building blocks are Dynamic Alignment Units (DAUs), which linearly transform their input with weight vectors that dynamically align with task-relevant patterns. As a result, CoDA-Nets model the classification prediction through a series of input-dependent linear transformations, allowing for linear decomposition of the output into individual input contributions. Given the alignment of the DAUs, the resulting contribution maps align with discriminative input patterns. These model-inherent decompositions are of high visual quality and outperform existing attribution methods under quantitative metrics. Further, CoDA-Nets constitute performant classifiers, achieving on par results to ResNet and VGG models on e.g. CIFAR-10 and TinyImagenet.
- Sanity Checks for Saliency Maps. In Advances in Neural Information Processing Systems (NeurIPS), 2018.
- Improved inception-residual convolutional neural network for object recognition. Neural Computing and Applications, 2020.
- Towards Robust Interpretability with Self-Explaining Neural Networks. In Advances in Neural Information Processing (NeurIPS), 2018.
- On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation. PLoS ONE, 2015.
- How to explain individual classification decisions. The Journal of Machine Learning Research, 2010.
- Approximating CNNs with Bag-of-local-Features models works surprisingly well on ImageNet. In International Conference on Learning Representations (ICLR). OpenReview.net, 2019.
- This Looks Like That: Deep Learning for Interpretable Image Recognition. In Advances in Neural Information Processing Systems (NeurIPS), 2019.
- RandAugment: Practical data augmentation with no separate search. CoRR, abs/1909.13719, 2019.
- The approximation of one matrix by another of lower rank. Psychometrika, 1936.
- Tiny ImageNet Visual Recognition Challenge. https://tiny-imagenet.herokuapp.com/. Accessed: 2020-11-10.
- MaxGain: Regularisation of Neural Networks by Constraining Activation Magnitudes. In Machine Learning and Knowledge Discovery in Databases - European Conference, ECML PKDD 2018, 2018.
- Deep Residual Learning for Image Recognition. In Conference on Computer Vision and Pattern Recognition (CVPR), 2016.
- Using Pre-Training Can Improve Model Robustness and Uncertainty. In Kamalika Chaudhuri and Ruslan Salakhutdinov, editors, Proceedings of Machine Learning Research (PMLR), 2019.
- DE-CapsNet: A Diverse Enhanced Capsule Network with Disperse Dynamic Routing. Applied Sciences, 2020.
- Dynamic filter networks. In Advances in Neural Information Processing Systems (NeurIPS), 2016.
- Alex Krizhevsky. Learning multiple layers of features from tiny images. Technical report, University of Toronto, 2009.
- learningai.io. VGGNet and Tiny ImageNet. https://learningai.io/projects/2017/06/29/tiny-imagenet.html. Accessed: 2020-11-08.
- MNIST handwritten digit database. ATT Labs [Online]. Available: http://yann.lecun.com/exdb/mnist, 2, 2010.
- A Unified Approach to Interpreting Model Predictions. In Advances in Neural Information Processing Systems (NeurIPS), 2017.
- On the Number of Linear Regions of Deep Neural Networks. In Advances in Neural Information Processing Systems (NeurIPS), 2014.
- Rectified Linear Units Improve Restricted Boltzmann Machines. In International Conference on Machine Learning (ICML), 2010.
- RISE: Randomized Input Sampling for Explanation of Black-box Models. In British Machine Vision Conference (BMVC), 2018.
- ”Why Should I Trust You?”: Explaining the predictions of any classifier. In International Conference on Knowledge Discovery and Data Mining (SIGKDD), 2016.
- Dynamic Routing Between Capsules. In Advances in Neural Information Processing Systems (NeurIPS), 2017.
- Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. In International Conference on Computer Vision (ICCV), 2017.
- Irhum Shafkat. Intuitively Understanding Convolutions for Deep Learning. https://towardsdatascience.com/intuitively-understanding-convolutions-for-deep-learning-1f6f42faee1#ad33, 2018.
- Learning Important Features Through Propagating Activation Differences. In International Conference on Machine Learning (ICML), 2017.
- Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps. In International Conference on Learning Representations (ICLR), Workshop, 2014.
- Very Deep Convolutional Networks for Large-Scale Image Recognition. In Yoshua Bengio and Yann LeCun, editors, International Conference on Learning Representations (ICLR), 2015.
- Striving for Simplicity: The All Convolutional Net. In International Conference on Learning Representations (ICLR), Workshop, 2015.
- Full-Gradient Representation for Neural Network Visualization. In Advances in Neural Information Processing Systems (NeurIPS), 2019.
- Lei Sun. ResNet on Tiny ImageNet. http://cs231n.stanford.edu/reports/2017/pdfs/12.pdf, 2016. Accessed: 2020-11-16.
- Axiomatic Attribution for Deep Networks. In Doina Precup and Yee Whye Teh, editors, International Conference on Machine Learning (ICML), 2017.
- An improved residual network model for image recognition using a combination of snapshot ensembles and the cutout technique. Multimedia Tools and Applications, 2020.
- Show, attend and tell: Neural image caption generation with visual attention. In International Conference on Machine Learning (ICML), 2015.
- On compressing deep models by low rank and sparse decomposition. In Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
- Wide Residual Networks. In British Machine Vision Conference (BMVC), 2016.
- Visualizing and Understanding Convolutional Networks. In European Conference on Computer Vision (ECCV), 2014.
- Top-Down Neural Attention by Excitation Backprop. Int. J. Comput. Vis., 2018.
- Learning Deep Features for Discriminative Localization. In Conference on Computer Vision and Pattern Recognition (CVPR), 2016.
Collections
Sign up for free to add this paper to one or more collections.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.