Targeted Activation Penalties Help CNNs Ignore Spurious Signals (2311.12813v2)
Abstract: Neural networks (NNs) can learn to rely on spurious signals in the training data, leading to poor generalisation. Recent methods tackle this problem by training NNs with additional ground-truth annotations of such signals. These methods may, however, let spurious signals re-emerge in deep convolutional NNs (CNNs). We propose Targeted Activation Penalty (TAP), a new method tackling the same problem by penalising activations to control the re-emergence of spurious signals in deep CNNs, while also lowering training times and memory usage. In addition, ground-truth annotations can be expensive to obtain. We show that TAP still works well with annotations generated by pre-trained models as effective substitutes of ground-truth annotations. We demonstrate the power of TAP against two state-of-the-art baselines on the MNIST benchmark and on two clinical image datasets, using four different CNN architectures.
- Post Hoc Explanations May Be Ineffective For Detecting Unknown Spurious Correlation. In The Tenth International Conference on Learning Representations.
- Auditing Black-box Models for Indirect Influence. Knowledge and Information Systems, 54(1): 95–122.
- Explainable Artificial Intelligence (XAI): Concepts, Taxonomies, Opportunities and Challenges toward Responsible AI. arxiv:1910.10045.
- Fully Automatic Knee Osteoarthritis Severity Grading Using Deep Neural Networks with a Novel Ordinal Loss. Computerized Medical Imaging and Graphics, 75: 84–92.
- A Typology for Exploring the Mitigation of Shortcut Behaviour. Nature Machine Intelligence, 5(3): 319–330.
- Impact of Feedback Type on Explanatory Interactive Learning. In Foundations of Intelligent Systems, volume 13515 of Lecture Notes in Computer Science, 127–137. Springer International Publishing.
- Deep Residual Learning for Image Recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 770–778. IEEE.
- Women Also Snowboard: Overcoming Bias in Captioning Models. In Computer Vision – ECCV 2018, volume 11207 of Lecture Notes in Computer Science, 793–811. Springer International Publishing.
- Distilling the Knowledge in a Neural Network. arxiv:1503.02531.
- Densely Connected Convolutional Networks. In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2261–2269. IEEE.
- CheXtransfer: Performance and Parameter Efficiency of ImageNet Models for Chest X-Ray Interpretation. In Proceedings of the Conference on Health, Inference, and Learning, 116–124. Association for Computing Machinery.
- Identifying Medical Diagnoses and Treatable Diseases by Image-Based Deep Learning. Cell, 172(5): 1122–1131.e9.
- Unmasking Clever Hans Predictors and Assessing What Machines Really Learn. Nature Communications, 10(1): 1096.
- Deep Learning. Nature, 521(7553): 436–444.
- The MNIST Database of Handwritten Digits. http://yann.lecun.com/exdb/mnist/. Accessed: 2023-08-10.
- Deep Learning for Chest Radiograph Diagnosis: A Retrospective Comparison of the CheXNeXt Algorithm to Practicing Radiologists. PLOS Medicine, 15(11): e1002686.
- ”Why Should I Trust You?”: Explaining the Predictions of Any Classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 1135–1144. ACM.
- Interpretations Are Useful: Penalizing Explanations to Align Neural Networks with Prior Knowledge. In 37th International Conference on Machine Learning, ICML 2020, 8086–8096.
- Right for the Right Reasons: Training Differentiable Models by Constraining Their Explanations. In IJCAI International Joint Conference on Artificial Intelligence, 2662–2670.
- Making Deep Neural Networks Right for the Right Scientific Reasons by Interacting with Their Explanations. Nature Machine Intelligence, 2(8): 476–486.
- Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization. International Journal of Computer Vision, 128(2): 336–359.
- Right for Better Reasons: Training Differentiable Models by Constraining Their Influence Functions. In 35th AAAI Conference on Artificial Intelligence, AAAI 2021, volume 11A, 9533–9540. ISBN 9781713835974.
- Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps. arxiv:1312.6034.
- Very Deep Convolutional Networks for Large-Scale Image Recognition. arxiv:1409.1556.
- Striving for Simplicity: The All Convolutional Net. arxiv:1412.6806.
- Knowledge Transfer with Jacobian Matching. In Proceedings of the 35th International Conference on Machine Learning.
- Axiomatic Attribution for Deep Networks. In Proceedings of the 34th International Conference on Machine Learning.
- Explanatory Interactive Machine Learning. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society, 239–245.
- Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer. In 5th International Conference on Learning Representations, ICLR 2017 - Conference Track Proceedings, 1–13.
- Variable Generalization Performance of a Deep Learning Model to Detect Pneumonia in Chest Radiographs: A Cross-Sectional Study. PLOS Medicine, 15(11): e1002683.
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days freePaper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.