Papers
Topics
Authors
Recent
2000 character limit reached

Targeted Activation Penalties Help CNNs Ignore Spurious Signals (2311.12813v2)

Published 22 Sep 2023 in cs.CV and cs.LG

Abstract: Neural networks (NNs) can learn to rely on spurious signals in the training data, leading to poor generalisation. Recent methods tackle this problem by training NNs with additional ground-truth annotations of such signals. These methods may, however, let spurious signals re-emerge in deep convolutional NNs (CNNs). We propose Targeted Activation Penalty (TAP), a new method tackling the same problem by penalising activations to control the re-emergence of spurious signals in deep CNNs, while also lowering training times and memory usage. In addition, ground-truth annotations can be expensive to obtain. We show that TAP still works well with annotations generated by pre-trained models as effective substitutes of ground-truth annotations. We demonstrate the power of TAP against two state-of-the-art baselines on the MNIST benchmark and on two clinical image datasets, using four different CNN architectures.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (30)
  1. Post Hoc Explanations May Be Ineffective For Detecting Unknown Spurious Correlation. In The Tenth International Conference on Learning Representations.
  2. Auditing Black-box Models for Indirect Influence. Knowledge and Information Systems, 54(1): 95–122.
  3. Explainable Artificial Intelligence (XAI): Concepts, Taxonomies, Opportunities and Challenges toward Responsible AI. arxiv:1910.10045.
  4. Fully Automatic Knee Osteoarthritis Severity Grading Using Deep Neural Networks with a Novel Ordinal Loss. Computerized Medical Imaging and Graphics, 75: 84–92.
  5. A Typology for Exploring the Mitigation of Shortcut Behaviour. Nature Machine Intelligence, 5(3): 319–330.
  6. Impact of Feedback Type on Explanatory Interactive Learning. In Foundations of Intelligent Systems, volume 13515 of Lecture Notes in Computer Science, 127–137. Springer International Publishing.
  7. Deep Residual Learning for Image Recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 770–778. IEEE.
  8. Women Also Snowboard: Overcoming Bias in Captioning Models. In Computer Vision – ECCV 2018, volume 11207 of Lecture Notes in Computer Science, 793–811. Springer International Publishing.
  9. Distilling the Knowledge in a Neural Network. arxiv:1503.02531.
  10. Densely Connected Convolutional Networks. In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2261–2269. IEEE.
  11. CheXtransfer: Performance and Parameter Efficiency of ImageNet Models for Chest X-Ray Interpretation. In Proceedings of the Conference on Health, Inference, and Learning, 116–124. Association for Computing Machinery.
  12. Identifying Medical Diagnoses and Treatable Diseases by Image-Based Deep Learning. Cell, 172(5): 1122–1131.e9.
  13. Unmasking Clever Hans Predictors and Assessing What Machines Really Learn. Nature Communications, 10(1): 1096.
  14. Deep Learning. Nature, 521(7553): 436–444.
  15. The MNIST Database of Handwritten Digits. http://yann.lecun.com/exdb/mnist/. Accessed: 2023-08-10.
  16. Deep Learning for Chest Radiograph Diagnosis: A Retrospective Comparison of the CheXNeXt Algorithm to Practicing Radiologists. PLOS Medicine, 15(11): e1002686.
  17. ”Why Should I Trust You?”: Explaining the Predictions of Any Classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 1135–1144. ACM.
  18. Interpretations Are Useful: Penalizing Explanations to Align Neural Networks with Prior Knowledge. In 37th International Conference on Machine Learning, ICML 2020, 8086–8096.
  19. Right for the Right Reasons: Training Differentiable Models by Constraining Their Explanations. In IJCAI International Joint Conference on Artificial Intelligence, 2662–2670.
  20. Making Deep Neural Networks Right for the Right Scientific Reasons by Interacting with Their Explanations. Nature Machine Intelligence, 2(8): 476–486.
  21. Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization. International Journal of Computer Vision, 128(2): 336–359.
  22. Right for Better Reasons: Training Differentiable Models by Constraining Their Influence Functions. In 35th AAAI Conference on Artificial Intelligence, AAAI 2021, volume 11A, 9533–9540. ISBN 9781713835974.
  23. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps. arxiv:1312.6034.
  24. Very Deep Convolutional Networks for Large-Scale Image Recognition. arxiv:1409.1556.
  25. Striving for Simplicity: The All Convolutional Net. arxiv:1412.6806.
  26. Knowledge Transfer with Jacobian Matching. In Proceedings of the 35th International Conference on Machine Learning.
  27. Axiomatic Attribution for Deep Networks. In Proceedings of the 34th International Conference on Machine Learning.
  28. Explanatory Interactive Machine Learning. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society, 239–245.
  29. Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer. In 5th International Conference on Learning Representations, ICLR 2017 - Conference Track Proceedings, 1–13.
  30. Variable Generalization Performance of a Deep Learning Model to Detect Pneumonia in Chest Radiographs: A Cross-Sectional Study. PLOS Medicine, 15(11): e1002683.
Citations (1)

Summary

We haven't generated a summary for this paper yet.

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.