Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Understanding the Role of Pathways in a Deep Neural Network (2402.18132v1)

Published 28 Feb 2024 in cs.CV and cs.NE

Abstract: Deep neural networks have demonstrated superior performance in artificial intelligence applications, but the opaqueness of their inner working mechanism is one major drawback in their application. The prevailing unit-based interpretation is a statistical observation of stimulus-response data, which fails to show a detailed internal process of inherent mechanisms of neural networks. In this work, we analyze a convolutional neural network (CNN) trained in the classification task and present an algorithm to extract the diffusion pathways of individual pixels to identify the locations of pixels in an input image associated with object classes. The pathways allow us to test the causal components which are important for classification and the pathway-based representations are clearly distinguishable between categories. We find that the few largest pathways of an individual pixel from an image tend to cross the feature maps in each layer that is important for classification. And the large pathways of images of the same category are more consistent in their trends than those of different categories. We also apply the pathways to understanding adversarial attacks, object completion, and movement perception. Further, the total number of pathways on feature maps in all layers can clearly discriminate the original, deformed, and target samples.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (33)
  1. Understanding the role of individual units in a deep neural network. Proceedings of the National Academy of Sciences, 117(48):30071–30078, 2020.
  2. Neural response interpretation through the lens of critical pathways. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 13528–13538, 2021.
  3. Interpret neural networks by identifying critical data routing paths. In proceedings of the IEEE conference on computer vision and pattern recognition, pages 8906–8914, 2018.
  4. Analyzing the noise robustness of deep neural networks. IEEE Transactions on Visualization and Computer Graphics, 27(7):3289–3304, 2020.
  5. Recurrent computations for visual pattern completion. Proceedings of the National Academy of Sciences, 115(35):8835–8840, 2018.
  6. Recurrence is required to capture the representational dynamics of the human visual system. Proceedings of the National Academy of Sciences, 116(43):21854–21863, 2019.
  7. Compositional convolutional neural networks: A robust and interpretable model for object recognition under occlusion. International Journal of Computer Vision, 129:736–760, 2021.
  8. Correlated variability modifies working memory fidelity in primate prefrontal neuronal ensembles. Proceedings of the National Academy of Sciences, 114(12):E2494–E2503, 2017.
  9. Autocrine bdnf–trkb signalling within a single dendritic spine. Nature, 538(7623):99–103, 2016.
  10. Understanding the computation of time using neural network models. Proceedings of the National Academy of Sciences, 117(19):10530–10540, 2020.
  11. Very deep convolutional networks for large-scale image recognition. ICLR, abs/1409.1556, 2015.
  12. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278–2324, 1998.
  13. Alex Krizhevsky. Learning multiple layers of features from tiny images. In Technical Report TR-2009, University of Toronto, Toronto, 2009.
  14. Explaining and harnessing adversarial examples. In Yoshua Bengio and Yann LeCun, editors, 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, 2015.
  15. Brain dissection: fmri-trained networks reveal spatial selectivity in the processing of natural images. bioRxiv, pages 2023–05, 2023.
  16. Rise: Randomized input sampling for explanation of black-box models. arXiv preprint arXiv:1806.07421, 2018.
  17. Axiomatic attribution for deep networks. In International conference on machine learning, pages 3319–3328. PMLR, 2017.
  18. " why should i trust you?" explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, pages 1135–1144, 2016.
  19. Interpretability beyond feature attribution: Quantitative testing with concept activation vectors (tcav). In International conference on machine learning, pages 2668–2677. PMLR, 2018.
  20. Incorporating natural language into vision models improves prediction and understanding of higher visual cortex. BioRxiv, pages 2022–09, 2022.
  21. Characterizing the ventral visual stream with response-optimized neural encoding models. Advances in Neural Information Processing Systems, 35:9389–9402, 2022.
  22. The dorsal visual pathway represents object-centered spatial relations for object recognition. Journal of Neuroscience, 42(23):4693–4710, 2022.
  23. Passivity and passification of memristor-based recurrent neural networks with additive time-varying delays. IEEE transactions on neural networks and learning systems, 26(9):2043–2057, 2014.
  24. Understanding neural networks and individual neuron importance via information-ordered cumulative ablation. IEEE Transactions on Neural Networks and Learning Systems, 33(12):7842–7852, 2021.
  25. Summit: Scaling deep learning interpretability by visualizing activation and attribution summarizations. IEEE Transactions on Visualization and Computer Graphics, 26(1):1096–1106, 2020.
  26. State estimation for genetic regulatory networks with two delay components by using second-order reciprocally convex approach. Neural Processing Letters, pages 1–19, 2022.
  27. Training interpretable convolutional neural networks by differentiating class-specific filters. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part II 16, pages 622–638. Springer, 2020.
  28. Further results on input-to-state stability of stochastic cohen–grossberg bam neural networks with probabilistic time-varying delays. Neural Processing Letters, pages 1–23, 2022.
  29. Analysis of markovian jump stochastic cohen–grossberg bam neural networks with time delays for exponential input-to-state stability. Neural Processing Letters, pages 1–18, 2023.
  30. Input-to-state stability of stochastic markovian jump genetic regulatory networks. Mathematics and Computers in Simulation, 2023.
  31. Farhanhubble. M2nist, 2017. https://www.kaggle.com/datasets/farhanhubble/multimnistm2nist/, Last accessed on 2022-11-01.
  32. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE international conference on computer vision, pages 618–626, 2017.
  33. Dynamic representation of partially occluded objects in primate prefrontal and visual cortex. Elife, 6:e25784, 2017.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Lei Lyu (6 papers)
  2. Chen Pang (4 papers)
  3. Jihua Wang (4 papers)
Citations (3)

Summary

We haven't generated a summary for this paper yet.