Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
175 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

On-Off Pattern Encoding and Path-Count Encoding as Deep Neural Network Representations (2401.09518v1)

Published 17 Jan 2024 in cs.CV

Abstract: Understanding the encoded representation of Deep Neural Networks (DNNs) has been a fundamental yet challenging objective. In this work, we focus on two possible directions for analyzing representations of DNNs by studying simple image classification tasks. Specifically, we consider \textit{On-Off pattern} and \textit{PathCount} for investigating how information is stored in deep representations. On-off pattern of a neuron is decided as on' oroff' depending on whether the neuron's activation after ReLU is non-zero or zero. PathCount is the number of paths that transmit non-zero energy from the input to a neuron. We investigate how neurons in the network encodes information by replacing each layer's activation with On-Off pattern or PathCount and evaluating its effect on classification performance. We also examine correlation between representation and PathCount. Finally, we show a possible way to improve an existing DNN interpretation method, Class Activation Map (CAM), by directly utilizing On-Off or PathCount.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (28)
  1. Hervé Abdi. The kendall rank correlation coefficient. Encyclopedia of Measurement and Statistics. Sage, Thousand Oaks, CA, pages 508–510, 2007.
  2. Towards better understanding of gradient-based attribution methods for deep neural networks. arXiv preprint arXiv:1711.06104, 2017.
  3. Improving deep neural networks for lvcsr using rectified linear units and dropout. In 2013 IEEE international conference on acoustics, speech and signal processing, pages 8609–8613. IEEE, 2013.
  4. Visualizing higher-layer features of a deep network. University of Montreal, 1341(3):1, 2009.
  5. Complexity of linear regions in deep networks. arXiv preprint arXiv:1901.09021, 2019.
  6. Deep relu networks have surprisingly few activation patterns. In Advances in Neural Information Processing Systems, pages 361–370, 2019.
  7. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
  8. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861, 2017.
  9. Imagenet classification with deep convolutional neural networks. In Advances in neural information processing systems, pages 1097–1105, 2012.
  10. Does resnet learn good general purpose features? In Proceedings of the 2017 International Conference on Artificial Intelligence, Automation and Control Technologies, pages 1–5, 2017.
  11. Inception and resnet features are (almost) equivalent. Cognitive Systems Research, 59:312–318, 2020.
  12. On the number of linear regions of deep neural networks. In Advances in neural information processing systems, pages 2924–2932, 2014.
  13. Rectified linear units improve restricted boltzmann machines. In ICML, 2010.
  14. Activation functions: Comparison of trends in practice and research for deep learning. arXiv preprint arXiv:1811.03378, 2018.
  15. On the number of response regions of deep feed forward networks with piece-wise linear activations. arXiv preprint arXiv:1312.6098, 2013.
  16. Adversarial defense through network profiling based path extraction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 4777–4786, 2019.
  17. Searching for activation functions. arXiv preprint arXiv:1710.05941, 2017.
  18. Evaluating the visualization of what a deep neural network has learned. IEEE transactions on neural networks and learning systems, 28(11):2660–2673, 2016.
  19. Restricting the flow: Information bottlenecks for attribution. arXiv preprint arXiv:2001.00396, 2020.
  20. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE international conference on computer vision, pages 618–626, 2017.
  21. Sagar Sharma. Activation functions in neural networks. Towards Data Science, 6, 2017.
  22. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556, 2014.
  23. Going deeper with convolutions. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1–9, 2015.
  24. Interpret neural networks by identifying critical data routing paths. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 8906–8914, 2018.
  25. Visualizing and comparing convolutional neural networks. arXiv preprint arXiv:1412.6631, 2014.
  26. Visualizing and understanding convolutional networks. In European conference on computer vision, pages 818–833. Springer, 2014.
  27. On rectified linear units for speech processing. In 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, pages 3517–3521. IEEE, 2013.
  28. Learning deep features for discriminative localization. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2921–2929, 2016.

Summary

We haven't generated a summary for this paper yet.