Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
173 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Parallel Algorithms for Exact Enumeration of Deep Neural Network Activation Regions (2403.00860v1)

Published 29 Feb 2024 in cs.LG, cs.AI, and cs.NE

Abstract: A feedforward neural network using rectified linear units constructs a mapping from inputs to outputs by partitioning its input space into a set of convex regions where points within a region share a single affine transformation. In order to understand how neural networks work, when and why they fail, and how they compare to biological intelligence, we need to understand the organization and formation of these regions. Step one is to design and implement algorithms for exact region enumeration in networks beyond toy examples. In this work, we present parallel algorithms for exact enumeration in deep (and shallow) neural networks. Our work has three main contributions: (1) we present a novel algorithm framework and parallel algorithms for region enumeration; (2) we implement one of our algorithms on a variety of network architectures and experimentally show how the number of regions dictates runtime; and (3) we show, using our algorithm's output, how the dimension of a region's affine transformation impacts further partitioning of the region by deeper layers. To our knowledge, we run our implemented algorithm on networks larger than all of the networks used in the existing region enumeration literature. Further, we experimentally demonstrate the importance of parallelism for region enumeration of any reasonably sized network.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (27)
  1. David Avis and Komei Fukuda. 1996. Reverse search for enumeration. Discrete Applied Mathematics 65, 1-3 (March 1996), 21–46. https://doi.org/10.1016/0166-218X(95)00026-N
  2. David Avis and Charles Jordan. 2016. A parallel framework for reverse search using mts. arXiv preprint arXiv:1610.07735 (2016).
  3. David Avis and Charles Jordan. 2018. mplrs: A scalable parallel vertex/facet enumeration code. Mathematical Programming Computation 10, 2 (2018), 267–302.
  4. David Avis and Charles Jordan. 2021. mts: a light framework for parallelizing tree search codes. Optimization methods and software 36, 2-3 (2021), 279–300.
  5. Randall Balestriero and Yann LeCun. 2023. Fast and Exact Enumeration of Deep Networks Partitions Regions. In ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). 1–5. https://doi.org/10.1109/ICASSP49357.2023.10095698
  6. Yoshua Bengio. 2009. Learning Deep Architectures for AI. Foundations and Trends in Machine Learning 2 (Jan. 2009).
  7. Monica Bianchini and Franco Scarselli. 2014. On the complexity of shallow and deep neural network classifiers.. In ESANN.
  8. Adrian ChmielewskiI-Anders. 2020. Activation Regions as a Proxy for Understanding Neural Networks. Ph. D. Dissertation.
  9. EMNIST: Extending MNIST to handwritten letters. In 2017 international joint conference on neural networks (IJCNN). IEEE, 2921–2926.
  10. Boris Hanin and David Rolnick. 2019. Deep ReLU Networks Have Surprisingly Few Activation Patterns. http://arxiv.org/abs/1906.00904 arXiv:1906.00904 [cs, math, stat].
  11. When Deep Learning Meets Polyhedral Theory: A Survey. (2023).
  12. Matthew Hutson. 2018. AI researchers allege that machine learning is alchemy. Science (May 2018). https://doi.org/10.1126/science.aau0577
  13. Diederik P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014).
  14. Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems 25 (2012).
  15. On the number of linear regions of deep neural networks. Advances in neural information processing systems 27 (2014).
  16. Sensitivity and Generalization in Neural Networks: an Empirical Study. http://arxiv.org/abs/1802.08760 arXiv:1802.08760 [cs, stat].
  17. Gurobi Optimization. 2014. Inc.,“Gurobi optimizer reference manual,” 2015. (2014).
  18. On the number of response regions of deep feed forward networks with piece-wise linear activations. arXiv preprint arXiv:1312.6098 (2013).
  19. Miroslav Rada and Michal Černý. 2018. A New Algorithm for Enumeration of Cells of Hyperplane Arrangements and a Comparison with Avis and Fukuda’s Reverse Search. SIAM Journal on Discrete Mathematics 32, 1 (Jan. 2018), 455–473. https://doi.org/10.1137/15M1027930
  20. Dissecting Deep Neural Networks. http://arxiv.org/abs/1910.03879 arXiv:1910.03879 [cs, stat].
  21. Ludwig Schläfli. 1901. Anzeige einer Abhandlung über die Theorie der vielfachen Kontinuität. Springer.
  22. Thiago Serra and Srikumar Ramalingam. 2019. Empirical Bounds on Linear Regions of Deep Rectifier Networks. http://arxiv.org/abs/1810.03370
  23. Bounding and Counting Linear Regions of Deep Neural Networks. In NeurIPS.
  24. Mastering the game of Go with deep neural networks and tree search. nature 529, 7587 (2016), 484–489.
  25. Nora Sleumer. 2000. Hyperplane arrangements: construction, visualization and application. Ph. D. Dissertation. ETH Zurich. https://doi.org/10.3929/ETHZ-A-003889994 Artwork Size: 104 S. Medium: application/pdf Pages: 104 S..
  26. TROPEX: AN ALGORITHM FOR EXTRACTING LINEAR TERMS IN DEEP NEURAL NETWORKS. (2021).
  27. Mattia Jacopo Villani and Nandi Schoots. 2023. Any Deep ReLU Network is Shallow. http://arxiv.org/abs/2306.11827 arXiv:2306.11827 [cs, stat].

Summary

We haven't generated a summary for this paper yet.