Efficient and Flexible Method for Reducing Moderate-size Deep Neural Networks with Condensation (2405.01041v2)
Abstract: Neural networks have been extensively applied to a variety of tasks, achieving astounding results. Applying neural networks in the scientific field is an important research direction that is gaining increasing attention. In scientific applications, the scale of neural networks is generally moderate-size, mainly to ensure the speed of inference during application. Additionally, comparing neural networks to traditional algorithms in scientific applications is inevitable. These applications often require rapid computations, making the reduction of neural network sizes increasingly important. Existing work has found that the powerful capabilities of neural networks are primarily due to their non-linearity. Theoretical work has discovered that under strong non-linearity, neurons in the same layer tend to behave similarly, a phenomenon known as condensation. Condensation offers an opportunity to reduce the scale of neural networks to a smaller subnetwork with similar performance. In this article, we propose a condensation reduction algorithm to verify the feasibility of this idea in practical problems. Our reduction method can currently be applied to both fully connected networks and convolutional networks, achieving positive results. In complex combustion acceleration tasks, we reduced the size of the neural network to 41.7% of its original scale while maintaining prediction accuracy. In the CIFAR10 image classification task, we reduced the network size to 11.5% of the original scale, still maintaining a satisfactory validation accuracy. Our method can be applied to most trained neural networks, reducing computational pressure and improving inference speed.
- Three ways to solve partial differential equations with neural networks—a review. GAMM-Mitteilungen, 44(2):e202100006, 2021.
- Shallow neural networks for fluid flow reconstruction with limited sensors. Proceedings of the Royal Society A, 476(2238):20200097, 2020.
- Neural operator: Graph kernel network for partial differential equations. arXiv preprint arXiv:2003.03485, 2020.
- Rethinking the value of network pruning. arXiv preprint arXiv:1810.05270, 2018.
- Phase diagram for two-layer relu neural networks at infinite-width limit. Journal of Machine Learning Research, 22(71):1–47, 2021.
- Solving differential equations using deep neural networks. Neurocomputing, 399:193–212, 2020.
- Graph neural networks for materials science and chemistry. Communications Materials, 3(1):93, 2022.
- Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. Journal of Computational physics, 378:686–707, 2019.
- Resource-efficient neural networks for embedded systems. Journal of Machine Learning Research, 25(50):1–51, 2024.
- Graph neural networks in particle physics. Machine Learning: Science and Technology, 2(2):021001, 2020.
- Astronomia ex machina: a history, primer and outlook on neural networks in astronomy. Royal Society Open Science, 10(5):221454, 2023.
- Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 4510–4520, 2018.
- DR Sarvamangala and Raghavendra V Kulkarni. Convolutional neural networks in medical image understanding: a survey. Evolutionary intelligence, 15(1):1–22, 2022.
- Pieter Wesseling. Introduction to multigrid methods. Technical report, 1995.
- Convolutional neural network pruning with structural redundancy reduction. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 14913–14922, 2021.
- Solving multiscale dynamical systems by deep learning. arXiv preprint arXiv:2401.01220, 2024.
- Graph neural networks and their current applications in bioinformatics. Frontiers in genetics, 12:690049, 2021.
- Machine learning: new ideas and tools in environmental science and engineering. Environmental Science & Technology, 55(19):12741–12754, 2021.
- Quasi-dns dataset of a piloted flame with inhomogeneous inlet conditions. Flow, Turbulence and Combustion, 104:997–1027, 2020.
- Embedding principle of loss landscape of deep neural networks. Advances in Neural Information Processing Systems, 34:14848–14859, 2021.