- The paper introduces a grow-and-prune paradigm that starts with a sparse seed and dynamically enhances the network using gradient-based growth and magnitude-based pruning.
- The paper achieves significant reductions in parameters and FLOPs, with models like LeNet-5 experiencing up to 74.3× fewer parameters and 43.7× reduced FLOPs without sacrificing accuracy.
- The paper demonstrates that NeST overcomes reliance on human-designed architectures, offering an automated synthesis approach ideal for resource-constrained applications.
NeST: A Neural Network Synthesis Tool Based on a Grow-and-Prune Paradigm
The paper presents NeST, a neural network synthesis tool designed to optimize deep neural network (DNN) architectures by utilizing a novel grow-and-prune paradigm. This approach specifically addresses the challenges associated with the derivation of optimal DNN architectures for large-scale applications. Conventional methods tend to develop increasingly deeper and larger networks, often resulting in significant redundancy. NeST aims to alleviate this issue by combining network growth and pruning into a coherent workflow that automates the generation of compact and high-performing DNNs.
The methodology behind NeST is inspired by the dynamic nature of synaptic connections in the human brain, where the number of connections first increases and then decreases over time. NeST begins with a sparse initial architecture, referred to as the "seed architecture," and iteratively enhances it. The growth phase involves adding neurons and connections based on gradient information to improve accuracy. This process is followed by a pruning phase that removes redundant elements identified by their magnitude, allowing the network to maintain compactness without compromising accuracy.
Experimental results demonstrate NeST's efficacy in synthesizing both the LeNet and AlexNet architectures. Specifically, NeST achieved a significant reduction in the number of parameters and floating-point operations (FLOPs) with no accuracy sacrifice: LeNet-300-100 and LeNet-5 saw parameter reductions of 70.2× and 74.3×, respectively, alongside reduced FLOPs of 79.4× and 43.7×. Similarly, for AlexNet and VGG-16, NeST reduced parameters (FLOPs) by 15.7× (4.6×) and 30.2× (8.6×), respectively, all while maintaining performance levels on par with baseline models.
Critically, NeST diverges from traditional DNN design paradigms by starting from a potentially arbitrary and sparse seed architecture, overcoming the reliance on human intuition for selecting baseline architectures. The grow-and-prune strategy fosters significant reductions in redundant components, leading to lightweight DNN models.
Theoretical implications of this research highlight an enhanced understanding of neural architecture search (NAS) and architecture design space. Practically, NeST promises substantial improvements in efficiency for deploying DNNs, particularly in resource-constrained environments such as mobile or edge devices, where computational capabilities and memory are limited.
Future research endeavors may explore extending NeST's application to more complex architectures like ResNet or DenseNet, addressing the associated temporospatial overheads in training large models within the grow-and-prune paradigm without utilizing memory-intensive masks.
In conclusion, NeST represents a notable advancement in automated DNN synthesis, marrying the biological inspiration of synaptic pruning with computational efficiency. Its contributions to achieving compact and efficient DNNs open avenues for deployment in a broader range of applications without performance compromises.