QuadraNet: Improving High-Order Neural Interaction Efficiency with Hardware-Aware Quadratic Neural Networks (2311.17956v1)
Abstract: Recent progress in computer vision-oriented neural network designs is mostly driven by capturing high-order neural interactions among inputs and features. And there emerged a variety of approaches to accomplish this, such as Transformers and its variants. However, these interactions generate a large amount of intermediate state and/or strong data dependency, leading to considerable memory consumption and computing cost, and therefore compromising the overall runtime performance. To address this challenge, we rethink the high-order interactive neural network design with a quadratic computing approach. Specifically, we propose QuadraNet -- a comprehensive model design methodology from neuron reconstruction to structural block and eventually to the overall neural network implementation. Leveraging quadratic neurons' intrinsic high-order advantages and dedicated computation optimization schemes, QuadraNet could effectively achieve optimal cognition and computation performance. Incorporating state-of-the-art hardware-aware neural architecture search and system integration techniques, QuadraNet could also be well generalized in different hardware constraint settings and deployment scenarios. The experiment shows thatQuadraNet achieves up to 1.5$\times$ throughput, 30% less memory footprint, and similar cognition performance, compared with the state-of-the-art high-order approaches.
- A. Krizhevsky, et al., “ImageNet Classification with Deep Convolutional Neural Networks,” in NeurIPS, 2012.
- K. He, et al., “Deep Residual Learning for Image Recognition,” in CVPR, 2016.
- A. Dosovitskiy, et al., “An Image Is Worth 16x16 Words: Transformers for Image Recognition at Scale,” in ICLR, 2021.
- Z. Liu, et al., “Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows,” in ICCV, 2021.
- S. Mehta et al., “MobileViT: Light-Weight, General-Purpose, and Mobile-Friendly Vision Transformer,” arXiv:2110.02178, 2021.
- M. Sandler, et al., “Mobilenetv2: Inverted Residuals and Linear Bottlenecks,” in CVPR, 2018.
- F. D. Keles, et al., “On the Computational Complexity of Self-Attention,” arXiv:2209.04881, 2022.
- S. Wang, et al., “Linformer: Self-attention with linear complexity,” arXiv preprint arXiv:2006.04768, 2020.
- Y. Rao, et al., “HorNet: Efficient High-Order Spatial Interactions with Recursive Gated Convolutions,” arXiv:2207.14284, 2022.
- G. G. Chrysos, et al., “P-Nets: Deep Polynomial Neural Networks,” in CVPR, 2020.
- F. Fan, et al., “A New Type of Neurons for Machine Learning,” International Journal for Numerical Methods in Biomedical Engineering, 2018.
- Z. Xu, et al., “QuadraLib: A Performant Quadratic Neural Network Library for Architecture Optimization and Design Exploration,” in MLSys, 2022.
- P. Mantini et al., “CQNN: Convolutional Quadratic Neural Networks,” in ICPR, 2021.
- C. Gong, et al., “NASViT: Neural Architecture Search for Efficient Vision Transformers with Gradient Conflict-Aware SuperNet Training,” in ICLR, 2021.
- T. Chen, et al., “TVM: An Automated End-to-End Optimizing Compiler for Deep Learning,” in OSDI, 2018.
- Z. Liu, et al., “A ConvNet for the 2020s,” in CVPR, 2022.
- H. Touvron, et al., “Training Data-Efficient Image Transformers & Distillation through Attention,” in ICML, 2021.
- K. Han, et al., “Transformer in Transformer,” in NeurIPS, 2021.
- W. Wang, et al., “Pyramid vision transformer: A versatile backbone for dense prediction without convolutions,” in ICCV, 2021.
- L. Yuan, et al., “Tokens-to-Token ViT: Training Vision Transformers from Scratch on ImageNet,” in ICCV, 2021.
- H. Cai, et al., “Proxylessnas: Direct neural architecture search on target task and hardware,” in ICLR, 2018.
- J. Deng, et al., “Imagenet: A large-scale hierarchical image database,” in CVPR, 2009.
- I. Loshchilov et al., “Decoupled weight decay regularization,” in ICLR, 2018.
- L. L. Zhang, et al., “nn-meter: towards accurate latency prediction of deep-learning model inference on diverse edge devices,” in MobiSys, 2021.
- Y. Rao, et al., “Global filter networks for image classification,” NeurIPS, 2021.
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days freePaper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.