Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Zen-NAS: A Zero-Shot NAS for High-Performance Deep Image Recognition (2102.01063v4)

Published 1 Feb 2021 in cs.CV and cs.LG

Abstract: Accuracy predictor is a key component in Neural Architecture Search (NAS) for ranking architectures. Building a high-quality accuracy predictor usually costs enormous computation. To address this issue, instead of using an accuracy predictor, we propose a novel zero-shot index dubbed Zen-Score to rank the architectures. The Zen-Score represents the network expressivity and positively correlates with the model accuracy. The calculation of Zen-Score only takes a few forward inferences through a randomly initialized network, without training network parameters. Built upon the Zen-Score, we further propose a new NAS algorithm, termed as Zen-NAS, by maximizing the Zen-Score of the target network under given inference budgets. Within less than half GPU day, Zen-NAS is able to directly search high performance architectures in a data-free style. Comparing with previous NAS methods, the proposed Zen-NAS is magnitude times faster on multiple server-side and mobile-side GPU platforms with state-of-the-art accuracy on ImageNet. Our source code and pre-trained models are released on https://github.com/idstcv/ZenNAS.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Ming Lin (65 papers)
  2. Pichao Wang (65 papers)
  3. Zhenhong Sun (12 papers)
  4. Hesen Chen (8 papers)
  5. Xiuyu Sun (25 papers)
  6. Qi Qian (54 papers)
  7. Hao Li (803 papers)
  8. Rong Jin (164 papers)
Citations (87)

Summary

Zen-NAS: A Zero-Shot NAS for High-Performance Image Recognition

This paper presents Zen-NAS, an innovative Neural Architecture Search (NAS) framework that eliminates the need for computationally intensive accuracy predictors. Current NAS methodologies are often hindered by the burdensome costs associated with training numerous networks for evaluating their architectures. Zen-NAS addresses this challenge through the introduction of the Zen-Score, a zero-cost proxy for ranking neural networks based on expressivity.

Zen-Score is calculated by performing forward inferences on a randomly initialized network, without necessitating any parameter training. This approach significantly reduces computational demands, enabling high-performance architecture search within just 0.5 GPU days, as opposed to traditional methods that require extensive resources. Zen-NAS leverages this rapid computation, optimizing architectures under various inference constraints while maintaining competitive accuracy levels on the ImageNet dataset.

The methodology outlined in the paper proposes using the Zen-Score as a novel index reflecting network expressivity—a metric correlating positively with accuracy. The computations are performed on randomly initialized networks, employing Gaussian inputs to swiftly approximate network expressivity. Notably, Zen-Score is shown to manage scale sensitivity effectively, particularly in handling Batch Normalization layers, making it adaptable to real-world network design challenges.

Zen-NAS differentiates itself by providing a zero-shot NAS approach focused on maximizing the Zen-Score of candidate architectures. This enables the rapid identification of high-performance networks without the typical iterative learning and fine-tuning processes. The empirical outcomes demonstrate substantial enhancements in speed and efficiency across various hardware settings, including server-side and mobile GPUs.

This research brings forth significant implications. Practically, Zen-NAS can democratize access to efficient NAS by reducing the computational barriers for researchers and developers without extensive hardware resources. From a theoretical perspective, the paper challenges conventional NAS paradigms, advocating for exploration in expressivity-driven, data-free NAS methods.

Potential future trajectories of this work could involve further refinement of the Zen-Score to account for more complex network operations or its integration into hybrid NAS approaches, combining zero-shot efficiency with scalable adjustment strategies. Moreover, as neural networks continue to proliferate in varied applications, extending Zen-NAS to address domain-specific performance requirements could be instrumental in advancing the versatility and applicability of automated neural architecture design.

In summary, Zen-NAS exhibits remarkable promise in reimagining NAS by channeling theoretical insights into practical, scalable solutions. It invites the research community to rethink architecture design, grounding future explorations in the efficient evaluation of expressivity while simplifying the computational landscape of NAS.

Github Logo Streamline Icon: https://streamlinehq.com

GitHub