Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

NAS-LID: Efficient Neural Architecture Search with Local Intrinsic Dimension (2211.12759v2)

Published 23 Nov 2022 in cs.CV, cs.AI, and cs.LG

Abstract: One-shot neural architecture search (NAS) substantially improves the search efficiency by training one supernet to estimate the performance of every possible child architecture (i.e., subnet). However, the inconsistency of characteristics among subnets incurs serious interference in the optimization, resulting in poor performance ranking correlation of subnets. Subsequent explorations decompose supernet weights via a particular criterion, e.g., gradient matching, to reduce the interference; yet they suffer from huge computational cost and low space separability. In this work, we propose a lightweight and effective local intrinsic dimension (LID)-based method NAS-LID. NAS-LID evaluates the geometrical properties of architectures by calculating the low-cost LID features layer-by-layer, and the similarity characterized by LID enjoys better separability compared with gradients, which thus effectively reduces the interference among subnets. Extensive experiments on NASBench-201 indicate that NAS-LID achieves superior performance with better efficiency. Specifically, compared to the gradient-driven method, NAS-LID can save up to 86% of GPU memory overhead when searching on NASBench-201. We also demonstrate the effectiveness of NAS-LID on ProxylessNAS and OFA spaces. Source code: https://github.com/marsggbo/NAS-LID.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Xin He (135 papers)
  2. Jiangchao Yao (74 papers)
  3. Yuxin Wang (132 papers)
  4. Zhenheng Tang (38 papers)
  5. Ka Chu Cheung (1 paper)
  6. Simon See (74 papers)
  7. Bo Han (282 papers)
  8. Xiaowen Chu (108 papers)
Citations (9)

Summary

We haven't generated a summary for this paper yet.