Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Discretization-Aware Architecture Search (2007.03154v1)

Published 7 Jul 2020 in cs.CV, cs.LG, and cs.NE

Abstract: The search cost of neural architecture search (NAS) has been largely reduced by weight-sharing methods. These methods optimize a super-network with all possible edges and operations, and determine the optimal sub-network by discretization, \textit{i.e.}, pruning off weak candidates. The discretization process, performed on either operations or edges, incurs significant inaccuracy and thus the quality of the final architecture is not guaranteed. This paper presents discretization-aware architecture search (DA\textsuperscript{2}S), with the core idea being adding a loss term to push the super-network towards the configuration of desired topology, so that the accuracy loss brought by discretization is largely alleviated. Experiments on standard image classification benchmarks demonstrate the superiority of our approach, in particular, under imbalanced target network configurations that were not studied before.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Yunjie Tian (17 papers)
  2. Chang Liu (864 papers)
  3. Lingxi Xie (137 papers)
  4. Jianbin Jiao (51 papers)
  5. Qixiang Ye (110 papers)
Citations (28)

Summary

We haven't generated a summary for this paper yet.