Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
126 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Survey on Evolutionary Neural Architecture Search (2008.10937v4)

Published 25 Aug 2020 in cs.NE

Abstract: Deep Neural Networks (DNNs) have achieved great success in many applications. The architectures of DNNs play a crucial role in their performance, which is usually manually designed with rich expertise. However, such a design process is labour intensive because of the trial-and-error process, and also not easy to realize due to the rare expertise in practice. Neural Architecture Search (NAS) is a type of technology that can design the architectures automatically. Among different methods to realize NAS, Evolutionary Computation (EC) methods have recently gained much attention and success. Unfortunately, there has not yet been a comprehensive summary of the EC-based NAS algorithms. This paper reviews over 200 papers of most recent EC-based NAS methods in light of the core components, to systematically discuss their design principles as well as justifications on the design. Furthermore, current challenges and issues are also discussed to identify future research in this emerging field.

Citations (353)

Summary

  • The paper presents a comprehensive review of over 200 research articles to identify ENAS core components and design principles.
  • It examines various encoding spaces—layer, block, and cell—and evolutionary algorithms like GA, GP, and PSO to automate DNN architecture design.
  • The paper also evaluates computational challenges and proposes strategies such as weight inheritance and early stopping to boost efficiency.

A Survey on Evolutionary Neural Architecture Search

The surveyed paper presents a thorough examination of Evolutionary Neural Architecture Search (ENAS), focusing on the applications and analytical methods that allow the automatic design of Deep Neural Networks (DNNs) architectures. The paper entailed a comprehensive review of over 200 research articles, encompassing disparate methods of ENAS with the aim of identifying core components, design principles, and the future trajectory of this field.

Key Insights and Components of ENAS

ENAS emerges as a promising avenue due to its capability to automate the architectural design of DNNs, traditionally a manual task requiring substantial time and expertise. The paper elaborates on various encoding spaces within ENAS—layer-based, block-based, and cell-based encoding spaces—outlining how they contribute to the potential architecture configurations of DNNs. It also details the challenges of these encoding spaces, such as the risk of overly-constraining designs and the heavy reliance on expert knowledge for effective initialization and search spaces.

A significant portion of the survey is dedicated to discussing the evolutionary process in ENAS, emphasizing evolutionary algorithms (EAs) such as Genetic Algorithms (GAs), Genetic Programming (GP), and Particle Swarm Optimization (PSO). These algorithms are pivotal in generating new neural architectures and are detailed in their strategies for population initialization, fitness evaluation, and selection processes. The paper methodically examines both single-objective and multi-objective optimization strategies. One key observation here is the predominance of GA-based ENAS methods, acknowledged for their ease of representing architecture and integration with standard evolutionary operators.

Challenges and Computational Efficiency

The paper candidly acknowledges the computational overhead associated with ENAS, as many methods demand extensive resources for architecture evaluation. Strategies such as weight inheritance, early stopping policies, and the use of reduced datasets and populations are evaluated for their efficacy in reducing computational costs. The survey brings to light the unique challenge of balancing evaluation accuracy with computational efficiency—a crucial consideration for the scalability of ENAS algorithms on large datasets like ImageNet.

Applications and Benchmarking

ENAS has exhibited substantial applications, primarily in image and signal processing domains, including image and speech recognition. The architecture's effectiveness has been proven largely in image classification tasks, with ENAS methodologies like LargeEvo and NSGANet demonstrating promising results on benchmark datasets such as CIFAR-10 and ImageNet. The paper also stresses the need for fair benchmarking standards in ENAS to provide consistent and reliable comparisons among various methodologies.

Theoretical and Practical Implications

The theoretical implications of ENAS primarily relate to its potential to democratize the design of neural network architectures by reducing the reliance on domain-specific expertise. Practically, the potential of ENAS is vast—encompassing any area where DNNs might be applied—implying an expansive horizon for practical implementations.

Future Directions

The paper outlines several challenges that future research must address, including enhancing the interpretability of discovered architectures, improving computational efficiency, and setting up fair comparisons across studies. Notably, the discussion suggests establishing standardized benchmarking datasets for NAS methods, offering a cohesive platform for evaluating the scalability of ENAS solutions.

In conclusion, the paper provides a comprehensive survey of ENAS, outlining the existing accomplishments and delineating areas necessitating further exploration. It positions ENAS as a robust mechanical framework poised to contribute significantly to the evolution of autonomous neural architecture design. This survey serves as a vital resource for researchers seeking to understand the landscape of ENAS and explore its untapped potential in AI advancements.