Papers
Topics
Authors
Recent
Search
2000 character limit reached

Graph Neural Architecture Search

Updated 23 April 2026
  • Graph Neural Architecture Search is a method that encodes neural networks as graphs to explore vast, expressive design spaces with efficiency.
  • It employs advanced techniques like graph generative models, variational autoencoders, and diffusion models to generate and optimize architectures.
  • The approach utilizes surrogate predictors and multiple optimization algorithms, such as reinforcement learning and evolutionary methods, to balance accuracy, latency, and robustness.

Graph Neural Architecture Search (NAS) automates the construction of neural architectures by directly encoding candidate networks as graphs and leveraging graph-based learning, optimization, and sampling strategies to efficiently navigate large, expressive, and topologically rich architecture spaces. Graph NAS encompasses both general-purpose neural network search framed in graph terms and specialized Graph Neural Network (GNN) architecture search.

1. Graph-based Encoding of Neural Architectures

Modern NAS formalizes neural architectures—both general and GNNs—as directed acyclic graphs (DAGs). Each node typically specifies an operation or layer type (e.g., convolution, activation, pooling, or skip), and edges indicate data flow. For cell-based networks (e.g., NAS for CNNs), a cell is itself a computational DAG, which, when stacked or composed, yields full architectures (Asthana et al., 2024, Li et al., 2022). Node feature embeddings and edge presence (in {0,1}n×n\{0,1\}^{n \times n} for nn nodes) fully define the space of possible architectures.

Graph encoding enables:

  • Embedding of architecture topology for learned performance prediction, as in GNN predictors and variational autoencoders (Zhang et al., 2024, Lukasik et al., 2020).
  • Graph generative modeling to sample new architectures, supporting distributional, multi-modal, and auto-regressive generation (Li et al., 2022).
  • Formulation of architecture search as global graph optimization, embedding reachability and graph kernels within acquisition function optimization (Xie et al., 29 May 2025).
  • Expressive search spaces allowing loops, dynamic termination, and branching—capturing the iterative, recursive, or conditional decision structures arising in practical architecture design (JastrzÄ™bski et al., 2018).

2. Search Spaces, Optimization, and Graph Innovations

Graph NAS differentiates itself through unified and highly expressive search spaces. In cell-based settings, the search space may encompass all DAGs up to NN nodes with labeled nodes/edges; for GNNs, the space often factors into choices of propagation (P) and transformation (T) operations, along with aggregators, attention mechanisms, layer combinations, skip strategies, pooling, and further architectural motifs (Wang et al., 30 Apr 2025, Zhang et al., 2022).

Notable advances include:

  • Disaggregation of P and T, with architectures represented as arbitrary-length sequences over a mixed P/T alphabet, supporting flexible, data-adaptive depth patterns (Wang et al., 30 Apr 2025, Zhang et al., 2022).
  • Mask operations and robustness-driven primitives for adversarially robust GNN NAS, leveraging graph-structure masking and explicit robustness metrics (Xie et al., 2023).
  • Compact macro-level search spaces constrained by practical design priors, supporting exhaustive tabular benchmarking (e.g., NAS-Bench-Graph with 26k unique GNN architectures) (Qin et al., 2022).
  • Hierarchical, branching, and dynamic search simulations, wherein the architecture search space itself is a graph, not a flat token sequence—leading to greater sample efficiency in RL-controlled NAS (JastrzÄ™bski et al., 2018).

3. Graph Generative Models and Diffusion Approaches

GraphNAS has moved beyond handcrafted or sequential sampling to deep generative modeling of architecture spaces. This includes:

  • Deep generative graph models that capture correlation among topological features, sampling new architectures via auto-regressive models parameterized by GNN-based encoders (Li et al., 2022).
  • Graph diffusion models, with neural architecture generation formalized as a discrete, conditional diffusion process—modeling both network structure and constraint satisfaction (e.g., latency, accuracy) in a single differentiable denoising network (Asthana et al., 2024).
  • Variational autoencoders (VAE) operating in latent space, enabling smooth interpolation and extrapolation across graph-encoded architectures and facilitating performance predictor-guided, gradient-based search (Lukasik et al., 2020).
  • Multi-modal and distributional search, yielding not only a point estimate for the best architecture but a full density over sets of high-performing candidates for ensembling or uncertainty quantification (Li et al., 2022).

Graph generative approaches enable high sample efficiency, direct constraint satisfaction, and the ability to search for architectures optimizing multiple objectives (e.g., accuracy and latency), as demonstrated in large-scale empirical evaluations (Asthana et al., 2024).

4. Learning-based Surrogate Models and Predictors

Graph-based performance predictors accelerate NAS by providing fast, surrogate-based scoring of candidate architectures without the need for full training:

  • GNN-based predictors (e.g., using Graph Isomorphism Networks) input the architecture DAG and output performance estimates; these predictors are critical for search methods such as Bayesian/NAS-Bayesian optimization (Zhang et al., 2024).
  • Dual-view predictors exploit both the original and reversed graph structure. Forward-and-Reverse (FR-NAS) methods align latent embeddings across both directions, yielding strong performance in low-data regimes and improving rank correlation with real validation accuracy by significant margins (Zhang et al., 2024).
  • Performance-predictive embedding spaces, such as those learned by smooth variational graph autoencoders (SVGe), support both regression and downstream search via Bayesian optimization (Lukasik et al., 2020).

Empirical benchmarks consistently show that GNN-embeddings for surrogate prediction outperform sequence-based or handcrafted feature predictors on standard NAS graph benchmarks (NAS-Bench-101, NAS-Bench-201) (Zhang et al., 2024, Lukasik et al., 2020).

5. Optimization Algorithms: RL, Evolution, Differentiable, and Global Methods

Graph NAS embraces a spectrum of optimization, sampling, and search algorithms:

  • Reinforcement learning controllers (often LSTM-based) sample action sequences that unfold to complete graph-encoded architectures, trained with REINFORCE to maximize observed architecture validation performance (Gao et al., 2019, Zhao et al., 2020).
  • Evolutionary and genetic algorithms operate on graph-encoded individuals, supporting mutation, crossover, and adaptive exploration-exploitation scheduling (AGOS) (Wang et al., 30 Apr 2025, Zhang et al., 2022).
  • Differentiable NAS leverages soft or probabilistic relaxations of discrete search spaces, employing Gumbel-softmax or continuous mixture weights for operators (e.g., Probabilistic Dual NAS, SANE) (Zhao et al., 2020, Zhao et al., 2021).
  • Mixed-integer global optimization: NAS-GOAT encodes the entire space of feasible cell-DAGs as a MIP, embeds structured GP kernels for surrogate modeling, and exactly solves global acquisition function optimization for batch BO (Xie et al., 29 May 2025).
  • Training-free and sparse-coding approaches sidestep weight optimization, exploiting linear GNN theory to directly optimize architecture selection through â„“1\ell_1-coded mixtures, delivering significant gains in computational efficiency (Xu et al., 2023).

Algorithm selection depends on the search space size, cost of candidate evaluation, and performance constraint structure. Integrated approaches (e.g., evolutionary search with periodic Bayesian hyperparameter refinement) further improve sample efficiency and adaptivity (Wang et al., 30 Apr 2025).

6. Empirical Benchmarks, Datasets, and Evaluations

Graph NAS methods have been rigorously benchmarked on a variety of well-established datasets and tabular search spaces. Key resources include:

  • NAS-Bench-101, NAS-Bench-201, NAS-Bench-301: cell-based image network benchmarks with exhaustive or surrogate-based performance tables for hundreds of thousands of architectures (Asthana et al., 2024, Li et al., 2022).
  • HW-NAS-Bench: hardware-constrained evaluation (e.g., latency on edge devices under various budgets) (Asthana et al., 2024).
  • NAS-Bench-Graph: macro-level GNN architecture benchmarking with 26,206 unique architectures, covering diverse graph datasets (Cora, PubMed, ogbn-arxiv, Amazon-Photo, etc.) and logging per-epoch performance and efficiency metrics (Qin et al., 2022).
  • Standard citation and protein-protein interaction datasets, with both transductive and inductive settings—supporting fair evaluation and transferability analysis (Gao et al., 2019, Wang et al., 30 Apr 2025, Qin et al., 2022).

Empirical findings underscore:

  • High validation-test accuracy, uniqueness, and architecture novelty from graph generative and graph-diffusion NAS methods, achieving >90%>90\% uniqueness across millions of generations (Asthana et al., 2024).
  • State-of-the-art macro-F1 and accuracy with advanced search strategies (e.g., ABG-NAS, DFG-NAS) compared to both manual GNNs and prior NAS strategies (Wang et al., 30 Apr 2025, Zhang et al., 2022).
  • Empirical recommendations: devote effort to deeper-layer decisions, jointly optimize efficiency/accuracy, and tailor the operator set per dataset class (Qin et al., 2022).

7. Practical Insights, Limitations, and Future Directions

Current evidence shows graph-based neural architecture search methods deliver improved sample efficiency, heightened expressiveness, and state-of-the-art downstream accuracy. Key principles and observations include:

  • Search spaces representing architectures as graphs (not linear sequences) capture design motifs such as loops, branching, and early termination far more succinctly, with demonstrated gains in sample efficiency and convergence (JastrzÄ™bski et al., 2018).
  • Deep graph generative models and graph diffusion-based sampling yield diverse, performant, and constraint-aware architectures with only a single trained generative model (Asthana et al., 2024, Li et al., 2022).
  • Surrogate and predictor models based on GNNs outperform non-graph-based surrogates for performance prediction and zero-cost search guidance (Zhang et al., 2024, Lukasik et al., 2020).
  • Training-free linear coding methods, leveraging initialization theory, match or exceed accuracy while reducing search wall time by several orders of magnitude (Xu et al., 2023).
  • Multi-objective, hardware-aware, and adversarially robust search spaces have been formalized, allowing simultaneous optimization of accuracy, latency, and robustness via graph-based search (Asthana et al., 2024, Xie et al., 2023).

Limitations remain: computational cost in large architecture spaces, tractability with increased operation heterogeneity, robust transfer across datasets, and handling of irregular graph structures. Future work may focus on hybrid one-shot, weight-sharing, and meta-learning strategies, expansion to more complex search spaces (heterogeneous graphs, multi-scale pooling), and integration of rich downstream task constraints (e.g., explainability, privacy, energy) (Wang et al., 30 Apr 2025, Zhang et al., 2022, Xie et al., 29 May 2025). Dataset-specific search remains necessary due to observed low cross-dataset transferability of optimal architectures (Qin et al., 2022).

In summary, graph neural architecture search provides a general and principled foundation for efficient exploration, generation, and evaluation of rich architecture spaces for both general neural networks and GNNs, with continued advances in learning methods, surrogate modeling, and global optimization driving the field forward.

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Graph Neural Architecture Search (NAS).