Papers
Topics
Authors
Recent
Search
2000 character limit reached

NAS-Based Pruning Techniques

Updated 24 February 2026
  • NAS-Based pruning is a technique that unifies neural architecture search with pruning to remove redundant weights during early training stages.
  • It leverages methods such as one-shot supernet pruning, dynamic distribution, and differentiable optimization to meet strict FLOPs and latency constraints.
  • Empirical results show that NAS-based pruning consistently outperforms traditional post-training methods, offering superior accuracy-resource trade-offs.

Pruning Before Training (PBT) refers to methods that exploit the hypothesis of over-parameterization in neural architectures by explicitly removing redundant or suboptimal weights, blocks, or operations early in the optimization pipeline—either before or during the initial training stages. In contrast to classical post-training pruning, which targets pretrained networks with magnitude or sensitivity-based criteria, PBT and NAS-pruning methods formally embed pruning as an integral part of the architecture search or as a zero-/one-shot optimization over network structure. This approach underlies a substantial family of techniques in efficient deep learning, from structured pruning with NAS integration, to differentiable pruning and joint architecture-sparsity search in both convolutional and transformer-based models.

1. Formal Problem Definitions

PBT formulations unify network pruning and neural architecture search (NAS) by elevating the pruning policy (e.g., per-layer mask or operation selection) to part of the discovery or optimization problem. The canonical formalism for channel/network pruning in the NAS regime is:

minm,θLtrain(θm) subject toFLOPs(θm)B, m0Smax\begin{align*} \min_{m,\, \theta}\quad &L_{\text{train}}(\theta \odot m) \ \text{subject to}\quad &\mathrm{FLOPs}(\theta \odot m) \leq B,\ &\| m \|_0 \leq S_{\max} \end{align*}

where θ\theta are model weights, mm is a binary mask over parameters or operations, LtrainL_{\text{train}} is the training loss, BB the target FLOPs budget, and SmaxS_{\max} an upper bound on retained parameters (Ghosh et al., 2023). This can be recast as an architecture selection or bi-level optimization problem, in which the mask is derived by minimizing validation loss post-fine-tuning:

m=argminm:FLOPs(θm)BLval(FineTune(θm))m^* = \arg\min_{m:\, \mathrm{FLOPs}(\theta \odot m) \leq B} L_{\text{val}}(\mathrm{FineTune}(\theta \odot m))

Generalizations extend mm to encode discrete operation choices, layer widths/depths, or structural motifs, with resource constraints on FLOPs, latency, or hardware-specific cost (Lin et al., 2021, Li et al., 2022, Dai et al., 2020). In transformer-based and LLM models, PBT encompasses binary masks over heads, neurons, or entire layers, and can include MLP widths or attention head counts as architectural degrees of freedom (Klein et al., 2024, Sarah et al., 2024, Malettira et al., 2 Feb 2026).

2. Core Methodologies and Algorithms

PBT/NAS-pruning methods comprise a diverse set of search and optimization frameworks:

  • One-shot and supernet-based pruning: Construct a weight-sharing supernet encoding the full pruning/search space, and optimize for subnetwork weights across pruned configurations. Post-training, the architectural subspace is explored (e.g., via evolutionary, Bayesian, or multi-objective optimization) to extract Pareto-optimal trade-offs (Sarah et al., 2024, Klein et al., 2024, Abebe et al., 15 Jan 2025).
  • Dynamic distribution pruning: Represent the candidate architecture set by joint categorical distributions across operations/edges, iteratively prune low-probability operations, and update the selection distributions via utility-driven momentum steps (Zheng et al., 2019). Pruning decisions are made at discrete rounds, with theoretical error bounds on mistaken exclusions.
  • Magnitude-based and structural pruning within NAS: Prune weights or architectural elements by global or layerwise criteria (e.g., magnitude, L1-norm, BN scale factor) applied to pretrained or supernet weights, often coupling the pruning to subsequent fine-tuning (Ghosh et al., 2023). Mask design can follow global or uniform per-layer sparsity constraints and be tuned to optimize subsequent fine-tune performance.
  • Differentiable and continuous pruning: Parameterize channel/layer sparsity via continuous latent variables (applying L1-regularization or soft thresholds), enabling joint optimization of architecture and weights through either standard gradients or proximal methods (Li et al., 2020, Li et al., 2022).
  • Reinforcement learning and evolutionary search: Model pruning/morphing choices as actions in a Markov decision process, with reward signals based on accuracy-resource trade-offs observed on proxy data or downstream hardware (Li et al., 2020, Laube et al., 2019). Improved differential evolution (IDE) and NSGA-II are frequent choices for discrete or multi-objective settings (Lin et al., 2021, Sarah et al., 2024).

A frequently adopted practical scheme is:

1
2
3
4
5
6
Input: pretrained model θ, dataset D, target FLOPs B
1. Compute weight scores (e.g., |θ_i|) for each parameter or block
2. For each layer or globally, select thresholds to meet global or per-layer sparsity targets
3. Apply mask m (set θ_i=0 where score < threshold)
4. Fine-tune θ ⊙ m on D until val-loss convergence
5. Evaluate and export pruned architecture
(Ghosh et al., 2023).

3. Search Spaces, Constraints, and Optimization Criteria

The definition and management of the architectural search space is central to PBT. Search spaces may comprise:

Constraint handling is often explicit (via rescaling population candidates to satisfy FLOPs/model size budgets (Lin et al., 2021)), Lagrangian (resource penalty terms in the objective (Li et al., 2022)), or via hardware-in-the-loop code generation with measured latency as the constraint (Li et al., 2020). Resource metrics include FLOPs, parameter count, latency (CPU/GPU/edge), sometimes measured via compiler-optimized code (Li et al., 2020).

Optimality is framed as Pareto front extraction with multi-objective trade-off between accuracy and resource consumption—typically using NSGA-II or hypervolume-improvement surrogates (Sarah et al., 2024, Klein et al., 2024). In many cases, the final subnet is chosen by maximizing validation accuracy at a fixed cost, or by scanning the non-dominated set.

4. Practical Algorithms and Empirical Performance

The empirical pipeline for PBT typically consists of:

  • Pretraining a supernet or full model;
  • Applying iterative or one-shot pruning to generate a batch of candidate subnetworks under sparsity/resource constraints;
  • Partial or full fine-tuning of these subnetworks (or all weight-shared subnetworks in case of supernets with "sandwich rule" (Abebe et al., 15 Jan 2025));
  • Ranking sub-networks on validation performance and resource footprint (e.g., actual measured latency via code generation and device profiling (Li et al., 2020));
  • Extracting the final network via mode-selection or optimal point on the Pareto frontier.

PBT approaches consistently produce models that outperform traditional post hoc pruning at equal compute, often yielding strictly higher accuracy–FLOPs trade-offs and substantial reductions in search cost:

Method Pruning Target Search/Fine-tune cost Δ Accuracy @ equal FLOPs Key Reference
Prune+Fine-tune FBNetV3 (ImageNet) 3–5× cheaper than full NAS +0.5–1.5% vs. NAS (Ghosh et al., 2023)
DDPNAS dynamic pruning MobileNet (ImageNet) 1.8 GPU-h (14× faster) 77.2% top-1 (SOTA under mobile) (Zheng et al., 2019)
AACP channel pruning ResNet50 (ImageNet) 42% FLOPs cut, little loss −0.18% vs. unpruned (Lin et al., 2021)
TraceNAS zero-shot LLaMA2-7B LLM 10× search cost reduction +5–9% over uniform pruning (Malettira et al., 2 Feb 2026)

These gains are robust across domains (classification, detection, segmentation, LLMs), architectures (ConvNets, Transformers, ViT), and deployment targets (mobile, server, edge).

5. Critical Analysis and Theoretical Considerations

PBT unifies traditional pruning with NAS by treating architecture and sparsity as search variables, not hand-designed heuristics. This enables:

  • Differential fine-grained control: per-layer/operation sparsity, structured pruning of blocks, heads, or layers;
  • Integration of hardware/compiler feedback: latency, energy, and memory profiles incorporated directly into the search/selection loop (Li et al., 2020);
  • Multi-objective discovery: explicit Pareto-front construction supporting deployment-informed trade-offs (Sarah et al., 2024, Klein et al., 2024);
  • Differentiable and meta-learning-based estimation: reduced search cost via proxy or differentiable predictors (gradient masks, accuracy estimators, Gumbel-softmax variables) (Li et al., 2020, Lin et al., 2021, Malettira et al., 2 Feb 2026).

Limitations include the potentially complex hyperparameter landscape (e.g., evolutionary search/generation rates), reliance on quality of accuracy predictors/cost proxies, and, in some methods, the assumption of sufficient weight-sharing and supernet generalization across pruned configurations. Most PBT variants do not directly model low-level hardware execution traces or non-differentiable deployment bottlenecks (e.g., kernel launch overhead), but exceptions include frameworks with tight compiler-in-the-loop latency measurement (Li et al., 2020).

6. Impact and Applications

PBT and NAS-pruning algorithms are widely used in:

The cumulative effect is to decouple resource-constrained deployment from the high cost of full-scale NAS, supporting mass deployment of efficient DNNs and LLMs with tractable tuning overheads and superior Pareto trade-offs.

7. Future Directions

Open research directions in PBT include:

  • Improved search and approximation methods for very high-dimensional or discrete search spaces (e.g., deep transformer pruning with long-range dependencies);
  • Incorporation of more realistic deployment metrics (energy, actual user-facing latency under varying loads);
  • Transferability and distillation across domains/tasks during pruning/NAS search, especially in multi-task or federated settings;
  • Joint quantization-pruning and low-rank compression within the same NAS-based search pipeline (Sarah et al., 2024, Abebe et al., 15 Jan 2025);
  • Theoretical analysis of convergence and generalization in differentiable pruning/architecture meta-optimization (Zheng et al., 2019, Li et al., 2022).

The PBT paradigm continues to generalize, with recent works demonstrating zero-shot, training-free identification of globally optimal subnetworks, surpassing the accuracy and speed trade-offs of classical magnitude- or sensitivity-based methods (Malettira et al., 2 Feb 2026).

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to NAS-Based Pruning.