Papers
Topics
Authors
Recent
2000 character limit reached

Energy-Aware Model Selection

Updated 7 December 2025
  • Energy-aware model selection is a framework that jointly optimizes predictive performance and energy consumption across AI models.
  • It employs optimization techniques such as Pareto-optimality, reinforcement learning, and meta-heuristics to balance utility with energy efficiency.
  • Empirical benchmarks demonstrate significant energy savings in real-world deployments, promoting sustainable and cost-effective AI solutions.

Energy-aware model selection is a formal approach to choosing among multiple candidate AI models, architectures, or configurations by optimizing explicitly for both predictive utility (such as accuracy, F1, or task-specific metrics) and energy consumption (variously measured in watts, joules, or proxies such as CPU-seconds or MACs). This framework is increasingly critical as the energy costs of machine learning—across datacenter, edge, and federated environments—scale with model size, deployment volume, and workload heterogeneity. Approaches span ensemble pruning, adaptive dynamic routing, federated optimization, hardware-specific compression, and online SLA-constrained selection, unified by explicit energy-performance trade-off formulations and quantitative, often empirical, measurement or modeling of energy usage.

1. Formal Criteria and Core Metrics

Energy-aware model selection defines objectives and constraints linking utility, U(m)U(m) (task-specific outcome, e.g., accuracy) and energy, E(m)E(m) (per-inference or cumulative). Typical formulations include:

  • Constraint form:

minmM  E(m)s.t.  U(m)(1Δ)U(mmax)\min_{m\in\mathcal{M}}\; E(m)\quad\text{s.t.}\; U(m)\geq(1-\Delta)\cdot U(m_{\text{max}})

where Δ\Delta is an allowed utility drop (e.g., 5%) (Barros et al., 2 Oct 2025).

  • Weighted (scalarized) trade-off:

maxm  [U(m)λE(m)]\max_{m}\;[U(m)-\lambda\,E(m)]

or, equivalently, score-based ranking functions:

R(m)=wφ(m)+(1w)U(m)R(m) = w\cdot\varphi(m)+(1-w)\cdot U(m)

with ww reflecting energy-performance prioritization (Tran et al., 22 Mar 2025).

Energy measurement methodologies range from direct hardware-instrumented joule readings (e.g., via CodeCarbon, NVML, Turbostat) to model-specific proxies like MACs, CPU-seconds, or, for edge CPUs, ISA-level event count-based regressions (Georgiou et al., 2021). Accuracy is always computed on meaningful task metrics (F1@k, BLEU, mAP, ROUGE-L, etc.).

2. Static, Dynamic, and Instance-Adaptive Selection

There are three broad classes of selection strategies:

  1. Static selection: A single optimal subset or configuration is used for all inputs and/or properties, maximizing overall utility or utility–energy trade-off on the validation set (Nijkamp et al., 2024).
  2. Dynamic (property-based) selection: The optimal subset/configuration is chosen per property, task, or client, i.e., for a set PP of properties, select S(pj)=argmaxSF1(S,pj)S^*(p_j) = \arg\max_S F1(S, p_j) (Nijkamp et al., 2024).
  3. Instance-adaptive selection: The specific model or ensemble subset is dynamically chosen per input instance, exploiting confidence scores or learned routing. “Green AI dynamic cascading” and routing techniques operationalize this by cascading through low-energy models, only escalating to higher-energy ones if confidence is low; or by explicitly routing each instance to the most likely-efficient model (Cruciani et al., 24 Sep 2025).

Instance-adaptive methods often yield the largest energy savings, particularly in workloads with highly variable input difficulty.

3. Algorithmic Approaches and Optimization Techniques

A variety of search and optimization strategies underpin energy-aware selection.

  • Exhaustive subset enumeration: For small ensembles (M<15|\mathcal{M}| < 15), all non-empty subsets are evaluated on validation data for both utility and energy (Nijkamp et al., 2024).
  • Meta-heuristics / Multi-objective optimization: When the configuration space is large, NSGA-II or similar samplers form Pareto fronts from sampled configurations, which are then ranked or further analyzed using techniques such as weighted gray relational analysis (Tundo et al., 2023).
  • Multi-armed bandits: Bandit-based algorithms allocate limited evaluation budget adaptively, focusing sampling on promising models given observed reward (composite of normalized accuracy and energy proxies) (Kannan, 2024).
  • Reinforcement learning: RL-based controllers, including deep Q-networks and MARL (QMIX for federated dual-selection), optimize over sequential and multi-agent energy-budgeted decision processes, with reward explicitly trading off global utility gain and energy use (Xia et al., 2024, Bullo et al., 2024).
  • Resource-efficient pruning/quantization: Joint quantization and structured/unstructured pruning, when guided by explicit energy–accuracy tradeoff scoring, allow production of model variants at varying energy and performance points, enabling post-hoc selection (Tran et al., 22 Mar 2025, Wang et al., 2020, Fang et al., 21 Nov 2025).
  • Online stochastic control with SLA guarantees: Algorithms such as MESS+ maintain a “virtual queue” tracking SLA-defined accuracy debt, and apply Lyapunov drift-plus-penalty scheduling to guarantee average performance under per-inference request energy minimization (Zhang et al., 2024).

4. Empirical Benchmarks and Quantitative Outcomes

Empirical evaluation across deployment paradigms demonstrates the efficacy of energy-aware selection:

  • Ensemble pruning (CUAD/Lease, DOCQMiner): Moving from full ensembles (100% energy, F1=0.15–0.22) to static or dynamic selection reduces energy to 60–76% and boosts F1 by 0.40–0.43; energy-aware variants cut further to 14–57% with only minimal F1 loss (Nijkamp et al., 2024).
  • Dynamic routing/cascading: On digit classification, dynamic cascading yields ≈95% of the most accurate model’s accuracy while reducing energy and latency by 20–25% (Cruciani et al., 24 Sep 2025).
  • SLAs in model zoos: Per-request SLA-driven selection achieves a 3.5× (translation) and 4.6× (summarization) energy reduction compared to always using the largest LLM, with no SLA violation (Zhang et al., 2024).
  • Global impact: Systematic model selection (“small is sufficient,” m_eff per task) could deliver a 27.8% reduction in global AI inference energy, yielding tens of TWh annual savings—equivalent to multiple large power reactors (Barros et al., 2 Oct 2025).
  • Hardware-adaptive compression: RL-driven joint quantization/pruning and dataflow selection on CNNs deliver 11–37× energy reductions for VGG-16, MobileNet, LeNet with ≤2% loss in accuracy (Wang et al., 2020). Layer-wise energy-prioritized quantization and weight selection further improve per-layer efficiency, with up to 58.6% reduction and ≤3% accuracy drop (Fang et al., 21 Nov 2025).
  • Federated FL client/model dual-selection: Dual-selection in non-IID battery-constrained FL settings yields higher global accuracy, longer system lifetimes, and more balanced energy use compared to static or greedy baselines (Xia et al., 2024).

5. Deployment Contexts and Application Scenarios

Energy-aware model selection is validated across several production and research domains:

  • Enterprise information extraction: DOCQMiner at Deloitte NL operationalizes static/dynamic selection and GreenQuotientIndex-driven trade-offs (Nijkamp et al., 2024).
  • Federated AIoT: Dual model/client selection via MARL addresses device heterogeneity and battery constraints, with layered model slicing (Xia et al., 2024).
  • Edge and energy-harvesting devices: Markov decision process controllers and incremental early exit policies enable robust operation under stochastic energy supply (Bullo et al., 2024).
  • Self-adaptive edge applications: NSGA-II and FSM-driven configuration selection achieves up to 81% energy savings with only 2–6% accuracy penalties in live pedestrian detection (Tundo et al., 2023).
  • Automated model zoo selection: GUIDE and GREEN frameworks automate Pareto-optimal selection in diverse task types, leveraging empirical databases and sub-second selection times (Smirnova et al., 30 Nov 2025, Betello et al., 2 May 2025).

6. Methodological Considerations and Extensions

Energy-aware model selection presupposes the availability of accurate energy measurement or estimation models, either hardware-instrumented or via proxies (MACs, CPU-seconds). For embedded/edge CPUs, static-event-count regressions can deliver <5% error per configuration (Georgiou et al., 2021). For accelerator-bound CNNs, per-layer switching-activity-based power models improve prioritization of compression effort (Fang et al., 21 Nov 2025).

Extensions include:

7. Limitations and Future Prospects

Documented limitations include overheads of meta-heuristic search in large configuration spaces, reliance on accurate energy and utility proxies, and the generalization of selection mechanisms to emerging architectures (e.g., LLM variants, MoEs, novel hardware) or tasks lacking clear confidence or utility surrogates (Smirnova et al., 30 Nov 2025, Barros et al., 2 Oct 2025, Betello et al., 2 May 2025). Strategies such as hierarchical learning, cluster-based partitioning, and continual/zero-shot learning for candidate models are actively considered to increase scalability and coverage (Xia et al., 2024, Betello et al., 2 May 2025).

Editor's term: “energy-aware model selection” thus denotes the class of AI system optimization strategies that jointly consider model utility and resource consumption—primarily energy—using methods that scale from offline Pareto sweep to online per-instance and per-budget adaptation, forming a foundation for sustainable and “Green AI” practice at-scale.

Whiteboard

Topic to Video (Beta)

Follow Topic

Get notified by email when new papers are published related to Energy-Aware Model Selection.