LRA-CMA: Adaptive Learning for CMA-ES
- LRA-CMA is a variant of CMA-ES that dynamically adapts learning rates via a signal-to-noise ratio criterion to enhance optimization efficiency.
- It employs online exponential moving averages of parameter updates to balance progress and stability in noisy and rugged search landscapes.
- Empirical evaluations demonstrate that LRA-CMA outperforms fixed-rate CMA-ES, achieving robust performance on challenging noisy and multimodal problems.
LRA-CMA, short for Learning Rate Adaptation for Covariance Matrix Adaptation Evolution Strategy, is a variant of CMA-ES that introduces a principled mechanism for online adaptation of learning rates based on a signal-to-noise ratio (SNR) criterion. Unlike standard CMA-ES, which uses fixed learning rates for its mean and covariance updates, LRA-CMA adaptively controls the magnitude of parameter updates to maintain a constant SNR in Fisher metric units. This design enables robust, tuning-free optimization on noisy and multimodal black-box functions while retaining the favorable practical characteristics of default CMA-ES (Nomura et al., 2023, Nomura et al., 29 Jan 2024).
1. Foundations of CMA-ES and the Role of the Learning Rate
CMA-ES operates by evolving a Gaussian search distribution, parameterized by mean and covariance , to minimize a black-box objective . At each iteration, candidate points are sampled and evaluated, with the top-ranked samples influencing the updates of and . In classical CMA-ES, these updates are controlled by fixed learning rates (). The learning rates directly affect the step sizes in natural gradient space, with large rates promoting rapid but potentially unstable adaptation, and small rates ensuring stability but risking inefficiency, especially in the presence of noise or rugged landscapes.
ODE-based analysis shows that the ideal learning rate must be sufficiently small to allow the stochastic discrete updates of CMA-ES to follow the path of the continuous-time natural gradient flow. On functions such as Rastrigin, the ODE for the mean and variance highlights that only for small enough learning rates do the trajectories reach the global optimum, avoiding divergence or stagnation (Nomura et al., 29 Jan 2024). This sensitivity to learning rate motivates the need for automatic adaptation.
2. Signal-to-Noise Ratio Criterion for Learning Rate Adaptation
LRA-CMA is built on the premise that each update to the parameters should maintain a prescribed ratio between the expected directional progress (signal) and the inherent stochastic fluctuation (noise). The SNR in Fisher-natural coordinates is defined as
where denotes the update measured in the local Fisher metric, with being the Fisher information matrix (, ).
To estimate the SNR online, LRA-CMA maintains exponential moving averages and of the updates and their squared norms: yielding the estimator
(Nomura et al., 2023, Nomura et al., 29 Jan 2024).
3. Learning Rate Adaptation Mechanism
The core mechanism in LRA-CMA is a multiplicative adaptation rule that adjusts the effective local learning rates , to enforce
for a target . At each generation,
where and denotes clipping to . The learning rate is adapted down in noisy or rugged phases and up in smooth phases, balancing robustness and efficiency.
Once learning rates are adapted, the proposed CMA-ES updates for and are rescaled by and : After the updates, the step size is corrected to maintain the optimal scaling : (Nomura et al., 2023, Nomura et al., 29 Jan 2024).
4. Algorithmic Structure and Pseudocode
The LRA-CMA-ES algorithm is best viewed as a modular extension of vanilla CMA-ES, with the SNR-based adaptation intervening just prior to parameter updates. A per-generation loop proceeds as:
| Stage | Main Operation | Intervention Point |
|---|---|---|
| Sample and evaluate | Standard CMA-ES sampling and ranking | — |
| Propose updates | Standard vanilla update formulas | — |
| Fisher metric transformation | Convert to local (unit-Fisher) coords | LRA-CMA mechanism |
| Exponential averaging | Update , for , | LRA-CMA mechanism |
| SNR estimation | Compute | LRA-CMA mechanism |
| Learning rate update | Update , | LRA-CMA mechanism |
| Rescale and apply update | Apply scaled to , | LRA-CMA mechanism |
| Step size correction | Update using ratio | LRA-CMA mechanism |
This staged structure ensures that LRA-CMA is a drop-in replacement for learning rate control in existing CMA-ES implementations (Nomura et al., 29 Jan 2024).
5. Empirical Evaluation and Performance Characteristics
Experiments on standard benchmarks (Sphere, Ellipsoid, Rosenbrock, Rastrigin, Schaffer, and noisy variants) in dimensions demonstrate that LRA-CMA achieves both problem-adaptive robustness and speed. On smooth unimodal problems, LRA-CMA matches the performance of CMA-ES with optimally fixed learning rates (SP1 for Sphere is , success). On multimodal or noisy functions, LRA-CMA obtains high success rates (e.g., Rastrigin: ; SP1 ) even when fixed-rate CMA-ES is either unstable (for large ) or inefficient (for small ). Success is defined as reaching within a – evaluation budget.
Notably, in strongly noisy scenarios (additive Gaussian noise, or ), fixed-rate CMA-ES often stalls, while LRA-CMA maintains steady progress by adaptively shrinking learning rates. LRA-CMA thereby achieves success down to . Population-size adaptation (PSA-CMA-ES) is competitive on noiseless multimodal functions but is outperformed by LRA-CMA under strong noise (Nomura et al., 2023, Nomura et al., 29 Jan 2024).
6. Practical Guidelines for Implementation
Recommended default hyperparameters for LRA-CMA-ES are (target SNR), , (exponential averaging), and (damping), with the learning rate factors initialized to one. If the optimization landscape is particularly noisy or multimodal, smaller , values are recommended for increased stability; for smooth unimodal problems, larger values yield faster adaptation.
LRA-CMA-ES permits retention of the default population size , avoiding the need for expensive population size adaptation or restarts. It integrates directly into any CMA-ES variant, including diagonal and separable schemes, by replacing parameter update steps with LRA-controlled updates. The only further constraint is to maintain , and to adjust in response to changes in (Nomura et al., 2023, Nomura et al., 29 Jan 2024).
7. Relation to Broader Optimization Methodology
LRA-CMA-ES can be formally interpreted as an instance of controlling the magnitude of natural gradient updates relative to their statistical estimation variance, functioning as an automatic mechanism for balancing progress versus robustness in stochastic search. Enforcing constant SNR ensures that neither drift (noise-dominated updates) nor slowness (excessively conservative updates) prevails over extended search horizons. This reflects a general pattern in ES design where stability and adaptivity are prioritized over tuning specific parameter schedules for each problem class. The SNR-based mechanism in LRA-CMA-ES offers a general principle extensible to other stochastic natural-gradient methods beyond ES.
Key references:
- "CMA-ES with Learning Rate Adaptation: Can CMA-ES with Default Population Size Solve Multimodal and Noisy Problems?" (Nomura et al., 2023)
- "CMA-ES with Learning Rate Adaptation" (Nomura et al., 29 Jan 2024)