CMA-ES with Learning Rate Adaptation (2401.15876v2)
Abstract: The covariance matrix adaptation evolution strategy (CMA-ES) is one of the most successful methods for solving continuous black-box optimization problems. A practically useful aspect of the CMA-ES is that it can be used without hyperparameter tuning. However, the hyperparameter settings still have a considerable impact on performance, especially for difficult tasks, such as solving multimodal or noisy problems. This study comprehensively explores the impact of learning rate on the CMA-ES performance and demonstrates the necessity of a small learning rate by considering ordinary differential equations. Thereafter, it discusses the setting of an ideal learning rate. Based on these discussions, we develop a novel learning rate adaptation mechanism for the CMA-ES that maintains a constant signal-to-noise ratio. Additionally, we investigate the behavior of the CMA-ES with the proposed learning rate adaptation mechanism through numerical experiments, and compare the results with those obtained for the CMA-ES with a fixed learning rate and with population size adaptation. The results show that the CMA-ES with the proposed learning rate adaptation works well for multimodal and/or noisy problems without extremely expensive learning rate tuning.
- Quality Gain Analysis of the Weighted Recombination Evolution Strategy on General Convex Quadratic Functions. Theoretical Computer Science 832 (2020), 42–67.
- An ODE Method to Prove the Geometric Convergence of Adaptive Stochastic Algorithms. Stochastic Processes and their Applications 145 (2022), 269–307.
- Youhei Akimoto and Nikolaus Hansen. 2020. Diagonal Acceleration for Covariance Matrix Adaptation Evolution Strategies. Evolutionary Computation 28, 3 (2020), 405–435.
- Bidirectional Relation between CMA Evolution Strategies and Natural Evolution Strategies. In International Conference on Parallel Problem Solving from Nature. 154–163.
- Shun-Ichi Amari and Scott C Douglas. 1998. Why Natural Gradient?. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Vol. 2. 1213–1216.
- Shun-ichi Amari and Hiroshi Nagaoka. 2000. Methods of Information Geometry. Vol. 191.
- Dirk V. Arnold. 2005. Optimal Weighted Recombination. In Foundations of Genetic Algorithms. 215–237.
- Anne Auger and Nikolaus Hansen. 2005. A Restart CMA Evolution Strategy with Increasing Population Size. In IEEE Congress on Evolutionary Computation, Vol. 2. IEEE, 1769–1776.
- Exploring optimal topology of thermal cloaks by CMA-ES. Applied Physics Letters 112, 6 (2018).
- Learning Rate Adaptation by Line Search in Evolution Strategies with Recombination. In Proceedings of the Genetic and Evolutionary Computation Conference. 630–638.
- David Ha and Jürgen Schmidhuber. 2018. World Models. arXiv preprint arXiv:1803.10122 (2018).
- Nikolaus Hansen. 2016. The CMA Evolution Strategy: A Tutorial. arXiv preprint arXiv:1604.00772 (2016).
- Nikolaus Hansen and Anne Auger. 2014. Principled Design of Continuous Stochastic Search: From Theory to Practice. In Theory and Principled Methods for the Design of Metaheuristics. Springer, 145–180.
- COCO: A Platform for Comparing Continuous Optimizers in a Black-Box Setting. Optimization Methods and Software 36, 1 (2021), 114–144.
- Nikolaus Hansen and Stefan Kern. 2004. Evaluating the CMA Evolution Strategy on Multimodal Test Functions. In International Conference on Parallel Problem Solving from Nature. 282–291.
- Nikolaus Hansen and Andreas Ostermeier. 2001. Completely Derandomized Self-Adaptation in Evolution Strategies. Evolutionary Computation 9, 2 (2001), 159–195.
- Michael Hellwig and Hans-Georg Beyer. 2016. Evolution Under Strong Noise: A Self-Adaptive Evolution Strategy Can Reach the Lower Performance Bound - The pcCMSA-ES. In International Conference on Parallel Problem Solving from Nature. 26–36.
- Action-Sufficient State Representation Learning for Control with Structural Constraints. In International Conference on Machine Learning. 9260–9279.
- Three Factors Influencing Minima in SGD. arXiv preprint arXiv:1711.04623 (2017).
- Modeling Visual Containment for Web Page Layout Optimization. In Computer Graphics Forum, Vol. 40. 33–44.
- Constrained Graphic Layout Generation via Latent Optimization. In Proceedings of the ACM International Conference on Multimedia. 88–96.
- Stochastic Differential Equations.
- Oswin Krause. 2019. Large-Scale Noise-Resilient Evolution-Strategies. In Proceedings of the Genetic and Evolutionary Computation Conference. 682–690.
- Maximum Likelihood-based Online Adaptation of Hyper-parameters in CMA-ES. In International Conference on Parallel Problem Solving from Nature. 70–79.
- Application of optimal control theory based on the evolution strategy (CMA-ES) to automatic berthing. Journal of Marine Science and Technology 25 (2020), 221–233.
- Hidekazu Miyazawa and Youhei Akimoto. 2017. Effect of the Mean Vector Learning Rate in CMA-ES. In Proceedings of the Genetic and Evolutionary Computation Conference. 721–728.
- Duc Manh Nguyen and Nikolaus Hansen. 2017. Benchmarking CMAES-APOP on the BBOB Noiseless Testbed. In Proceedings of the Genetic and Evolutionary Computation Conference Companion. 1756–1763.
- Kouhei Nishida and Youhei Akimoto. 2016. Population Size Adaptation for the CMA-ES Based on the Estimation Accuracy of the Natural Gradient. In Proceedings of the Genetic and Evolutionary Computation Conference. 237–244.
- Kouhei Nishida and Youhei Akimoto. 2018. PSA-CMA-ES: CMA-ES with Population Size Adaptation. In Proceedings of the Genetic and Evolutionary Computation Conference. 865–872.
- CMA-ES with Learning Rate Adaptation: Can CMA-ES with Default Population Size Solve Multimodal and Noisy Problems?. In Proceedings of the Genetic and Evolutionary Computation Conference. 839–847.
- Masahiro Nomura and Isao Ono. 2022. Towards a Principled Learning Rate Adaptation for Natural Evolution Strategies. In Applications of Evolutionary Computation. 721–737.
- Warm Starting CMA-ES for Hyperparameter Optimization. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 35. 9188–9196.
- Information-Geometric Optimization Algorithms: A Unifying Picture via Invariance Principles. Journal of Machine Learning Research 18, 18 (2017), 1–65.
- Evolving Losses for Unsupervised Video Representation Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 133–142.
- Lennart Oswald Purucker and Joeran Beel. 2023. CMA-ES for Post Hoc Ensembling in AutoML: A Great Success and Salvageable Failure. In AutoML Conference 2023.
- Level Generation for Angry Birds with Sequential VAE and Latent Variable Evolution. In Proceedings of the Genetic and Evolutionary Computation Conference. 1052–1060.
- Multi-Object Manipulation via Object-Centric Neural Scattering Functions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 9021–9031.
- Evolving Mario Levels in the Latent Space of a Deep Convolutional Generative Adversarial Network. In Proceedings of the Genetic and Evolutionary Computation Conference. 221–228.
- Masahiro Nomura (56 papers)
- Youhei Akimoto (54 papers)
- Isao Ono (10 papers)