Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

CMA-ES with Learning Rate Adaptation (2401.15876v2)

Published 29 Jan 2024 in cs.NE and math.OC

Abstract: The covariance matrix adaptation evolution strategy (CMA-ES) is one of the most successful methods for solving continuous black-box optimization problems. A practically useful aspect of the CMA-ES is that it can be used without hyperparameter tuning. However, the hyperparameter settings still have a considerable impact on performance, especially for difficult tasks, such as solving multimodal or noisy problems. This study comprehensively explores the impact of learning rate on the CMA-ES performance and demonstrates the necessity of a small learning rate by considering ordinary differential equations. Thereafter, it discusses the setting of an ideal learning rate. Based on these discussions, we develop a novel learning rate adaptation mechanism for the CMA-ES that maintains a constant signal-to-noise ratio. Additionally, we investigate the behavior of the CMA-ES with the proposed learning rate adaptation mechanism through numerical experiments, and compare the results with those obtained for the CMA-ES with a fixed learning rate and with population size adaptation. The results show that the CMA-ES with the proposed learning rate adaptation works well for multimodal and/or noisy problems without extremely expensive learning rate tuning.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (38)
  1. Quality Gain Analysis of the Weighted Recombination Evolution Strategy on General Convex Quadratic Functions. Theoretical Computer Science 832 (2020), 42–67.
  2. An ODE Method to Prove the Geometric Convergence of Adaptive Stochastic Algorithms. Stochastic Processes and their Applications 145 (2022), 269–307.
  3. Youhei Akimoto and Nikolaus Hansen. 2020. Diagonal Acceleration for Covariance Matrix Adaptation Evolution Strategies. Evolutionary Computation 28, 3 (2020), 405–435.
  4. Bidirectional Relation between CMA Evolution Strategies and Natural Evolution Strategies. In International Conference on Parallel Problem Solving from Nature. 154–163.
  5. Shun-Ichi Amari and Scott C Douglas. 1998. Why Natural Gradient?. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Vol. 2. 1213–1216.
  6. Shun-ichi Amari and Hiroshi Nagaoka. 2000. Methods of Information Geometry. Vol. 191.
  7. Dirk V. Arnold. 2005. Optimal Weighted Recombination. In Foundations of Genetic Algorithms. 215–237.
  8. Anne Auger and Nikolaus Hansen. 2005. A Restart CMA Evolution Strategy with Increasing Population Size. In IEEE Congress on Evolutionary Computation, Vol. 2. IEEE, 1769–1776.
  9. Exploring optimal topology of thermal cloaks by CMA-ES. Applied Physics Letters 112, 6 (2018).
  10. Learning Rate Adaptation by Line Search in Evolution Strategies with Recombination. In Proceedings of the Genetic and Evolutionary Computation Conference. 630–638.
  11. David Ha and Jürgen Schmidhuber. 2018. World Models. arXiv preprint arXiv:1803.10122 (2018).
  12. Nikolaus Hansen. 2016. The CMA Evolution Strategy: A Tutorial. arXiv preprint arXiv:1604.00772 (2016).
  13. Nikolaus Hansen and Anne Auger. 2014. Principled Design of Continuous Stochastic Search: From Theory to Practice. In Theory and Principled Methods for the Design of Metaheuristics. Springer, 145–180.
  14. COCO: A Platform for Comparing Continuous Optimizers in a Black-Box Setting. Optimization Methods and Software 36, 1 (2021), 114–144.
  15. Nikolaus Hansen and Stefan Kern. 2004. Evaluating the CMA Evolution Strategy on Multimodal Test Functions. In International Conference on Parallel Problem Solving from Nature. 282–291.
  16. Nikolaus Hansen and Andreas Ostermeier. 2001. Completely Derandomized Self-Adaptation in Evolution Strategies. Evolutionary Computation 9, 2 (2001), 159–195.
  17. Michael Hellwig and Hans-Georg Beyer. 2016. Evolution Under Strong Noise: A Self-Adaptive Evolution Strategy Can Reach the Lower Performance Bound - The pcCMSA-ES. In International Conference on Parallel Problem Solving from Nature. 26–36.
  18. Action-Sufficient State Representation Learning for Control with Structural Constraints. In International Conference on Machine Learning. 9260–9279.
  19. Three Factors Influencing Minima in SGD. arXiv preprint arXiv:1711.04623 (2017).
  20. Modeling Visual Containment for Web Page Layout Optimization. In Computer Graphics Forum, Vol. 40. 33–44.
  21. Constrained Graphic Layout Generation via Latent Optimization. In Proceedings of the ACM International Conference on Multimedia. 88–96.
  22. Stochastic Differential Equations.
  23. Oswin Krause. 2019. Large-Scale Noise-Resilient Evolution-Strategies. In Proceedings of the Genetic and Evolutionary Computation Conference. 682–690.
  24. Maximum Likelihood-based Online Adaptation of Hyper-parameters in CMA-ES. In International Conference on Parallel Problem Solving from Nature. 70–79.
  25. Application of optimal control theory based on the evolution strategy (CMA-ES) to automatic berthing. Journal of Marine Science and Technology 25 (2020), 221–233.
  26. Hidekazu Miyazawa and Youhei Akimoto. 2017. Effect of the Mean Vector Learning Rate in CMA-ES. In Proceedings of the Genetic and Evolutionary Computation Conference. 721–728.
  27. Duc Manh Nguyen and Nikolaus Hansen. 2017. Benchmarking CMAES-APOP on the BBOB Noiseless Testbed. In Proceedings of the Genetic and Evolutionary Computation Conference Companion. 1756–1763.
  28. Kouhei Nishida and Youhei Akimoto. 2016. Population Size Adaptation for the CMA-ES Based on the Estimation Accuracy of the Natural Gradient. In Proceedings of the Genetic and Evolutionary Computation Conference. 237–244.
  29. Kouhei Nishida and Youhei Akimoto. 2018. PSA-CMA-ES: CMA-ES with Population Size Adaptation. In Proceedings of the Genetic and Evolutionary Computation Conference. 865–872.
  30. CMA-ES with Learning Rate Adaptation: Can CMA-ES with Default Population Size Solve Multimodal and Noisy Problems?. In Proceedings of the Genetic and Evolutionary Computation Conference. 839–847.
  31. Masahiro Nomura and Isao Ono. 2022. Towards a Principled Learning Rate Adaptation for Natural Evolution Strategies. In Applications of Evolutionary Computation. 721–737.
  32. Warm Starting CMA-ES for Hyperparameter Optimization. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 35. 9188–9196.
  33. Information-Geometric Optimization Algorithms: A Unifying Picture via Invariance Principles. Journal of Machine Learning Research 18, 18 (2017), 1–65.
  34. Evolving Losses for Unsupervised Video Representation Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 133–142.
  35. Lennart Oswald Purucker and Joeran Beel. 2023. CMA-ES for Post Hoc Ensembling in AutoML: A Great Success and Salvageable Failure. In AutoML Conference 2023.
  36. Level Generation for Angry Birds with Sequential VAE and Latent Variable Evolution. In Proceedings of the Genetic and Evolutionary Computation Conference. 1052–1060.
  37. Multi-Object Manipulation via Object-Centric Neural Scattering Functions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 9021–9031.
  38. Evolving Mario Levels in the Latent Space of a Deep Convolutional Generative Adversarial Network. In Proceedings of the Genetic and Evolutionary Computation Conference. 221–228.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Masahiro Nomura (56 papers)
  2. Youhei Akimoto (54 papers)
  3. Isao Ono (10 papers)

Summary

We haven't generated a summary for this paper yet.