Evolution Transformer: In-Context Evolutionary Optimization (2403.02985v1)
Abstract: Evolutionary optimization algorithms are often derived from loose biological analogies and struggle to leverage information obtained during the sequential course of optimization. An alternative promising approach is to leverage data and directly discover powerful optimization principles via meta-optimization. In this work, we follow such a paradigm and introduce Evolution Transformer, a causal Transformer architecture, which can flexibly characterize a family of Evolution Strategies. Given a trajectory of evaluations and search distribution statistics, Evolution Transformer outputs a performance-improving update to the search distribution. The architecture imposes a set of suitable inductive biases, i.e. the invariance of the distribution update to the order of population members within a generation and equivariance to the order of the search dimensions. We train the model weights using Evolutionary Algorithm Distillation, a technique for supervised optimization of sequence models using teacher algorithm trajectories. The resulting model exhibits strong in-context optimization performance and shows strong generalization capabilities to otherwise challenging neuroevolution tasks. We analyze the resulting properties of the Evolution Transformer and propose a technique to fully self-referentially train the Evolution Transformer, starting from a random initialization and bootstrapping its own learning progress. We provide an open source implementation under https://github.com/RobertTLange/evosax.
- JAX: composable transformations of Python+NumPy programs. (2018). http://github.com/google/jax
- Learning to learn without gradient descent by gradient descent. In International Conference on Machine Learning. PMLR, 748–756.
- Towards learning universal hyperparameter optimizers with transformers. Advances in Neural Information Processing Systems 35 (2022), 32053–32068.
- Multi-step Planning for Automated Hyperparameter Optimization with OptFormer. arXiv preprint arXiv:2210.04971 (2022).
- Brax–A Differentiable Physics Engine for Large Scale Rigid Body Simulation. arXiv preprint arXiv:2106.13281 (2021).
- Nikolaus Hansen. 2006. The CMA evolution strategy: a comparing review. Towards a new evolutionary computation: Advances in the estimation of distribution algorithms (2006), 75–102.
- Real-parameter black-box optimization benchmarking 2010: Experimental setup. Ph.D. Dissertation. INRIA.
- Array programming with NumPy. Nature 585, 7825 (2020), 357–362.
- John D Hunter. 2007. Matplotlib: A 2D graphics environment. IEEE Annals of the History of Computing 9, 03 (2007), 90–95.
- Population based training of neural networks. arXiv preprint arXiv:1711.09846 (2017).
- Perceiver: General perception with iterative attention. In International conference on machine learning. PMLR, 4651–4664.
- General-purpose in-context learning by meta-learning transformers. arXiv preprint arXiv:2212.04458 (2022).
- Louis Kirsch and Jürgen Schmidhuber. 2022. Eliminating meta optimization through self-referential meta learning. arXiv preprint arXiv:2212.14392 (2022).
- Self-attention between datapoints: Going beyond individual input-output pairs in deep learning. Advances in Neural Information Processing Systems 34 (2021), 28742–28756.
- Generative pretraining for black-box optimization. arXiv preprint arXiv:2206.10786 (2022).
- Discovering Attention-Based Genetic Algorithms via Meta-Black-Box Optimization. In Proceedings of the Genetic and Evolutionary Computation Conference. 929–937.
- Discovering evolution strategies via meta-black-box optimization. In Proceedings of the Companion Conference on Genetic and Evolutionary Computation. 29–30.
- NeuroEvoBench: Benchmarking Evolutionary Optimizers for Deep Learning Applications. Advances in Neural Information Processing Systems 36 (2024).
- Robert Tjarko Lange. 2021. MLE-Infrastructure: A Set of Lightweight Tools for Distributed Machine Learning Experimentation. (2021). http://github.com/mle-infrastructure
- Robert Tjarko Lange. 2022a. evosax: JAX-based Evolution Strategies. arXiv preprint arXiv:2212.04180 (2022).
- Robert Tjarko Lange. 2022b. gymnax: A JAX-based Reinforcement Learning Environment Library. (2022). http://github.com/RobertTLange/gymnax
- Large Language Models As Evolution Strategies. arXiv preprint arXiv:2402.18381 (2024).
- In-context reinforcement learning with algorithm distillation. arXiv preprint arXiv:2210.14215 (2022).
- Set transformer: A framework for attention-based permutation-invariant neural networks. In International conference on machine learning. PMLR, 3744–3753.
- Structured state space models for in-context reinforcement learning. arXiv preprint arXiv:2303.03982 (2023).
- Arbitrary Order Meta-Learning with Simple Population-Based Evolution. arXiv preprint arXiv:2303.09478 (2023).
- Training learned optimizers with randomly initialized learned optimizers. arXiv preprint arXiv:2101.07367 (2021).
- Improving language understanding by generative pre-training. (2018).
- Exploring the limits of transfer learning with a unified text-to-text transformer. The Journal of Machine Learning Research 21, 1 (2020), 5485–5551.
- Ingo Rechenberg. 1978. Evolutionsstrategien. In Simulationsmethoden in der Medizin und Biologie. Springer, 83–114.
- Raymond Ros and Nikolaus Hansen. 2008. A simple modification in CMA-ES achieving linear time and space complexity. In International conference on parallel problem solving from nature. Springer, 296–305.
- Evolution strategies as a scalable alternative to reinforcement learning. arXiv preprint arXiv:1703.03864 (2017).
- High dimensions and heavy tails for natural evolution strategies. In Proceedings of the 13th annual conference on Genetic and evolutionary computation. 845–852.
- Jürgen Schmidhuber. 1987. Evolutionary principles in self-referential learning, or on learning how to learn: the meta-meta-… hook. Ph.D. Dissertation. Technische Universität München.
- Deep neuroevolution: Genetic algorithms are a competitive alternative for training deep neural networks for reinforcement learning. arXiv preprint (2017).
- Yujin Tang and David Ha. 2021. The sensory neuron as a transformer: Permutation-invariant neural networks for reinforcement learning. Advances in Neural Information Processing Systems 34 (2021), 22574–22587.
- EvoJAX: Hardware-Accelerated Neuroevolution. arXiv preprint arXiv:2202.05008 (2022).
- Adaptive Agents Team. 2023. Human-timescale adaptation in an open-ended task space. arXiv preprint arXiv:2301.07608 (2023).
- Attention is all you need. Advances in neural information processing systems 30 (2017).
- Michael L Waskom. 2021. Seaborn: statistical data visualization. Journal of Open Source Software 6, 60 (2021), 3021.
- Natural evolution strategies. The Journal of Machine Learning Research 15, 1 (2014).
- Robert Tjarko Lange (21 papers)
- Yingtao Tian (32 papers)
- Yujin Tang (31 papers)