Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Deep Neuroevolution: Genetic Algorithms Are a Competitive Alternative for Training Deep Neural Networks for Reinforcement Learning (1712.06567v3)

Published 18 Dec 2017 in cs.NE and cs.LG

Abstract: Deep artificial neural networks (DNNs) are typically trained via gradient-based learning algorithms, namely backpropagation. Evolution strategies (ES) can rival backprop-based algorithms such as Q-learning and policy gradients on challenging deep reinforcement learning (RL) problems. However, ES can be considered a gradient-based algorithm because it performs stochastic gradient descent via an operation similar to a finite-difference approximation of the gradient. That raises the question of whether non-gradient-based evolutionary algorithms can work at DNN scales. Here we demonstrate they can: we evolve the weights of a DNN with a simple, gradient-free, population-based genetic algorithm (GA) and it performs well on hard deep RL problems, including Atari and humanoid locomotion. The Deep GA successfully evolves networks with over four million free parameters, the largest neural networks ever evolved with a traditional evolutionary algorithm. These results (1) expand our sense of the scale at which GAs can operate, (2) suggest intriguingly that in some cases following the gradient is not the best choice for optimizing performance, and (3) make immediately available the multitude of neuroevolution techniques that improve performance. We demonstrate the latter by showing that combining DNNs with novelty search, which encourages exploration on tasks with deceptive or sparse reward functions, can solve a high-dimensional problem on which reward-maximizing algorithms (e.g.\ DQN, A3C, ES, and the GA) fail. Additionally, the Deep GA is faster than ES, A3C, and DQN (it can train Atari in ${\raise.17ex\hbox{$\scriptstyle\sim$}}$4 hours on one desktop or ${\raise.17ex\hbox{$\scriptstyle\sim$}}$1 hour distributed on 720 cores), and enables a state-of-the-art, up to 10,000-fold compact encoding technique.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Felipe Petroski Such (14 papers)
  2. Vashisht Madhavan (7 papers)
  3. Edoardo Conti (5 papers)
  4. Joel Lehman (34 papers)
  5. Kenneth O. Stanley (33 papers)
  6. Jeff Clune (65 papers)
Citations (673)

Summary

Overview of "Deep Neuroevolution: Genetic Algorithms are a Competitive Alternative for Training Deep Neural Networks for Reinforcement Learning"

The paper investigates the application of Genetic Algorithms (GAs) to train Deep Neural Networks (DNNs) within the context of Reinforcement Learning (RL) tasks. Traditionally, DNNs have been trained using gradient-based methods such as backpropagation, Q-learning, and policy gradient methods. However, the authors propose that non-gradient-based evolutionary algorithms, specifically Genetic Algorithms, can be a viable alternative.

Key Contributions

  1. Scale and Performance: The researchers demonstrate that GAs can evolve DNNs with over four million parameters, which is significant given the historical assumption that GAs would be infeasible at this scale. The Deep GA performed well on complex RL problems like Atari games and humanoid locomotion tasks.
  2. Comparison with Other Methods: The paper provides a comparative analysis between GAs and other RL algorithms, such as DQN, A3C, and Evolution Strategies (ES). Results show that GAs can be competitive, even outperforming these methods in certain tasks.
  3. Novelty Search Integration: By integrating a novelty search, the paper showcases how exploration strategies from the neuroevolution community can be effectively used with high-dimensional networks to overcome deceptive problems, achieving success where traditional reward-maximizing algorithms fail.
  4. Efficiency: One of the significant advantages highlighted is the computational efficiency of GAs. The Deep GA can train complex tasks faster than ES, A3C, and DQN, primarily due to better parallelization capabilities.
  5. Compact Encoding: A novel method for storing large neural networks efficiently is introduced, allowing for a significant reduction in memory usage. This encoding facilitates distributed training, enabling broader accessibility for individual researchers.

Implications and Future Directions

The implications of these findings are both practical and theoretical:

  • Practical: The ability of GAs to quickly find effective solutions in large parameter spaces opens new possibilities for leveraging neuroevolution in industrial applications, especially where gradient-based methods struggle.
  • Theoretical: The paper challenges the conventional belief that gradient-following is necessary for optimizing DNNs, highlighting that alternative search strategies can also be effective. This could lead to further research into hybrid algorithms that combine the strengths of both gradient-based and gradient-free methods.
  • Exploration Strategies: The success of integrating novelty search suggests further research into adaptive exploration strategies could enhance GA performance and potentially improve other RL algorithms.

In conclusion, this paper argues for reconsidering the role of Genetic Algorithms in the training of deep neural networks, especially for reinforcement learning tasks. This perspective might influence future AI research directions, encouraging a re-evaluation of “old” algorithms in light of modern computational capabilities.

Youtube Logo Streamline Icon: https://streamlinehq.com