Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Deep Learning-Based Operators for Evolutionary Algorithms (2407.10477v1)

Published 15 Jul 2024 in cs.NE and cs.LG

Abstract: We present two novel domain-independent genetic operators that harness the capabilities of deep learning: a crossover operator for genetic algorithms and a mutation operator for genetic programming. Deep Neural Crossover leverages the capabilities of deep reinforcement learning and an encoder-decoder architecture to select offspring genes. BERT mutation masks multiple gp-tree nodes and then tries to replace these masks with nodes that will most likely improve the individual's fitness. We show the efficacy of both operators through experimentation.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (29)
  1. Ahmed, Z.H.: Genetic algorithm for the traveling salesman problem using sequential constructive crossover operator. International Journal of Biometrics & Bioinformatics (IJBB) 3(6), 96 (2010)
  2. In: 2019 International Conference on Electrical, Computer and Communication Engineering (ECCE), pp. 1–6. IEEE (2019). DOI 10.1109/ECACE.2019.8679367
  3. Morgan Kaufmann Publishers Inc. (1998)
  4. arXiv preprint arXiv:1611.09940 (2016)
  5. IEEE Transactions on Evolutionary Computation 26(4), 793–799 (2022)
  6. UCI Machine Learning Repository (2014). URL https://doi.org/10.24432/C5VW2C. DOI: 10.24432/C5VW2C
  7. arXiv preprint arXiv:1810.04805 (2018)
  8. In: International conference on parallel problem solving from nature, pp. 78–87. Springer (1994)
  9. Springer (2015)
  10. Friedman, J.H.: Multivariate adaptive regression splines. The annals of statistics 19(1), 1–67 (1991)
  11. gplearn github. https://github.com/trevorstephens/gplearn
  12. American Mathematical Soc. (1996)
  13. Transactions of the association for computational linguistics 8, 64–77 (2020)
  14. In: 2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC), pp. 4204–4209. IEEE (2020)
  15. In: Handbook of Evolutionary Machine Learning, pp. 331–366. Springer (2023)
  16. Applied Soft Computing 146, 110,680 (2023)
  17. Expert Systems with Applications 189, 116,042 (2022)
  18. Magalhaes-Mendes, J.: A comparative study of crossover operators for genetic algorithms to solve the job shop scheduling problem. WSEAS transactions on computers 12(4), 164–173 (2013)
  19. In: First international conference on genetic algorithms in engineering systems: innovations and applications, pp. 487–492. IET (1995)
  20. Evolutionary Computation 6(3), 231–252 (1998). DOI 10.1162/evco.1998.6.3.231. URL https://doi.org/10.1162/evco.1998.6.3.231
  21. Draft, US Army Space and Missile Defense Command, Huntsville, Alabama, USA (2002)
  22. In: Advances in Swarm Intelligence: Third International Conference, ICSI 2012, Shenzhen, China, June 17-20, 2012 Proceedings, Part I 3, pp. 414–421. Springer (2012)
  23. In: Proceedings of the Genetic and Evolutionary Computation Conference (to be published), GECCO ’24. Association for Computing Machinery (2024). URL https://arxiv.org/abs/2403.11159
  24. In: 15th international conference on soft computing, Mendel, vol. 9, pp. 73–91 (2009)
  25. In: European Conference on Genetic Programming, pp. 152–161. Springer (2002)
  26. Advances in Neural Information Processing Systems 28 (2015). URL https://proceedings.neurips.cc/paper_files/paper/2015/file/29921001f2f04bd3baee84a12e98098f-Paper.pdf
  27. Williams, R.J.: Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine learning 8, 229–256 (1992)
  28. arXiv preprint arXiv:1910.03771 (2019)
  29. Yeh, I.C.: Concrete Compressive Strength. UCI Machine Learning Repository (2007). DOI: https://doi.org/10.24432/C5PK67

Summary

  • The paper proposes the Deep Neural Crossover (DNC) operator that employs deep reinforcement learning to capture complex gene correlations for enhanced GA performance.
  • The paper introduces BERT Mutation, which adapts masked language modeling to optimize GP mutation, achieving superior fitness on benchmark tests.
  • The paper demonstrates that applying transfer-learning reduces DRL computational overhead, broadening applicability across scheduling and combinatorial optimization tasks.

Deep Learning-Based Operators for Evolutionary Algorithms

Overview

The paper "Deep Learning-Based Operators for Evolutionary Algorithms" by Eliad Shem-Tov, Moshe Sipper, and Achiya Elyasaf introduces two novel deep learning-driven operators designed to improve the efficacy of genetic algorithms (GAs) and genetic programming (GP). Specifically, the paper presents a multi-parent crossover operator dubbed Deep Neural Crossover (DNC) and a mutation operator based on the architecture of BERT, named BERT Mutation.

Deep Neural Crossover (DNC)

The DNC operator leverages deep reinforcement learning (DRL) coupled with an encoder-decoder architecture to dynamically select advantageous genes during the crossover process in GAs. Unlike traditional crossover methods such as one-point or uniform crossover, which rely on random selection, DNC uses an RNN-based model to encode parental genomes and a pointing mechanism to decode and generate offspring by learning gene correlations. This approach eliminates biases inherent in traditional methods and allows for capturing both linear and nonlinear relationships between genes.

The paper highlights the computational intensity of training a DRL model but proposes a transfer-learning strategy. This involves initially training the DNC on a specific problem within a domain and subsequently applying the trained model to other problems in the same domain, thus reducing the computational overhead.

In evaluation, DNC outperformed all traditional crossover methods in terms of solution quality over benchmark domains such as Graph Coloring and Bin Packing, showing significant improvements with notably better fitness scores and convergence rates.

BERT Mutation

BERT Mutation adapts principles from NLP, particularly the Masked LLMing (MLM) of BERT to the GP mutation process. This operator masks multiple nodes in a GP tree and uses a trained BERT model to predict replacements that are likely to enhance individual fitness. Instead of relying on ground-truth-based training, BERT Mutation employs reinforcement learning, optimizing for fitness improvement as a reward. The method takes into account the structural context of the GP tree, ensuring valid replacements through a constrained sampling process.

The experimental results demonstrate that BERT Mutation significantly outperforms traditional GP mutation operators such as point mutation, subtree mutation, and hoist mutation across various benchmark datasets, including both standard regression problems and specially designed symbolic regression tasks. Notably, BERT Mutation achieves superior fitness while maintaining competitive runtime performance.

Implications and Future Directions

The practical implications of these findings are manifold. By utilizing deep learning to inform crossover and mutation processes, the evolutionary algorithms can explore the solution space more effectively, potentially leading to better optimization outcomes in various domains including scheduling, feature selection, and combinatorial problems. The transfer-learning concept applied in DNC showcases an efficient way to mitigate the computational load typically associated with DRL.

Theoretically, these findings suggest a compelling synergy between evolutionary algorithms and deep learning techniques, opening up avenues for further research. Possible future directions include expanding the DNC approach to additional domains, exploring the state maintenance capabilities of BERT Mutation more extensively in dynamic environments, and integrating these operators into hybrid evolutionary frameworks.

Conclusion

In conclusion, the methods proposed by Shem-Tov et al. mark a significant step forward in enhancing the operability and effectiveness of evolutionary algorithms through deep learning techniques. By addressing traditional shortcomings in crossover and mutation operators, these innovations set a new benchmark for domain-independent genetic operators, promising broad applicability and substantial performance gains across various evolutionary computing tasks.