Large Language Model-Driven Curriculum Design for Mobile Networks (2405.18039v2)
Abstract: This study introduces an innovative framework that employs LLMs to automate the design and generation of curricula for reinforcement learning (RL). As mobile networks evolve towards the 6G era, managing their increasing complexity and dynamic nature poses significant challenges. Conventional RL approaches often suffer from slow convergence and poor generalization due to conflicting objectives and the large state and action spaces associated with mobile networks. To address these shortcomings, we introduce curriculum learning, a method that systematically exposes the RL agent to progressively challenging tasks, improving convergence and generalization. However, curriculum design typically requires extensive domain knowledge and manual human effort. Our framework mitigates this by utilizing the generative capabilities of LLMs to automate the curriculum design process, significantly reducing human effort while improving the RL agent's convergence and performance. We deploy our approach within a simulated mobile network environment and demonstrate improved RL convergence rates, generalization to unseen scenarios, and overall performance enhancements. As a case study, we consider autonomous coordination and user association in mobile networks. Our obtained results highlight the potential of combining LLM-based curriculum generation with RL for managing next-generation wireless networks, marking a significant step towards fully autonomous network operations.
- W. Saad, M. Bennis, and M. Chen, “A vision of 6G wireless systems: Applications, trends, technologies, and open research problems,” IEEE Network, vol. 34, no. 3, pp. 134–142, 2020.
- D. Liu, L. Wang, Y. Chen, M. Elkashlan, K.-K. Wong, R. Schober, and L. Hanzo, “User association in 5G networks: A survey and an outlook,” IEEE Commun Surveys & Tuts, vol. 18, no. 2, pp. 1018–1044, 2016.
- O. Alhussein and W. Zhuang, “Dynamic topology design of NFV-enabled services using deep reinforcement learning,” IEEE Trans. Cognitive Commun. Netw., vol. 8, no. 2, pp. 1228–1238, 2022.
- Z. Xiong, Y. Zhang, D. Niyato, R. Deng, P. Wang, and L.-C. Wang, “Deep reinforcement learning for mobile 5G and beyond: Fundamentals, applications, and challenges,” IEEE Veh. Technol. Mag., vol. 14, no. 2, pp. 44–52, 2019.
- S. Narvekar, B. Peng, M. Leonetti, J. Sinapov, M. E. Taylor, and P. Stone, “Curriculum learning for reinforcement learning domains: A framework and survey,” J. Mach Learning Research, vol. 21, no. 181, pp. 1–50, 2020.
- M. Sana, A. De Domenico, W. Yu, Y. Lostanlen, and E. C. Strinati, “Multi-agent reinforcement learning for adaptive user association in dynamic mmwave networks,” IEEE Trans. Wireless Commun., vol. 19, no. 10, pp. 6520–6534, 2020.
- S. Schneider, H. Karl, R. Khalili, and A. Hecker, “Multi-agent deep reinforcement learning for coordinated multipoint in mobile networks,” IEEE Trans. Netw. Service Manag., 2023.
- Y. Bengio, J. Louradour, R. Collobert, and J. Weston, “Curriculum learning,” in Proc. ICML, 2009, pp. 41–48.
- “Awesome curriuclum learning,” https://github.com/Openning07/awesome-curriculum-learning, 2022.
- A. Graves, M. G. Bellemare, J. Menick, R. Munos, and K. Kavukcuoglu, “Automated curriculum learning for neural networks,” in Proc. ICML. PMLR, 2017, pp. 1311–1320.
- S. Narvekar, J. Sinapov, and P. Stone, “Autonomous task sequencing for customized curriculum design in reinforcement learning.” in IJCAI, 2017, pp. 2536–2542.
- Y. Du, P. Abbeel, and A. Grover, “It takes four to tango: Multiagent self play for automatic curriculum generation,” in Proc. ICLR, 2022.
- E. Hyytiä and J. Virtamo, “Random waypoint mobility model in cellular networks,” Wireless Networks, vol. 13, no. 2, pp. 177–188, 2007.
- A. Medeisis and A. Kajackas, “On the use of the universal okumura-hata propagation prediction model in rural areas,” in Proc. IEEE VTC, vol. 3, 2000, pp. 1815–1818.
- J. Schulman, F. Wolski, P. Dhariwal, A. Radford, and O. Klimov, “Proximal policy optimization algorithms,” CoRR, vol. abs/1707.06347, 2017.
- S. Schneider, S. Werner, R. Khalili, A. Hecker, and H. Karl, “mobile-env: An open platform for reinforcement learning in wireless mobile networks,” in Proc. IEEE/IFIP NOMS, 2022, pp. 1–3.
- O. Erak, “LLM-CL,” https://github.com/OmarErak/LLM-CL, 2024.
- OpenAI, J. Achiam, S. Adler, S. Agarwal, L. Ahmad, and I. A. et al., “GPT-4 technical report,” 2024.
- Omar Erak (5 papers)
- Omar Alhussein (18 papers)
- Shimaa Naser (11 papers)
- Nouf Alabbasi (4 papers)
- De Mi (19 papers)
- Sami Muhaidat (71 papers)