Network Slicing via Transfer Learning aided Distributed Deep Reinforcement Learning (2301.03262v2)
Abstract: Deep reinforcement learning (DRL) has been increasingly employed to handle the dynamic and complex resource management in network slicing. The deployment of DRL policies in real networks, however, is complicated by heterogeneous cell conditions. In this paper, we propose a novel transfer learning (TL) aided multi-agent deep reinforcement learning (MADRL) approach with inter-agent similarity analysis for inter-cell inter-slice resource partitioning. First, we design a coordinated MADRL method with information sharing to intelligently partition resource to slices and manage inter-cell interference. Second, we propose an integrated TL method to transfer the learned DRL policies among different local agents for accelerating the policy deployment. The method is composed of a new domain and task similarity measurement approach and a new knowledge transfer approach, which resolves the problem of from whom to transfer and how to transfer. We evaluated the proposed solution with extensive simulations in a system-level simulator and show that our approach outperforms the state-of-the-art solutions in terms of performance, convergence speed and sample efficiency. Moreover, by applying TL, we achieve an additional gain over 27% higher than the coordinate MADRL approach without TL.
- Y. Liu, J. Ding, and X. Liu, “A constrained reinforcement learning based approach for network slicing,” in 2020 IEEE 28th International Conference on Network Protocols (ICNP), 2020, pp. 1–6.
- Q. Liu, T. Han, N. Zhang, and Y. Wang, “DeepSlicing: Deep reinforcement learning assisted resource allocation for network slicing,” in IEEE GLOBECOM, 2020, pp. 1–6.
- N. Zhao, Y.-C. Liang, D. T. Niyato, Y. Pei, M. Wu, and Y. Jiang, “Deep reinforcement learning for user association and resource allocation in heterogeneous cellular networks,” IEEE Transactions on Wireless Communications, vol. 18, pp. 5141–5152, 2019.
- Y. Shao, R. Li, Z. Zhao, and H. Zhang, “Graph attention network-based drl for network slicing management in dense cellular networks,” IEEE WCNC, pp. 1–6, 2021.
- S. J. Pan and Q. Yang, “A survey on transfer learning,” IEEE Transactions on knowledge and data engineering, vol. 22, no. 10, pp. 1345–1359, 2009.
- C. T. Nguyen et al., “Transfer learning for future wireless networks: A comprehensive survey,” arXiv preprint arXiv:2102.07572, 2021.
- M. Wang, Y. Lin, Q. Tian, and G. Si, “Transfer learning promotes 6g wireless communications: recent advances and future challenges,” IEEE Transactions on Reliability, 2021.
- C. Parera, Q. Liao et al., “Transfer learning for tilt-dependent radio map prediction,” IEEE Transactions on Cognitive Communications and Networking, vol. 6, no. 2, pp. 829–843, 2020.
- F. Zhuang, Z. Qi, K. Duan, D. Xi, Y. Zhu, H. Zhu, H. Xiong, and Q. He, “A comprehensive survey on transfer learning,” Proceedings of the IEEE, vol. 109, no. 1, pp. 43–76, 2020.
- M. E. Taylor and P. Stone, “Transfer learning for reinforcement learning domains: A survey,” J. Mach. Learn. Res., vol. 10, pp. 1633–1685, 2009.
- Z. Zhu, K. Lin, and J. Zhou, “Transfer learning in deep reinforcement learning: A survey,” CoRR, vol. abs/2009.07888, 2020. [Online]. Available: https://arxiv.org/abs/2009.07888
- A. M. Nagib, H. Abou-zeid, and H. S. Hassanein, “Transfer learning-based accelerated deep reinforcement learning for 5G RAN slicing,” IEEE 46th LCN, pp. 249–256, 2021.
- T. Mai, H. Yao et al., “Transfer reinforcement learning aided distributed network slicing resource optimization in industrial IoT,” IEEE Transactions on Industrial Informatics, 2021.
- T. Hu, Q. Liao, Q. Liu, D. Wellington, and G. Carle, “Inter-cell slicing resource partitioning via coordinated multi-agent deep reinforcement learning,” in IEEE ICC, 2022.
- R. L. G. Cavalcante, Q. Liao, and S. Stańczak, “Connections between spectral properties of asymptotic mappings and solutions to wireless network problems,” IEEE Transactions on Signal Processing, vol. 67, pp. 2747–2760, 2019.
- S. Fujimoto et al., “Addressing function approximation error in Actor-Critic methods,” ArXiv, vol. abs/1802.09477, 2018.
- K. Sohn, H. Lee, and X. Yan, “Learning structured output representation using deep conditional generative models,” in NIPS, 2015.