Inter-Cell Network Slicing With Transfer Learning Empowered Multi-Agent Deep Reinforcement Learning (2306.11552v1)
Abstract: Network slicing enables operators to efficiently support diverse applications on a common physical infrastructure. The ever-increasing densification of network deployment leads to complex and non-trivial inter-cell interference, which requires more than inaccurate analytic models to dynamically optimize resource management for network slices. In this paper, we develop a DIRP algorithm with multiple deep reinforcement learning (DRL) agents to cooperatively optimize resource partition in individual cells to fulfill the requirements of each slice, based on two alternative reward functions. Nevertheless, existing DRL approaches usually tie the pretrained model parameters to specific network environments with poor transferability, which raises practical deployment concerns in large-scale mobile networks. Hence, we design a novel transfer learning-aided DIRP (TL-DIRP) algorithm to ease the transfer of DIRP agents across different network environments in terms of sample efficiency, model reproducibility, and algorithm scalability. The TL-DIRP algorithm first centrally trains a generalized model and then transfers the "generalist" to each local agent as "specialist" with distributed finetuning and execution. TL-DIRP consists of two steps: 1) centralized training of a generalized distributed model, 2) transferring the "generalist" to each "specialist" with distributed finetuning and execution. The numerical results show that not only DIRP outperforms existing baseline approaches in terms of faster convergence and higher reward, but more importantly, TL-DIRP significantly improves the service performance, with reduced exploration cost, accelerated convergence rate, and enhanced model reproducibility. As compared to a traffic-aware baseline, TL-DIRP provides about 15% less violation ratio of the quality of service (QoS) for the worst slice service and 8.8% less violation on the average service QoS.
- T. Hu, Q. Liao, Q. Liu, D. Wellington, and G. Carle, “Inter-cell slicing resource partitioning via coordinated multi-agent deep reinforcement learning,” in IEEE International Conference Communications (ICC), 2022.
- A. Ksentini and N. Nikaein, “Toward enforcing network slicing on RAN: Flexibility and resources abstraction,” IEEE Communications Magazine, vol. 55, no. 6, pp. 102–108, 2017.
- P. L. Vo, M. N. Nguyen, T. A. Le, and N. H. Tran, “Slicing the edge: Resource allocation for RAN network slicing,” IEEE Wireless Communications Letters, vol. 7, no. 6, pp. 970–973, 2018.
- R. A. Addad, M. Bagaa, T. Taleb, D. Dutra, and H. Flinck, “Optimization model for cross-domain network slices in 5G networks,” IEEE Transactions on Mobile Computing, vol. 19, pp. 1156–1169, 2020.
- H. Beshley, M. Beshley, M. Medvetskyi, and J. Pyrih, “QoS-aware optimal radio resource allocation method for machine-type communications in 5G LTE and beyond cellular networks,” Wirel. Commun. Mob. Comput., vol. 2021, pp. 9 966 366:1–9 966 366:18, 2021.
- F. Fossati, S. Moretti, P. Perny, and S. Secci, “Multi-resource allocation for network slicing,” IEEE/ACM Transactions on Networking, vol. 28, pp. 1311–1324, 2020.
- T. Ma, Y. Zhang, F. Wang, D. Wang, and D. Guo, “Slicing resource allocation for eMBB and URLLC in 5G RAN,” Wirel. Commun. Mob. Comput., vol. 2020, pp. 6 290 375:1–6 290 375:11, 2020.
- 3GPP, TS 23.501, “System architecture for the 5G System (5GS), V17.4.0,” 3GPP, March 2022.
- H. Mao, M. Alizadeh, I. Menache, and S. Kandula, “Resource management with deep reinforcement learning,” Proceedings of the 15th ACM Workshop on Hot Topics in Networks, 2016.
- Y. Liu, J. Ding, and X. Liu, “A constrained reinforcement learning based approach for network slicing,” in IEEE 28th International Conference on Network Protocols (ICNP), 2020, pp. 1–6.
- Q. Liu, T. Han, N. Zhang, and Y. Wang, “DeepSlicing: Deep reinforcement learning assisted resource allocation for network slicing,” in IEEE Global Communications Conference (GLOBECOM), 2020, pp. 1–6.
- R. Li, Z. Zhao, Q. Sun, C.-L. I, C. Yang, X. Chen, M. Zhao, and H. Zhang, “Deep reinforcement learning for resource management in network slicing,” IEEE Access, vol. 6, pp. 74 429–74 441, 2018.
- I. Alqerm and B. Shihada, “A cooperative online learning scheme for resource allocation in 5G systems,” 2016 IEEE International Conference on Communications (ICC), pp. 1–7, 2016.
- N. Zhao, Y.-C. Liang, D. T. Niyato, Y. Pei, M. Wu, and Y. Jiang, “Deep reinforcement learning for user association and resource allocation in heterogeneous cellular networks,” IEEE Transactions on Wireless Communications, vol. 18, pp. 5141–5152, 2019.
- Y. Shao, R. Li, Z. Zhao, and H. Zhang, “Graph attention network-based DRL for network slicing management in dense cellular networks,” 2021 IEEE Wireless Communications and Networking Conference (WCNC), pp. 1–6, 2021.
- H. Nie, S. Li, and Y. Liu, “Multi-agent deep reinforcement learning for resource allocation in the multi-objective HetNet,” 2021 International Wireless Communications and Mobile Computing (IWCMC), pp. 116–121, 2021.
- S. J. Pan and Q. Yang, “A survey on transfer learning,” IEEE Transactions on knowledge and data engineering, vol. 22, no. 10, pp. 1345–1359, 2009.
- C. T. Nguyen, N. Van Huynh, N. H. Chu, Y. M. Saputra, D. T. Hoang, D. N. Nguyen, Q.-V. Pham, D. Niyato, E. Dutkiewicz, and W.-J. Hwang, “Transfer learning for future wireless networks: A comprehensive survey,” arXiv preprint arXiv:2102.07572, 2021.
- M. Wang, Y. Lin, Q. Tian, and G. Si, “Transfer learning promotes 6g wireless communications: Recent advances and future challenges,” IEEE Transactions on Reliability, 2021.
- C. Parera, Q. Liao, I. Malanchini, C. Tatino, A. E. C. Redondi, and M. Cesana, “Transfer learning for tilt-dependent radio map prediction,” IEEE Transactions on Cognitive Communications and Networking, vol. 6, no. 2, pp. 829–843, 2020.
- M. E. Taylor, P. Stone, and Y. Liu, “Transfer learning via inter-task mappings for temporal difference learning,” J. Mach. Learn. Res., vol. 8, pp. 2125–2167, 2007.
- F. Zhuang, Z. Qi, K. Duan, D. Xi, Y. Zhu, H. Zhu, H. Xiong, and Q. He, “A comprehensive survey on transfer learning,” Proceedings of the IEEE, vol. 109, no. 1, pp. 43–76, 2020.
- M. E. Taylor and P. Stone, “Transfer learning for reinforcement learning domains: A survey,” J. Mach. Learn. Res., vol. 10, pp. 1633–1685, 2009.
- Z. Zhu, K. Lin, and J. Zhou, “Transfer learning in deep reinforcement learning: A survey,” arXiv preprint arXiv:2009.07888, 2020.
- A. M. Nagib, H. Abou-Zeid, and H. S. Hassanein, “Transfer learning-based accelerated deep reinforcement learning for 5G RAN slicing,” in 2021 IEEE 46th Conference on Local Computer Networks (LCN). IEEE, 2021, pp. 249–256.
- T. Mai, H. Yao, N. Zhang, W. He, D. Guo, and M. Guizani, “Transfer reinforcement learning aided distributed network slicing resource optimization in industrial IoT,” IEEE Transactions on Industrial Informatics, 2021.
- H. Zafar, Z. Utkovski, M. Kasparick, and S. Stańczak, “Transfer learning in multi-agent reinforcement learning with double Q-networks for distributed resource sharing in V2X communication,” ArXiv, vol. abs/2107.06195, 2021.
- 3GPP, TS 28.530, “Technical Specification Group Services and System Aspects; Management and orchestration; Concepts, use cases and requirements, V17.2.0,” 3GPP, December 2021.
- J. Mo and J. Walrand, “Fair end-to-end window-based congestion control,” IEEE/ACM Transactions on Networking, vol. 8, no. 5, pp. 556–567, 2000.
- T. Bonald, L. Massoulié, A. Proutiere, and J. Virtamo, “A queueing analysis of max-min fairness, proportional fairness and balanced fairness,” Queueing systems, vol. 53, no. 1, pp. 65–84, 2006.
- J. Ewing, “Autonomic performance optimization with application to self-architecting software systems,” Ph.D. dissertation, 04 2015.
- R. L. G. Cavalcante, Q. Liao, and S. Stańczak, “Connections between spectral properties of asymptotic mappings and solutions to wireless network problems,” IEEE Transactions on Signal Processing, vol. 67, pp. 2747–2760, 2019.
- V. Sciancalepore, I. Filippini, V. Mancuso, A. Capone, and A. Banchs, “A multi-traffic inter-cell interference coordination scheme in dense cellular networks,” IEEE/ACM Transactions on Networking, vol. 26, pp. 2361–2375, 2018.
- Z. Xu, J. Tang, J. Meng, W. Zhang, Y. Wang, C. Liu, and D. Yang, “Experience-driven networking: A deep reinforcement learning based approach,” IEEE INFOCOM 2018 - IEEE Conference on Computer Communications, pp. 1871–1879, 2018.
- H. Song, L. Liu, J. D. Ashdown, and Y. C. Yi, “A deep reinforcement learning framework for spectrum management in dynamic spectrum access,” IEEE Internet of Things Journal, vol. 8, pp. 11 208–11 218, 2021.
- H. Peng and X. S. Shen, “Deep reinforcement learning based resource management for multi-access edge computing in vehicular networks,” IEEE Transactions on Network Science and Engineering, vol. 7, pp. 2416–2428, 2020.
- D. Xu, P. Qiao, and Y. Dou, “Aggregation transfer learning for multi-agent reinforcement learning,” 2021 2nd International Conference on Big Data & Artificial Intelligence & Software Engineering (ICBASE), pp. 547–551, 2021.
- J. N. Foerster, Y. M. Assael, N. de Freitas, and S. Whiteson, “Learning to communicate with deep multi-agent reinforcement learning,” in NIPS, 2016.
- V. Konda and J. Tsitsiklis, “Actor-Critic algorithms,” in NIPS, 1999.
- S. Fujimoto, H. V. Hoof, and D. Meger, “Addressing function approximation error in Actor-Critic methods,” ArXiv, vol. abs/1802.09477, 2018.
- D. Silver, G. Lever, N. Heess, T. Degris, D. Wierstra, and M. A. Riedmiller, “Deterministic policy gradient algorithms,” in ICML, 2014.
- G. Kang, L. Jiang, Y. Yang, and A. G. Hauptmann, “Contrastive adaptation network for unsupervised domain adaptation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 4893–4902.
- N. S. Networks, “White paper: Self-organizing network (SON): Introducing the nokia siemens networks SON suite-an efficient, future-proof platform for SON.” October, 2009.
- J. Meinilä, P. Kyösti, L. Hentilä, T. Jämsä, E. Suikkanen, E. Kunnari, and M. Narandžić, “Wireless world initiative new radio - Winner+,” 2010.