Generative AI-enabled Mobile Tactical Multimedia Networks: Distribution, Generation, and Perception (2401.06386v1)
Abstract: Mobile multimedia networks (MMNs) demonstrate great potential in delivering low-latency and high-quality entertainment and tactical applications, such as short-video sharing, online conferencing, and battlefield surveillance. For instance, in tactical surveillance of battlefields, scalability and sustainability are indispensable for maintaining large-scale military multimedia applications in MMNs. Therefore, many data-driven networking solutions are leveraged to optimize streaming strategies based on real-time traffic analysis and resource monitoring. In addition, generative AI (GAI) can not only increase the efficiency of existing data-driven solutions through data augmentation but also develop potential capabilities for MMNs, including AI-generated content (AIGC) and AI-aided perception. In this article, we propose the framework of GAI-enabled MMNs that leverage the capabilities of GAI in data and content synthesis to distribute high-quality and immersive interactive content in wireless networks. Specifically, we outline the framework of GAI-enabled MMNs and then introduce its three main features, including distribution, generation, and perception. Furthermore, we propose a second-score auction mechanism for allocating network resources by considering GAI model values and other metrics jointly. The experimental results show that the proposed auction mechanism can effectively increase social welfare by allocating resources and models with the highest user satisfaction.
- T. Fujihashi, T. Koike-Akino, and T. Watanabe, “Soft delivery: Survey on a new paradigm for wireless and mobile multimedia streaming,” ACM Computing Surveys, 2021.
- S. Bond-Taylor, A. Leach, Y. Long, and C. G. Willcocks, “Deep generative modelling: A comparative review of vaes, gans, normalizing flows, energy-based and autoregressive models,” IEEE transactions on pattern analysis and machine intelligence, 2021.
- Y. Zhang, Y. Liu, L. Guo, and J. Y. Lee, “Measurement of a large-scale short-video service over mobile and wireless networks,” IEEE Transactions on Mobile Computing, 2022.
- N. Aldausari, A. Sowmya, N. Marcus, and G. Mohammadi, “Video generative adversarial networks: a review,” ACM Computing Surveys (CSUR), vol. 55, no. 2, pp. 1–25, 2022.
- T. Li, V. Sivaraman, L. Fan, M. Alizadeh, and D. Katabi, “Reparo: Loss-resilient generative codec for video conferencing,” arXiv preprint arXiv:2305.14135, 2023.
- P. Ardhianto, Y. P. Santosa, C. Moniaga, M. P. Utami, C. Dewi, H. J. Christanto, A. P. S. Chen, et al., “Generative deep learning for visual animation in landscapes design,” Scientific Programming, vol. 2023, 2023.
- L. Wang, B. Huang, Z. Zhao, Z. Tong, Y. He, Y. Wang, Y. Wang, and Y. Qiao, “Videomae v2: Scaling video masked autoencoders with dual masking,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 14549–14560, 2023.
- Z. Hu and D. Xu, “Videocontrolnet: A motion-guided video-to-video translation framework by using diffusion model with controlnet,” arXiv preprint arXiv:2307.14073, 2023.
- Q. Zhou, R. Li, S. Guo, Y. Liu, J. Guo, and Z. Xu, “Cadm: Codec-aware diffusion modeling for neural-enhanced video streaming,” arXiv preprint arXiv:2211.08428, 2022.
- R. Lee, S. I. Venieris, and N. D. Lane, “Deep neural network–based enhancement for image and video streaming systems: A survey and future directions,” ACM Computing Surveys (CSUR), vol. 54, no. 8, pp. 1–30, 2021.
- F. Reda, J. Kontkanen, E. Tabellion, D. Sun, C. Pantofaru, and B. Curless, “Film: Frame interpolation for large motion,” in European Conference on Computer Vision, pp. 250–266, Springer, 2022.
- R. Yang, Y. Yang, J. Marino, and S. Mandt, “Insights from generative modeling for neural video compression,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023.
- K. Khan, “A taxonomy for generative adversarial networks in dynamic adaptive streaming over http,” International Journal of Multidisciplinary Research and Analysis, 2023.
- H. Weng, L. Li, Q. Cheng, W. Chen, and Z. Han, “Content caching policy based on gan and distributional reinforcement learning,” in ICC 2020-2020 IEEE International Conference on Communications (ICC), pp. 1–5, IEEE, 2020.
- M. Choi, W. J. Yun, and J. Kim, “Delay-sensitive and power-efficient quality control of dynamic video streaming using adaptive super-resolution,” arXiv preprint arXiv:2110.05783, 2021.
- Minrui Xu (57 papers)
- Dusit Niyato (671 papers)
- Jiawen Kang (204 papers)
- Zehui Xiong (177 papers)
- Song Guo (138 papers)
- Yuguang Fang (55 papers)
- Dong In Kim (168 papers)