Dynamic and Super-Personalized Media Ecosystem Driven by Generative AI: Unpredictable Plays Never Repeating The Same (2402.12412v1)
Abstract: This paper introduces a media service model that exploits AI video generators at the receive end. This proposal deviates from the traditional multimedia ecosystem, completely relying on in-house production, by shifting part of the content creation onto the receiver. We bring a semantic process into the framework, allowing the distribution network to provide service elements that prompt the content generator, rather than distributing encoded data of fully finished programs. The service elements include fine-tailored text descriptions, lightweight image data of some objects, or application programming interfaces, comprehensively referred to as semantic sources, and the user terminal translates the received semantic data into video frames. Empowered by the random nature of generative AI, the users could then experience super-personalized services accordingly. The proposed idea incorporates the situations in which the user receives different service providers' element packages; a sequence of packages over time, or multiple packages at the same time. Given promised in-context coherence and content integrity, the combinatory dynamics will amplify the service diversity, allowing the users to always chance upon new experiences. This work particularly aims at short-form videos and advertisements, which the users would easily feel fatigued by seeing the same frame sequence every time. In those use cases, the content provider's role will be recast as scripting semantic sources, transformed from a thorough producer. Overall, this work explores a new form of media ecosystem facilitated by receiver-embedded generative models, featuring both random content dynamics and enhanced delivery efficiency simultaneously.
- Huawei 6G Research Team, “6G: The next horizon,” Huawei, Tech. Rep., Nov. 2021, Accessed 30 Dec. 2023. [Online]. Available: https://www.huawei.com/en/huaweitech/future-technologies/6g-the-next-horizon
- Samsung Research, “6G: The next hyper-connected experience for all,” White Paper, Dec. 2020. Accessed 30 Dec. 2023. [Online]. Available: https://cdn.codeground.org/nsr/downloads/researchareas/20201201_6G_ Vision_web.pdf
- H. Jiang, “A latent space theory for emergent abilities in large language models,” arXiv:2304.09960v3, Sep. 2023.
- [Online] https://audiocraft.metademolab.com
- P. Salembier and J. R. Smith, “MPEG-7 multimedia description schemes,” IEEE Trans. Circ. Syst. Video Tech., vol. 11, no. 6, pp. 748-759, Jun. 2001.
- X. Luo, H.-H. Chen, and Q. Luo, “Semantic communications: Overview, open issues, and future research directions,” IEEE Wireless Commun., 2022.
- M. Smith, “When AI unplugs, all bets are off,” IEEE Spectrum, Dec. 2023. [Online]. Available: https://spectrum.ieee.org/personal-ai-assistant
- J. Song, C. Meng, and S. Ermon, “Denoising diffusion implicit models,” in Proc. ICLR, Jan. 2021.
- P. Dhariwal and A. Nichol, “Diffusion models beat GANs on image synthesis,” in Proc. NeurIPS, 2021.
- PikaLab. Pika Lab discord server. https://www.pika.art. 2023.
- V. Liu and L. B. Chilton, “Design guidelines for prompt engineering text-to-image generative models,” in Proc. CHI’22, New Orleans, LA, USA, May 2022.
- [Online] https://openart.ai/presets
- [Online] https://flowgpt.com/prompt-engineering
- J. M. Martinez, “MPEG-7: Overview of MPEG-7 description tools, Part 2,” IEEE MultiMedia, vol. 9, no. 3, pp. 83-93, 2002.
- Broadcast and On-line Services: Search, select, and rightful use of content on personal storage systems (“TV-Anytime”), ETSI Standard TS 102.822, 2023.
- M. Javaheripi and S. Bubeck, “Phi-2: The surprising power of small language models,” Microsoft Research Blog, Dec. 2023. [Online] https://www.microsoft.com/en-us/research/blog/phi-2-the-surprising-power-of-small-language-models/
- K. Shum, S. Diao, and T. Zhang, “Automatic prompt augmentation and selection with chain-of-thought from labeled data,” arXiv:2302.12822, Feb. 2023.
- K. Liu, “Thoughts and prayers,” Slate, Jan. 2019.
- M. Corl, V. Anishchenko, and T. Mondal, “BPS ATSC 3.0 broadcast emission time stabilization system proof-of concept,” in Proc. NAB Broadcast Engin. Inform. Technol. (BEIT) Conf., Las Vegas, USA, Apr. 2023.
- Sungjun Ahn (1 paper)
- Hyun-Jeong Yim (1 paper)
- Youngwan Lee (18 papers)
- Sung-Ik Park (1 paper)