Trustworthy AI-Generative Content for Intelligent Network Service: Robustness, Security, and Fairness
Abstract: AI-generated content (AIGC) models, represented by LLMs (LLM), have revolutionized content creation. High-speed next-generation communication technology is an ideal platform for providing powerful AIGC network services. At the same time, advanced AIGC techniques can also make future network services more intelligent, especially various online content generation services. However, the significant untrustworthiness concerns of current AIGC models, such as robustness, security, and fairness, greatly affect the credibility of intelligent network services, especially in ensuring secure AIGC services. This paper proposes TrustGAIN, a trustworthy AIGC framework that incorporates robust, secure, and fair network services. We first discuss the robustness to adversarial attacks faced by AIGC models in network systems and the corresponding protection issues. Subsequently, we emphasize the importance of avoiding unsafe and illegal services and ensuring the fairness of the AIGC network services. Then as a case study, we propose a novel sentiment analysis-based detection method to guide the robust detection of unsafe content in network services. We conduct our experiments on fake news, malicious code, and unsafe review datasets to represent LLM application scenarios. Our results indicate that TrustGAIN is an exploration of future networks that can support trustworthy AIGC network services.
- J. Yang, H. Jin, R. Tang, X. Han, Q. Feng, H. Jiang, S. Zhong, B. Yin, and X. Hu, “Harnessing the power of llms in practice: A survey on chatgpt and beyond,” ACM Trans. Knowl. Discov. Data, vol. 18, no. 6, apr 2024.
- Y. Chen, R. Li, Z. Zhao, C. Peng, J. Wu, E. Hossain, and H. Zhang, “Netgpt: An ai-native network architecture for provisioning beyond personalized generative services,” IEEE Network, 2024.
- S. Li, X. Lin, H. Xu, K. Hua, X. Jin, G. Li, and J. Li, “Multi-agent rl-based industrial aigc service offloading over wireless edge networks,” 2024.
- A. Wan, E. Wallace, S. Shen, and D. Klein, “Poisoning language models during instruction tuning,” arXiv preprint arXiv:2305.00944, 2023.
- H. Du, D. Niyato, J. Kang, Z. Xiong, K.-Y. Lam, Y. Fang, and Y. Li, “Spear or shield: Leveraging generative ai to tackle security threats of intelligent network services,” 2023.
- X. Fang, S. Che, M. Mao, H. Zhang, M. Zhao, and X. Zhao, “Bias of ai-generated content: An examination of news produced by large language models,” arXiv preprint arXiv:2309.09825, 2023.
- L. N. Tidjon and F. Khomh, “Threat assessment in machine learning based systems,” arXiv preprint arXiv:2207.00091, 2022.
- I. Shumailov, Y. Zhao, D. Bates, N. Papernot, R. Mullins, and R. Anderson, “Sponge examples: Energy-latency attacks on neural networks,” in 2021 IEEE European symposium on security and privacy (EuroS&P). IEEE, 2021, pp. 212–231.
- N. Lukas, A. Salem, R. Sim, S. Tople, L. Wutschitz, and S. Zanella-Béguelin, “Analyzing leakage of personally identifiable information in language models,” in 2023 IEEE Symposium on Security and Privacy (SP). IEEE, 2023, pp. 346–363.
- J. Wang, H. Du, D. Niyato, J. Kang, Z. Xiong, D. Rajan, S. Mao, and X. Shen, “A unified framework for guiding generative ai with wireless perception in resource constrained mobile edge networks,” IEEE Transactions on Mobile Computing, 2024.
- R. Naik and B. Nushi, “Social biases through the text-to-image generation lens,” arXiv preprint arXiv:2304.06034, 2023.
- M. D’Incà, C. Tzelepis, I. Patras, and N. Sebe, “Improving fairness using vision-language driven image augmentation,” arXiv preprint arXiv:2311.01573, 2023.
- C. T. Teo, M. Abdollahzadeh, and N.-M. Cheung, “Fair generative models via transfer learning,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 37, no. 2, 2023, pp. 2429–2437.
- J. Rando, D. Paleka, D. Lindner, L. Heim, and F. Tramèr, “Red-teaming the stable diffusion safety filter,” arXiv preprint arXiv:2210.04610, 2022.
- J. Wang, Z. Yan, J. Lan, E. Bertino, and W. Pedrycz, “Trustguard: Gnn-based robust and explainable trust evaluation with dynamicity support,” IEEE Transactions on Dependable and Secure Computing, pp. 1–18, 2024.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.