CI-Bench: Benchmarking Contextual Integrity of AI Assistants on Synthetic Data (2409.13903v1)
Abstract: Advances in generative AI point towards a new era of personalized applications that perform diverse tasks on behalf of users. While general AI assistants have yet to fully emerge, their potential to share personal data raises significant privacy challenges. This paper introduces CI-Bench, a comprehensive synthetic benchmark for evaluating the ability of AI assistants to protect personal information during model inference. Leveraging the Contextual Integrity framework, our benchmark enables systematic assessment of information flow across important context dimensions, including roles, information types, and transmission principles. We present a novel, scalable, multi-step synthetic data pipeline for generating natural communications, including dialogues and emails. Unlike previous work with smaller, narrowly focused evaluations, we present a novel, scalable, multi-step data pipeline that synthetically generates natural communications, including dialogues and emails, which we use to generate 44 thousand test samples across eight domains. Additionally, we formulate and evaluate a naive AI assistant to demonstrate the need for further study and careful training towards personal assistant tasks. We envision CI-Bench as a valuable tool for guiding future LLM development, deployment, system design, and dataset construction, ultimately contributing to the development of AI assistants that align with users' privacy expectations.
- Air gap: Protecting privacy-conscious conversational agents. arXiv [cs.CR].
- Constitutional AI: Harmlessness from AI Feedback. arXiv [cs.CL].
- Privacy and contextual integrity: framework and applications. In 2006 IEEE Symposium on Security and Privacy (S&P’06). IEEE.
- Extracting Training Data from Large Language Models. arXiv [cs.CR].
- Extending context window of large language models via positional interpolation. arXiv preprint arXiv:2306.15595.
- Fischer, K. A. 2023. Reflective linguistic programming (rlp): A stepping stone in socially-aware agi (socialagi). arXiv preprint arXiv:2305.12647.
- The ethics of advanced ai assistants. arXiv preprint arXiv:2404.16244.
- Gemini Team,. 2023. Gemini: a family of highly capable multimodal models. arXiv preprint arXiv:2312.11805.
- Operationalizing Contextual Integrity in Privacy-Conscious Assistants. arXiv preprint arXiv:2408.02373.
- Memory sandbox: Transparent and interactive memory management for conversational agents. In Adjunct Proceedings of the 36th Annual ACM Symposium on User Interface Software and Technology, 1–3.
- Likert scale: Explored and explained. British journal of applied science & technology.
- Pema: Plug-in external memory adaptation for language models. arXiv preprint arXiv:2311.08590.
- Aquilis: Using Contextual Integrity for Privacy Protection on Mobile Devices. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol., 4(4): 1–28.
- Can LLMs keep a secret? Testing privacy implications of language models via contextual integrity theory. arXiv [cs.AI].
- Simplyretrieve: A private and lightweight retrieval-centric generative ai tool. arXiv preprint arXiv:2308.03983.
- Nissenbaum, H. 2004. Privacy as Contextual Integrity. Wash Law Rev., 79(1): 119.
- Nissenbaum, H. 2009. Privacy in Context. Stanford University Press.
- Training language models to follow instructions with human feedback. Adv. Neural Inf. Process. Syst., abs/2203.02155.
- Yarn: Efficient context window extension of large language models. arXiv preprint arXiv:2309.00071.
- Sayplan: Grounding large language models using 3d scene graphs for scalable robot task planning. In 7th Annual Conference on Robot Learning.
- VACCINE: Using Contextual Integrity For Data Leakage Detection. In The World Wide Web Conference, WWW ’19, 1702–1712. New York, NY, USA: Association for Computing Machinery.
- Trustllm: Trustworthiness in large language models. arXiv preprint arXiv:2401.05561.
- DecodingTrust: A comprehensive assessment of trustworthiness in GPT models. arXiv [cs.CL].
- A survey on large language model based autonomous agents. Frontiers of Computer Science, 18(6): 1–26.
- Aligning large Language Models with human: A survey. arXiv [cs.CL].
- Memorizing Transformers. arXiv [cs.LG].
- Harnessing the Power of LLMs in Practice: A Survey on ChatGPT and Beyond. arXiv [cs.CL].
Collections
Sign up for free to add this paper to one or more collections.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.