Can LLMs get help from other LLMs without revealing private information? (2404.01041v2)
Abstract: Cascades are a common type of machine learning systems in which a large, remote model can be queried if a local model is not able to accurately label a user's data by itself. Serving stacks for LLMs increasingly use cascades due to their ability to preserve task performance while dramatically reducing inference costs. However, applying cascade systems in situations where the local model has access to sensitive data constitutes a significant privacy risk for users since such data could be forwarded to the remote model. In this work, we show the feasibility of applying cascade systems in such setups by equipping the local model with privacy-preserving techniques that reduce the risk of leaking private information when querying the remote model. To quantify information leakage in such setups, we introduce two privacy measures. We then propose a system that leverages the recently introduced social learning paradigm in which LLMs collaboratively learn from each other by exchanging natural language. Using this paradigm, we demonstrate on several datasets that our methods minimize the privacy loss while at the same time improving task performance compared to a non-cascade baseline.
- Social learning theory, volume 1. Englewood cliffs Prentice Hall, 1977.
- Language models are few-shot learners, 2020.
- Frugalgpt: How to use large language models while reducing cost and improving performance, 2023.
- Training verifiers to solve math word problems. arXiv preprint arXiv:2110.14168, 2021.
- Senteval: An evaluation toolkit for universal sentence representations. arXiv preprint arXiv:1803.05449, 2018.
- Cynthia Dwork. Differential privacy. In International colloquium on automata, languages, and programming, pp. 1–12. Springer, 2006.
- Limiting privacy breaches in privacy preserving data mining. In Proceedings of the twenty-second ACM SIGMOD-SIGACT-SIGART symposium on Principles of database systems, pp. 211–222, 2003.
- Google. Gemini: a family of highly capable multimodal models. arXiv preprint arXiv:2312.11805, 2023.
- Training compute-optimal large language models, 2022.
- What can we learn privately? SIAM Journal on Computing, 40(3):793–826, 2011.
- Rlaif: Scaling reinforcement learning from human feedback with ai feedback, 2023.
- Cascadebert: Accelerating inference of pre-trained language models via calibrated complete models cascade, 2021.
- Anonymisation models for text data: State of the art, challenges and future directions. In Chengqing Zong, Fei Xia, Wenjie Li, and Roberto Navigli (eds.), Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 4188–4203, Online, August 2021. Association for Computational Linguistics. doi: 10.18653/v1/2021.acl-long.323. URL https://aclanthology.org/2021.acl-long.323.
- Prompt injection attack against llm-integrated applications, 2024.
- Towards efficient generative large language model serving: A survey from algorithms to systems, 2023.
- Social learning: Towards collaborative learning with large language models. arXiv preprint arXiv:2312.11441, 2023.
- Robust de-anonymization of large sparse datasets. In 2008 IEEE Symposium on Security and Privacy (sp 2008), pp. 111–125. IEEE, 2008.
- Helen Nissenbaum. Privacy as contextual integrity. Wash. L. Rev., 79:119, 2004.
- OpenAI. GPT-4 Technical Report, 2023.
- Are emergent abilities of large language models a mirage?, 2023.
- Bleurt: Learning robust metrics for text generation. arXiv preprint arXiv:2004.04696, 2020.
- Synthetic prompting: Generating chain-of-thought demonstrations for large language models. In International Conference on Machine Learning, pp. 30706–30775. PMLR, 2023.
- Beyond the imitation game: Quantifying and extrapolating the capabilities of language models. arXiv preprint arXiv:2206.04615, 2022.
- Latanya Sweeney. k-anonymity: A model for protecting privacy. International journal of uncertainty, fuzziness and knowledge-based systems, 10(05):557–570, 2002.
- Jörg Tiedemann. The tatoeba translation challenge–realistic data sets for low resource and multilingual mt. arXiv preprint arXiv:2010.06354, 2020.
- Recovering from privacy-preserving masking with large language models, 2023.
- Self-instruct: Aligning language models with self-generated instructions, 2023.
- Stanley L Warner. Randomized response: A survey technique for eliminating evasive answer bias. Journal of the American Statistical Association, 60(309):63–69, 1965.
- Privacy-preserving in-context learning for large language models. In The Twelfth International Conference on Learning Representations, 2023.
- Large language model cascades with mixture of thoughts representations for cost-efficient reasoning, 2024.