FewFedPIT: Towards Privacy-preserving and Few-shot Federated Instruction Tuning (2403.06131v2)
Abstract: Instruction tuning has been identified as a crucial technique for optimizing the performance of LLMs in generating human-aligned responses. Nonetheless, gathering diversified and superior-quality instruction data for such tuning presents notable obstacles, especially in domains with rigid privacy provisions. Federated instruction tuning (FedIT) has emerged as a promising solution, by consolidating collaborative training across multiple data owners, thereby resulting in a privacy-preserving learning model. However, FedIT encounters limitations such as scarcity of instructional data and risk of exposure to training data extraction attacks. In this paper, we propose a novel federated algorithm, FewFedPIT, designed to simultaneously enhance privacy protection and model performance of federated few-shot learning. FewFedPITcomprises three vital components on the client side: (1) synthetic data generation, which utilizes LLMs' in-context learning capacity to generate synthetic data autonomously, thus expanding the local database; (2) parameter isolation training, which individually updates the public parameters in the synthetic data and the private parameters in the local data, consequently mitigating the noise impact of the synthetic data; (3) local aggregation sharing, which mixes public and private parameters before uploading, effectively preventing data extraction attacks. Extensive experiments on three open-source datasets demonstrate the effectiveness of FewFedPITin, enhancing privacy preservation and improving federated few-shot performance.
- Pythia: A suite for analyzing large language models across training and scaling. In International Conference on Machine Learning, pages 2397–2430. PMLR.
- Language models are realistic tabular data generators. arXiv preprint arXiv:2210.06280.
- What does it mean for a language model to preserve privacy? In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency, pages 2280–2292.
- Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901.
- Extracting training data from large language models. In 30th USENIX Security Symposium (USENIX Security 21), pages 2633–2650.
- Alpagasus: Training a better alpaca with fewer data. arXiv preprint arXiv:2307.08701.
- Fewfedweight: Few-shot federated learning framework across multiple nlp tasks. arXiv preprint arXiv:2212.08354.
- Alpacafarm: A simulation framework for methods that learn from human feedback. arXiv preprint arXiv:2305.14387.
- Fate-llm: A industrial grade federated learning framework for large language models. arXiv preprint arXiv:2310.10049.
- A theoretical analysis of the repetition problem in text generation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 12848–12856.
- The pile: An 800gb dataset of diverse text for language modeling. arXiv preprint arXiv:2101.00027.
- Measuring the effects of non-identical data distribution for federated visual classification. arXiv preprint arXiv:1909.06335.
- Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685.
- Federated learning: Strategies for improving communication efficiency. arXiv preprint arXiv:1610.05492.
- Federatedscope-llm: A comprehensive package for fine-tuning large language models in federated learning. arXiv preprint arXiv:2309.00363.
- From quantity to quality: Boosting llm performance with self-guided data selection for instruction tuning. arXiv preprint arXiv:2308.12032.
- Self-alignment with instruction backtranslation. arXiv preprint arXiv:2308.06259.
- Chatdoctor: A medical chat model fine-tuned on a large language model meta-ai (llama) using medical domain knowledge. Cureus, 15(6).
- Fednlp: Benchmarking federated learning methods for natural language processing tasks. arXiv preprint arXiv:2104.08815.
- Chin-Yew Lin. 2004. Rouge: A package for automatic evaluation of summaries. In Text summarization branches out, pages 74–81.
- Communication-efficient learning of deep networks from decentralized data. In Artificial intelligence and statistics, pages 1273–1282. PMLR.
- Scalable extraction of training data from (production) language models. arXiv preprint arXiv:2311.17035.
- R OpenAI. 2023. Gpt-4 technical report. arxiv 2303.08774. View in Article, 2:13.
- Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35:27730–27744.
- West-of-n: Synthetic preference generation for improved reward modeling. arXiv preprint arXiv:2401.12086.
- The RefinedWeb dataset for Falcon LLM: outperforming curated corpora with web data, and web data only. arXiv preprint arXiv:2306.01116.
- Matt Post. 2018. A call for clarity in reporting bleu scores. arXiv preprint arXiv:1804.08771.
- A split-and-privatize framework for large language model fine-tuning. arXiv preprint arXiv:2312.15603.
- Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288.
- Large language models are not fair evaluators. arXiv preprint arXiv:2305.17926.
- Self-instruct: Aligning language model with self generated instructions. arXiv preprint arXiv:2212.10560.
- Super-naturalinstructions: Generalization via declarative instructions on 1600+ nlp tasks. arXiv preprint arXiv:2204.07705.
- Openfedllm: Training large language models on decentralized private data via federated learning. arXiv preprint arXiv:2402.06954.
- Self-rewarding language models. arXiv preprint arXiv:2401.10020.
- Towards building the federated gpt: Federated instruction tuning. arXiv preprint arXiv:2305.05644.
- Instruction tuning for large language models: A survey. arXiv preprint arXiv:2308.10792.
- Alpacare: Instruction-tuned large language models for medical application. arXiv preprint arXiv:2310.14558.
- Fedlegal: The first real-world federated learning benchmark for legal nlp. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 3492–3507.
- Fedpetuning: When federated learning meets the parameter-efficient tuning methods of pre-trained language models. In Annual Meeting of the Association of Computational Linguistics 2023, pages 9963–9977. Association for Computational Linguistics (ACL).
- Fedprompt: Communication-efficient and privacy-preserving prompt tuning in federated learning. In ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 1–5. IEEE.
- Judging llm-as-a-judge with mt-bench and chatbot arena. Advances in Neural Information Processing Systems, 36.
- Zhuo Zhang (42 papers)
- Jingyuan Zhang (50 papers)
- Jintao Huang (12 papers)
- Lizhen Qu (68 papers)
- Hongzhi Zhang (33 papers)
- Zenglin Xu (145 papers)
- Qifan Wang (129 papers)
- Xun Zhou (62 papers)