LAB: Large-Scale Alignment for ChatBots (2403.01081v3)
Abstract: This work introduces LAB (Large-scale Alignment for chatBots), a novel methodology designed to overcome the scalability challenges in the instruction-tuning phase of LLM training. Leveraging a taxonomy-guided synthetic data generation process and a multi-phase tuning framework, LAB significantly reduces reliance on expensive human annotations and proprietary models like GPT-4. We demonstrate that LAB-trained models can achieve competitive performance across several benchmarks compared to models trained with traditional human-annotated or GPT-4 generated synthetic data. Thus offering a scalable, cost-effective solution for enhancing LLM capabilities and instruction-following behaviors without the drawbacks of catastrophic forgetting, marking a step forward in the efficient training of LLMs for a wide range of applications.
- Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge, March 2018.
- Training Verifiers to Solve Math Word Problems, November 2021.
- The false promise of imitating proprietary llms. arXiv preprint arXiv:2305.15717, 2023.
- Measuring massive multitask language understanding. In International Conference on Learning Representations, 2020.
- Measuring massive multitask language understanding, 2021.
- Mistral 7B, October 2023.
- Mixtral of Experts, January 2024.
- Synthetic data (almost) from scratch: Generalized instruction tuning for language models, 2024.
- Self-alignment with instruction backtranslation, 2023.
- The flan collection: Designing data and methods for effective instruction tuning, 2023.
- Orca 2: Teaching Small Language Models How to Reason. https://arxiv.org/abs/2311.11045v2, November 2023.
- Orca: Progressive learning from complex explanation traces of gpt-4, 2023.
- Training language models to follow instructions with human feedback, 2022.
- Direct preference optimization: Your language model is secretly a reward model, 2023.
- WinoGrande: An Adversarial Winograd Schema Challenge at Scale, November 2019.
- Learning to summarize from human feedback, 2022.
- Stanford alpaca: An instruction-following llama model. https://github.com/tatsu-lab/stanford_alpaca, 2023.
- Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288, 2023.
- Musique: Multihop questions via single-hop question composition, 2022.
- Zephyr: Direct Distillation of LM Alignment, October 2023.
- Self-Instruct: Aligning Language Models with Self-Generated Instructions, May 2023.
- Finetuned language models are zero-shot learners. In International Conference on Learning Representations.
- et al. Xiang Yue. Mammoth: Building math generalist models through hybrid instruction tuning. arXiv preprint arXiv:2309.05653, 2023.
- WizardLM: Empowering Large Language Models to Follow Complex Instructions, June 2023.
- Learning to mine aligned code and natural language pairs from stack overflow. In International Conference on Mining Software Repositories, MSR, pp. 476–486. ACM, 2018. doi: https://doi.org/10.1145/3196398.3196408.
- HellaSwag: Can a Machine Really Finish Your Sentence?, May 2019.
- Judging llm-as-a-judge with mt-bench and chatbot arena, 2023.
- Judging llm-as-a-judge with mt-bench and chatbot arena. Advances in Neural Information Processing Systems, 36, 2024.