Becoming self-instruct: introducing early stopping criteria for minimal instruct tuning (2307.03692v1)
Abstract: In this paper, we introduce the Instruction Following Score (IFS), a metric that detects LLMs' ability to follow instructions. The metric has a dual purpose. First, IFS can be used to distinguish between base and instruct models. We benchmark publicly available base and instruct models, and show that the ratio of well formatted responses to partial and full sentences can be an effective measure between those two model classes. Secondly, the metric can be used as an early stopping criteria for instruct tuning. We compute IFS for Supervised Fine-Tuning (SFT) of 7B and 13B LLaMA models, showing that models learn to follow instructions relatively early in the training process, and the further finetuning can result in changes in the underlying base model semantics. As an example of semantics change we show the objectivity of model predictions, as defined by an auxiliary metric ObjecQA. We show that in this particular case, semantic changes are the steepest when the IFS tends to plateau. We hope that decomposing instruct tuning into IFS and semantic factors starts a new trend in better controllable instruct tuning and opens possibilities for designing minimal instruct interfaces querying foundation models.
- “Stanford Alpaca: An Instruction-following LLaMA model” In GitHub repository GitHub, https://github.com/tatsu-lab/stanford_alpaca, 2023
- “Self-Instruct: Aligning Language Models with Self-Generated Instructions”, 2023 arXiv:2212.10560 [cs.CL]
- “The Flan Collection: Designing Data and Methods for Effective Instruction Tuning”, 2023 arXiv:2301.13688 [cs.AI]
- “LIMA: Less Is More for Alignment”, 2023 arXiv:2305.11206 [cs.CL]
- “GPT4All: Training an Assistant-style Chatbot with Large Scale Data Distillation from GPT-3.5-Turbo” In GitHub repository GitHub, https://github.com/nomic-ai/gpt4all, 2023
- “LLaMA: Open and Efficient Foundation Language Models”, 2023 arXiv:2302.13971 [cs.CL]
- “OPT: Open Pre-trained Transformer Language Models”, 2022 arXiv:2205.01068 [cs.CL]
- “The Pile: An 800GB Dataset of Diverse Text for Language Modeling” In arXiv preprint arXiv:2101.00027, 2020
- Writer “Palmyra LLMs empower secure, enterprise-grade generative AI for business. Writer Blog”, 2023 URL: https://writer.com/blog/palmyra/
- “The False Promise of Imitating Proprietary LLMs”, 2023 arXiv:2305.15717 [cs.CL]
- OpenAI “ChatGPT: Optimizing language models for dialogue.”, 2022 URL: https://online-chatgpt.com/
- Sundar Pichai “An important next step on our AI journey. Google AI Blog”, 2023 URL: https://blog.google/intl/en-africa/products/explore-get-answers/an-important-next-step-on-our-ai-journey/
- AnthropicAI “Introducing Claude”, 2023 URL: https://www.anthropic.com/index/introducing-claude
- Geoffrey Hinton, Oriol Vinyals and Jeff Dean “Distilling the Knowledge in a Neural Network”, 2015 arXiv:1503.02531 [stat.ML]
- “Holistic Evaluation of Language Models”, 2022 arXiv:2211.09110 [cs.CL]
- “Natural Questions: a Benchmark for Question Answering Research” In Transactions of the Association of Computational Linguistics, 2019
- Huggingface “Open LLM Leaderboard” Accessed: 2023-06-10, 2023 URL: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
- “A framework for few-shot language model evaluation” Zenodo, 2021 DOI: 10.5281/zenodo.5371628
- “Maybe Only 0.5% Data is Needed: A Preliminary Exploration of Low Training Data Instruction Tuning”, 2023 arXiv:2305.09246 [cs.AI]
- “Fine-Tuning can Distort Pretrained Features and Underperform Out-of-Distribution”, 2022 arXiv:2202.10054 [cs.LG]
- “Language Models are Few-Shot Learners”, 2020 arXiv:2005.14165 [cs.CL]
- “Language Models are Unsupervised Multitask Learners”, 2018 URL: https://d4mucfpksywv.cloudfront.net/better-language-models/language-models.pdf
- “Training language models to follow instructions with human feedback”, 2022 arXiv:2203.02155 [cs.CL]
- “Proximal Policy Optimization Algorithms”, 2017 arXiv:1707.06347 [cs.LG]
- “Measuring Massive Multitask Language Understanding”, 2021 arXiv:2009.03300 [cs.CY]
- “OpenAssistant Conversations – Democratizing Large Language Model Alignment”, 2023 arXiv:2304.07327 [cs.CL]
- Huggingface “AutoTrain: Create powerful AI models without code”, 2023 URL: https://huggingface.co/autotrain
- “Emergent Abilities of Large Language Models”, 2022 arXiv:2206.07682 [cs.CL]
Collections
Sign up for free to add this paper to one or more collections.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.