Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
134 tokens/sec
GPT-4o
10 tokens/sec
Gemini 2.5 Pro Pro
47 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Token-Efficient Leverage Learning in Large Language Models (2404.00914v1)

Published 1 Apr 2024 in cs.CL, cs.AI, and cs.LG

Abstract: LLMs have excelled in various tasks but perform better in high-resource scenarios, which presents challenges in low-resource scenarios. Data scarcity and the inherent difficulty of adapting LLMs to specific tasks compound the challenge. To address the twin hurdles, we introduce \textbf{Leverage Learning}. We present a streamlined implement of this methodology called Token-Efficient Leverage Learning (TELL). TELL showcases the potential of Leverage Learning, demonstrating effectiveness across various LLMs and low-resource tasks, ranging from $104$ to $106$ tokens. It reduces task data requirements by up to nearly an order of magnitude compared to conventional Supervised Fine-Tuning (SFT) while delivering competitive performance. With the same amount of task data, TELL leads in improving task performance compared to SFT. We discuss the mechanism of Leverage Learning, suggesting it aligns with quantization hypothesis and explore its promising potential through empirical testing.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (34)
  1. Better fine-tuning by reducing representational collapse. In Proc. of ICLR. OpenReview.net, 2021. URL https://openreview.net/forum?id=OQ08SN70M1V.
  2. Qwen technical report. 2023.
  3. Scaling instruction-finetuned language models. 2022.
  4. QLoRA: Efficient finetuning of quantized LLMs. 2023.
  5. How abilities in large language models are affected by supervised fine-tuning data composition. 2023.
  6. The unreasonable effectiveness of few-shot learning for machine translation. 2023a.
  7. The unreasonable effectiveness of few-shot learning for machine translation. 2023b.
  8. Google. Gemma-7b-it. https://huggingface.co/google/gemma-7b-it, 2024.
  9. Large language models are reasoning teachers. 2022.
  10. Lora: Low-rank adaptation of large language models, 2021.
  11. How to fine-tune models with few samples: Update, data augmentation, and test-time augmentation. 2022.
  12. Platypus: Quick, cheap, and powerful refinement of LLMs. 2023a.
  13. RLAIF: Scaling reinforcement learning from human feedback with AI feedback. 2023b.
  14. HaluEval: A large-scale hallucination evaluation benchmark for large language models. 2023.
  15. The quantization model of neural scaling. 2023.
  16. Can a suit of armor conduct electricity? a new dataset for open book question answering. In Proc. of EMNLP, pp.  2381–2391, Brussels, Belgium, 2018. Association for Computational Linguistics. doi: 10.18653/v1/D18-1260. URL https://aclanthology.org/D18-1260.
  17. Rethinking the role of demonstrations: What makes in-context learning work? In Proc. of EMNLP, pp.  11048–11064, Abu Dhabi, United Arab Emirates, 2022. Association for Computational Linguistics. URL https://aclanthology.org/2022.emnlp-main.759.
  18. OpenAI. Gpt-3.5 turbo. https://openai.com/, 2023.
  19. Learning compact metrics for MT. In Proc. of EMNLP, pp.  751–762, Online and Punta Cana, Dominican Republic, 2021. Association for Computational Linguistics. doi: 10.18653/v1/2021.emnlp-main.58. URL https://aclanthology.org/2021.emnlp-main.58.
  20. Direct preference optimization: Your language model is secretly a reward model. 2023.
  21. Are emergent abilities of large language models a mirage? 2023.
  22. Proximal policy optimization algorithms. 2017.
  23. A comparative study between full-parameter and LoRA-based fine-tuning on chinese instruction data for instruction following large language model. 2023.
  24. Stanford alpaca: An instruction-following llama model. https://github.com/tatsu-lab/stanford_alpaca, 2023.
  25. The ModelScope Team. Swift:scalable lightweight infrastructure for fine-tuning. https://github.com/modelscope/swift, 2024.
  26. Llama 2: Open foundation and fine-tuned chat models. 2023.
  27. Do prompt-based models really understand the meaning of their prompts? In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp.  2300–2344, Seattle, United States, 2022. Association for Computational Linguistics. doi: 10.18653/v1/2022.naacl-main.167. URL https://aclanthology.org/2022.naacl-main.167.
  28. Chain-of-thought prompting elicits reasoning in large language models. 2022.
  29. ChatHome: Development and evaluation of a domain-specific language model for home renovation. 2023.
  30. WaveCoder: Widespread and versatile enhanced instruction tuning with refined data generation. 2023.
  31. Revisiting few-sample BERT fine-tuning. In Proc. of ICLR. OpenReview.net, 2021. URL https://openreview.net/forum?id=cO1IH43yUF.
  32. Calibrate before use: Improving few-shot performance of language models. In Marina Meila and Tong Zhang (eds.), Proc. of ICML, volume 139 of Proceedings of Machine Learning Research, pp.  12697–12706. PMLR, 2021. URL http://proceedings.mlr.press/v139/zhao21c.html.
  33. AGIEval: A human-centric benchmark for evaluating foundation models. 2023.
  34. LIMA: Less is more for alignment. 2023.

Summary

We haven't generated a summary for this paper yet.