L-TUNING: Synchronized Label Tuning for Prompt and Prefix in LLMs (2402.01643v2)
Abstract: Efficiently fine-tuning LLMs for specific tasks presents a considerable challenge in natural language processing. Traditional methods, like prompt or prefix tuning, typically rely on arbitrary tokens for training, leading to prolonged training times and generalized token use across various class labels. To address these issues, this paper introduces L-Tuning, an efficient fine-tuning approach designed for classification tasks within the Natural Language Inference (NLI) framework. Diverging from conventional methods, L-Tuning focuses on the fine-tuning of label tokens processed through a pre-trained LLM, thereby harnessing its pre-existing semantic knowledge. This technique not only improves the fine-tuning accuracy and efficiency but also facilitates the generation of distinct label embeddings for each class, enhancing the model's training nuance. Our experimental results indicate a significant improvement in training efficiency and classification accuracy with L-Tuning compared to traditional approaches, marking a promising advancement in fine-tuning LLMs for complex language tasks.
- BERT: Pre-training of deep bidirectional transformers for language understanding. In Jill Burstein, Christy Doran, and Thamar Solorio (eds.), Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 4171–4186, Minneapolis, Minnesota, June 2019. Association for Computational Linguistics. doi: 10.18653/v1/N19-1423. URL https://aclanthology.org/N19-1423.
- Openagi: When llm meets domain experts. arXiv preprint arXiv:2304.04370, 2023.
- Ppt: Pre-trained prompt tuning for few-shot learning. arXiv preprint arXiv:2109.04332, 2021.
- Ptr: Prompt tuning with rules for text classification. AI Open, 3:182–192, 2022.
- Deberta: Decoding-enhanced bert with disentangled attention, 2021.
- Contrastive learning for universal zero-shot nli with cross-lingual sentence embeddings. In Proceedings of the 3rd Workshop on Multi-lingual Representation Learning (MRL), pp. 239–252, 2023.
- The power of scale for parameter-efficient prompt tuning. arXiv preprint arXiv:2104.08691, 2021.
- P-tuning: Prompt tuning can be comparable to fine-tuning across scales and tasks. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pp. 61–68, 2022.
- Roberta: A robustly optimized bert pretraining approach, 2019.
- The refinedweb dataset for falcon llm: Outperforming curated corpora with web data, and web data only, 2023.
- Soft prompt tuning for augmenting dense retrieval with large language models. arXiv preprint arXiv:2307.08303, 2023.
- Self-attention encoding and pooling for speaker recognition. arXiv preprint arXiv:2008.01077, 2020.
- Llama 2: Open foundation and fine-tuned chat models, 2023.
- GLUE: A multi-task benchmark and analysis platform for natural language understanding. 2019. In the Proceedings of ICLR.
- Bloom: A 176b-parameter open-access multilingual language model, 2023.
- Md. Shohanur Islam Sobuj (5 papers)
- Asif Mahmud (2 papers)
- Nusrat Jahan Prottasha (12 papers)
- Prakash Bhat (6 papers)
- Md. Kowsher (3 papers)