LoraMap: Harnessing the Power of LoRA Connections (2408.16264v2)
Abstract: Fact-checking techniques can mitigate hallucinations in LLMs, a prominent issue in specialized domains. As parameter-efficient techniques such as Low-Rank Adaptation (LoRA) can overcome substantial computational overhead, some studies have explored the integration of multiple LoRAs. While previous studies focus on parallel integration, this paper investigates methods to establish connections among multiple LoRAs. We create three reasoning datasets tailored to fact-checking and fine-tune individual LoRAs, allowing them to view and reason from diverse perspectives. Then, we explore strategies for allocating these reasoning LoRAs and introduce LoraMap, an approach to map connections between them. The results of the fact-checking task demonstrate that the performance of LoraMap is superior to LoraHub, an existing method for integrating LoRAs. LoraMap also outperforms with significantly fewer trainable parameters than LoraConcat, which concatenates LoRAs and further fine-tunes them.
- Satanjeev Banerjee and Alon Lavie. 2005. Meteor: An automatic metric for mt evaluation with improved correlation with human judgments. In Proceedings of the acl workshop on intrinsic and extrinsic evaluation measures for machine translation and/or summarization, pages 65–72, Prague, Czech Republic. Association for Computational Linguistics.
- Longformer: The long-document transformer. Computing Research Repository, arXiv:2004.05150. Version 2.
- Automated fact-checking in dialogue: Are specialized models needed? In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 16009–16020, Singapore. Association for Computational Linguistics.
- Factool: Factuality detection in generative ai–a tool augmented framework for multi-task and multi-domain scenarios. Computing Research Repository, arXiv:2307.13528.
- Palm: Scaling language modeling with pathways. Journal of Machine Learning Research, 24(240):1–113.
- Scaling instruction-finetuned language models. Computing Research Repository, arXiv:2210.11416.
- Llm.int8(): 8-bit matrix multiplication for transformers at scale. In Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022.
- Qlora: Efficient finetuning of quantized llms. Advances in Neural Information Processing Systems, 36.
- Loramoe: Revolutionizing mixture of experts for maintaining world knowledge in language model alignment. Computing Research Repository, arXiv:2312.09979.
- Improving factuality and reasoning in language models through multiagent debate. Computing Research Repository, arXiv:2305.14325.
- Higher layers need more lora experts. Computing Research Repository, arXiv:2402.08562.
- A survey on automated fact-checking. Transactions of the Association for Computational Linguistics, 10:178–206.
- Dialfact: A benchmark for fact-checking in dialogue. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 3785–3801, Dublin, Ireland. Association for Computational Linguistics.
- Parameter-efficient transfer learning for nlp. In International Conference on Machine Learning, pages 2790–2799. PMLR.
- Lora: Low-rank adaptation of large language models. In International Conference on Learning Representations, Online. Association for Computational Linguistics.
- Lorahub: Efficient cross-task generalization via dynamic lora composition. volume arXiv:2307.13269.
- Survey of hallucination in natural language generation. ACM Computing Surveys, 55(12):1–38.
- Neema Kotonya and Francesca Toni. 2020. Explainable automated fact-checking for public health claims. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 7740–7754, Online. Association for Computational Linguistics.
- Llms as factual reasoners: Insights from existing benchmarks and beyond. Computing Research Repository, arXiv:2305.14540.
- The power of scale for parameter-efficient prompt tuning. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 3045–3059. Association for Computational Linguistics.
- Mixlora: Enhancing large language models fine-tuning with lora-based mixture of experts. Computing Research Repository, arXiv:2404.15159.
- Self-checker: Plug-and-play modules for fact-checking with large language models. Computing Research Repository, arXiv:2305.14623.
- Xiang Lisa Li and Percy Liang. 2021. Prefix-tuning: Optimizing continuous prompts for generation. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 4582–4597. Association for Computational Linguistics.
- Chin-Yew Lin. 2004. Rouge: A package for automatic evaluation of summaries. Text Summarization Branches Out, Association for Computational Linguistics, pages 74–81.
- Chin-Yew Lin and Franz Josef Och. 2004. Automatic evaluation of machine translation quality using longest common subsequence and skip-bigram statistics. In Proceedings of the 42nd Annual Meeting of the Association for Computational Linguistics (ACL-04), pages 605–612, Barcelona, Spain. Association for Computational Linguistics.
- Moelora: An moe-based parameter efficient fine-tuning method for multi-task medical applications. Computing Research Repository, arXiv:2310.18339.
- Factscore: Fine-grained atomic evaluation of factual precision in long form text generation. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 12076–12100. Association for Computational Linguistics.
- OpenAI. 2023. Gpt-4 technical report. Computing Research Repository, arXiv:2303.08774.
- Qacheck: A demonstration system for question-guided multi-hop fact-checking. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 264–273. Association for Computational Linguistics.
- Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics, pages 311–318, Philadelphia, Pennsylvania. Association for Computational Linguistics.
- Adapterhub: A framework for adapting transformers. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 46–54. Association for Computational Linguistics.
- Covid-fact: Fact extraction and verification of real-world claims on covid-19 pandemic. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, ACL/IJCNLP 2021, (Volume 1:Long Papers), pages 2116–2129, Online. Association for Computational Linguistics.
- Evidence-based fact-checking of health-related claims. In In Findings of the Association for Computational Linguistics: EMNLP 2021, pages 3499–3512, Punta Cana, Dominican Republic. Association for Computational Linguistics.
- Fever: a large-scale dataset for fact extraction and verification. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 809–819, New Orleans, Louisiana. Association for Computational Linguistics.
- Llama: Open and efficient foundation language models. Computing Research Repository, arXiv:2302.13971.
- Fact or fiction: Verifying scientific claims. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 7534–7550, Online. Association for Computational Linguistics.
- Multivers: Improving scientific claim verification with weak supervision and full-document context. In Findings of the Association for Computational Linguistics: NAACL 2022, pages 61–76. Association for Computational Linguistics.
- Survey on factuality in large language models: Knowledge, retrieval and domain-specificity. Computing Research Repository, arXiv:2310.07521.
- Factcheck-bench: Fine-grained evaluation benchmark for automatic fact-checkers. Computing Research Repository, arXiv:2311.09000.
- Chain-of-thought prompting elicits reasoning in large language models. volume 35, pages 24824–24837, New Orleans, Louisiana, United States of America. NeurIPS.
- Bertscore: Evaluating text generation with bert. International Conference on Learning Representations.
- Xuan Zhang and Wei Gao. 2023. Towards llm-based fact verification on news claims with a hierarchical step-by-step prompting method. In Proceedings of the 13th International Joint Conference on Natural Language Processing and the 3rd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics (Volume 1: Long Papers), pages 996–1011. Association for Computational Linguistics.
Collections
Sign up for free to add this paper to one or more collections.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.