Retrieval Augmented Generation for Domain-specific Question Answering (2404.14760v2)
Abstract: Question answering (QA) has become an important application in the advanced development of LLMs. General pre-trained LLMs for question-answering are not trained to properly understand the knowledge or terminology for a specific domain, such as finance, healthcare, education, and customer service for a product. To better cater to domain-specific understanding, we build an in-house question-answering system for Adobe products. We propose a novel framework to compile a large question-answer database and develop the approach for retrieval-aware finetuning of a LLM. We showcase that fine-tuning the retriever leads to major improvements in the final generation. Our overall approach reduces hallucinations during generation while keeping in context the latest retrieval information for contextual grounding.
- Sanat Sharma (10 papers)
- David Seunghyun Yoon (3 papers)
- Franck Dernoncourt (161 papers)
- Dewang Sultania (3 papers)
- Karishma Bagga (2 papers)
- Mengjiao Zhang (4 papers)
- Trung Bui (79 papers)
- Varun Kotte (1 paper)