Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Evaluation of RAG Metrics for Question Answering in the Telecom Domain (2407.12873v1)

Published 15 Jul 2024 in cs.CL, cs.IR, cs.LG, and cs.AI

Abstract: Retrieval Augmented Generation (RAG) is widely used to enable LLMs perform Question Answering (QA) tasks in various domains. However, RAG based on open-source LLM for specialized domains has challenges of evaluating generated responses. A popular framework in the literature is the RAG Assessment (RAGAS), a publicly available library which uses LLMs for evaluation. One disadvantage of RAGAS is the lack of details of derivation of numerical value of the evaluation metrics. One of the outcomes of this work is a modified version of this package for few metrics (faithfulness, context relevance, answer relevance, answer correctness, answer similarity and factual correctness) through which we provide the intermediate outputs of the prompts by using any LLMs. Next, we analyse the expert evaluations of the output of the modified RAGAS package and observe the challenges of using it in the telecom domain. We also study the effect of the metrics under correct vs. wrong retrieval and observe that few of the metrics have higher values for correct retrieval. We also study for differences in metrics between base embeddings and those domain adapted via pre-training and fine-tuning. Finally, we comment on the suitability and challenges of using these metrics for in-the-wild telecom QA task.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Sujoy Roychowdhury (9 papers)
  2. Sumit Soman (18 papers)
  3. Neeraj Gunda (3 papers)
  4. Vansh Chhabra (2 papers)
  5. Sai Krishna Bala (3 papers)
  6. H G Ranjani (1 paper)
Citations (6)
X Twitter Logo Streamline Icon: https://streamlinehq.com