Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

How Well Do Large Language Models Truly Ground? (2311.09069v2)

Published 15 Nov 2023 in cs.CL and cs.AI

Abstract: To reduce issues like hallucinations and lack of control in LLMs, a common method is to generate responses by grounding on external contexts given as input, known as knowledge-augmented models. However, previous research often narrowly defines "grounding" as just having the correct answer, which does not ensure the reliability of the entire response. To overcome this, we propose a stricter definition of grounding: a model is truly grounded if it (1) fully utilizes the necessary knowledge from the provided context, and (2) stays within the limits of that knowledge. We introduce a new dataset and a grounding metric to evaluate model capability under the definition. We perform experiments across 25 LLMs of different sizes and training methods and provide insights into factors that influence grounding performance. Our findings contribute to a better understanding of how to improve grounding capabilities and suggest an area of improvement toward more reliable and controllable LLM applications.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Hyunji Lee (19 papers)
  2. Sejune Joo (4 papers)
  3. Chaeeun Kim (5 papers)
  4. Joel Jang (30 papers)
  5. Doyoung Kim (19 papers)
  6. Kyoung-Woon On (19 papers)
  7. Minjoon Seo (82 papers)
Citations (3)