Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning Fine-Grained Grounded Citations for Attributed Large Language Models (2408.04568v1)

Published 8 Aug 2024 in cs.CL and cs.AI

Abstract: Despite the impressive performance on information-seeking tasks, LLMs still struggle with hallucinations. Attributed LLMs, which augment generated text with in-line citations, have shown potential in mitigating hallucinations and improving verifiability. However, current approaches suffer from suboptimal citation quality due to their reliance on in-context learning. Furthermore, the practice of citing only coarse document identifiers makes it challenging for users to perform fine-grained verification. In this work, we introduce FRONT, a training framework designed to teach LLMs to generate Fine-Grained Grounded Citations. By grounding model outputs in fine-grained supporting quotes, these quotes guide the generation of grounded and consistent responses, not only improving citation quality but also facilitating fine-grained verification. Experiments on the ALCE benchmark demonstrate the efficacy of FRONT in generating superior grounded responses and highly supportive citations. With LLaMA-2-7B, the framework significantly outperforms all the baselines, achieving an average of 14.21% improvement in citation quality across all datasets, even surpassing ChatGPT.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (11)
  1. Lei Huang (175 papers)
  2. Xiaocheng Feng (54 papers)
  3. Weitao Ma (12 papers)
  4. Yuxuan Gu (17 papers)
  5. Weihong Zhong (15 papers)
  6. Xiachong Feng (28 papers)
  7. Weijiang Yu (23 papers)
  8. Weihua Peng (12 papers)
  9. Duyu Tang (65 papers)
  10. Dandan Tu (25 papers)
  11. Bing Qin (186 papers)
Citations (2)
X Twitter Logo Streamline Icon: https://streamlinehq.com