Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Making Long-Context Language Models Better Multi-Hop Reasoners (2408.03246v1)

Published 6 Aug 2024 in cs.CL

Abstract: Recent advancements in long-context modeling have enhanced LLMs (LMs) for complex tasks across multiple NLP applications. Despite this progress, we find that these models struggle with multi-hop reasoning and exhibit decreased performance in the presence of noisy contexts. In this paper, we introduce Reasoning with Attributions, a novel approach that prompts LMs to supply attributions for each assertion during their reasoning. We validate our approach through experiments on three multi-hop datasets, employing both proprietary and open-source models, and demonstrate its efficacy and resilience. Furthermore, we explore methods to augment reasoning capabilities via fine-tuning and offer an attribution-annotated dataset and a specialized training strategy. Our fine-tuned model achieves competitive performance on multi-hop reasoning benchmarks, closely paralleling proprietary LMs such as ChatGPT and Claude-instant.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Yanyang Li (22 papers)
  2. Shuo Liang (3 papers)
  3. Michael R. Lyu (176 papers)
  4. Liwei Wang (239 papers)
Citations (3)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets