Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

An Analysis and Mitigation of the Reversal Curse (2311.07468v3)

Published 13 Nov 2023 in cs.CL, cs.AI, and cs.LG

Abstract: Recent research observed a noteworthy phenomenon in LLMs, referred to as the reversal curse.'' The reversal curse is that when dealing with two entities, denoted as $a$ and $b$, connected by their relation $R$ and its inverse $R^{-1}$, LLMs excel in handling sequences in the form of$aRb$,'' but encounter challenges when processing $bR^{-1}a$,'' whether in generation or comprehension. For instance, GPT-4 can accurately respond to the queryTom Cruise's mother is?'' with Mary Lee Pfeiffer,'' but it struggles to provide a satisfactory answer when askedMary Lee Pfeiffer's son is?'' In this paper, we undertake the first-ever study of how the reversal curse happens in LLMs. Our investigations reveal that the reversal curse can stem from the specific training objectives, which become particularly evident in the widespread use of next-token prediction within most causal LLMs. We hope this initial investigation can draw more attention to the reversal curse, as well as other underlying limitations in current LLMs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Ang Lv (19 papers)
  2. Kaiyi Zhang (11 papers)
  3. Shufang Xie (29 papers)
  4. Quan Tu (16 papers)
  5. Yuhan Chen (39 papers)
  6. Ji-Rong Wen (299 papers)
  7. Rui Yan (250 papers)
Citations (19)
Youtube Logo Streamline Icon: https://streamlinehq.com