Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Adaptive Token Biaser: Knowledge Editing via Biasing Key Entities (2406.12468v1)

Published 18 Jun 2024 in cs.CL and cs.AI

Abstract: The parametric knowledge memorized by LLMs becomes outdated quickly. In-context editing (ICE) is currently the most effective method for updating the knowledge of LLMs. Recent advancements involve enhancing ICE by modifying the decoding strategy, obviating the need for altering internal model structures or adjusting external prompts. However, this enhancement operates across the entire sequence generation, encompassing a plethora of non-critical tokens. In this work, we introduce $\textbf{A}$daptive $\textbf{T}$oken $\textbf{Bias}$er ($\textbf{ATBias}$), a new decoding technique designed to enhance ICE. It focuses on the tokens that are mostly related to knowledge during decoding, biasing their logits by matching key entities related to new and parametric knowledge. Experimental results show that ATBias significantly enhances ICE performance, achieving up to a 32.3% improvement over state-of-the-art ICE methods while incurring only half the latency. ATBias not only improves the knowledge editing capabilities of ICE but can also be widely applied to LLMs with negligible cost.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Baolong Bi (23 papers)
  2. Shenghua Liu (33 papers)
  3. Yiwei Wang (119 papers)
  4. Lingrui Mei (20 papers)
  5. Hongcheng Gao (28 papers)
  6. Yilong Xu (6 papers)
  7. Xueqi Cheng (274 papers)
Citations (7)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets