Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Mitigating the Influence of Distractor Tasks in LMs with Prior-Aware Decoding (2401.17692v3)

Published 31 Jan 2024 in cs.CL

Abstract: The broad capabilities of LLMs (LMs) can be limited by their sensitivity to distractor tasks: LMs can infer secondary tasks from the prompt in addition to the intended one, leading to unwanted outputs. For example, prompt injection attacks can cause models to deviate from explicit directives. In some 'inverse scaling' cases, this unwanted behaviour actually worsens as models scale up to at least 540B parameters. We present a theoretical framework that interprets LMs as a product of experts that combine multiple data generation processes. Based on this framework, we demonstrate prior-aware decoding (PAD) - a simple contrastive inference method to reduce the influence of distractor tasks. We apply PAD to eleven models, across four datasets, and find improvements in 41 out of 44 task-model combinations, with a median increase in task completion proportion of 40%. The results suggest a promising direction for further development towards more reliable LLMs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Raymond Douglas (4 papers)
  2. Andis Draguns (8 papers)
  3. Tomáš Gavenčiak (13 papers)
Citations (1)
X Twitter Logo Streamline Icon: https://streamlinehq.com