Perplexity-Trap: PLM-Based Retrievers Overrate Low Perplexity Documents
The paper "Perplexity-Trap: PLM-Based Retrievers Overrate Low Perplexity Documents" investigates the phenomenon that LLM generated content is often given higher relevance scores by Pretrained LLM (PLM) based retrieval systems, even when its semantic quality is on par with human-written content. This inclination towards LLM-generated content is referred to as source bias and poses a risk to the integrity of the information access ecosystem. The authors hypothesize that PLM-based retrievers prioritize documents with lower perplexity scores, which are more commonly associated with LLM-rewritten documents.
The research introduces a compelling causal graph-based framework to diagnose and separate the causal effect of document perplexity from estimated relevance, arguing that this separation is crucial to understanding the roots of source bias in these retrievers. The paper theorizes that this bias is due to the alignment between the gradients of loss functions from language modeling and retrieval tasks. The authors propose a solution through the Causal Diagnosis and Correction (CDC) method, which attempts to separate the bias induced by perplexity from overall relevance scores. This inference-time debiasing approach is shown to effectively adjust for this bias without compromising the rank quality across domains.
In empirical terms, the research demonstrates the prevalence of the causal effect of perplexity over various PLM-based retrievers on distinct domain datasets using a robust two-stage regression process. It is revealed that the bias is statistically significant, although relatively modest, indicating a systematic tilt towards content with lower perplexity values. These findings endorse the hypothesis that perplexity, unrelated to purely semantic matching, is an unintended factor affecting retrieval relevance scores.
From a theoretical perspective, the paper thoroughly analyzes the interplay between retrieval and language modeling tasks. It explicates how the linear relationship between their gradient structures contributes to the biased retrieval process. This demonstrates that improving a retriever’s language modeling capacity could inadvertently heighten its bias towards low-perplexity documents, thus resulting in a trade-off between retrieval efficacy and exacerbation of source bias.
Practically, these findings suggest a predictable compromise for retrieval systems designed using PLM architectures: enhanced language modeling abilities that improve ranking performance may simultaneously intensify source bias. The paper’s proposed CDC approach signifies a crucial development, offering a promising pathway to mitigate such bias while preserving, or even enhancing, ranking performance.
Looking forward, this research opens avenues for further investigation into other non-causal features that might influence retrieval biases and how similar causal frameworks can be employed to mitigate them. The study also raises questions about the extent to which retrieval systems should debias towards human-written content, balancing information quality and content diversity. As LLMs continue to proliferate, addressing source bias presents both a critical challenge and an opportunity for optimizing information retrieval systems.