Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Decorrelate Irrelevant, Purify Relevant: Overcome Textual Spurious Correlations from a Feature Perspective (2202.08048v2)

Published 16 Feb 2022 in cs.CL, cs.AI, cs.IT, and math.IT

Abstract: Natural language understanding (NLU) models tend to rely on spurious correlations (i.e., dataset bias) to achieve high performance on in-distribution datasets but poor performance on out-of-distribution ones. Most of the existing debiasing methods often identify and weaken these samples with biased features (i.e., superficial surface features that cause such spurious correlations). However, down-weighting these samples obstructs the model in learning from the non-biased parts of these samples. To tackle this challenge, in this paper, we propose to eliminate spurious correlations in a fine-grained manner from a feature space perspective. Specifically, we introduce Random Fourier Features and weighted re-sampling to decorrelate the dependencies between features to mitigate spurious correlations. After obtaining decorrelated features, we further design a mutual-information-based method to purify them, which forces the model to learn features that are more relevant to tasks. Extensive experiments on two well-studied NLU tasks demonstrate that our method is superior to other comparative approaches.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Shihan Dou (46 papers)
  2. Rui Zheng (79 papers)
  3. Ting Wu (31 papers)
  4. Junjie Shan (12 papers)
  5. Qi Zhang (785 papers)
  6. Yueming Wu (16 papers)
  7. Xuanjing Huang (287 papers)
  8. Songyang Gao (28 papers)
Citations (8)

Summary

We haven't generated a summary for this paper yet.