Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Boosting Few-Shot Text Classification via Distribution Estimation (2303.16764v1)

Published 26 Mar 2023 in cs.CL

Abstract: Distribution estimation has been demonstrated as one of the most effective approaches in dealing with few-shot image classification, as the low-level patterns and underlying representations can be easily transferred across different tasks in computer vision domain. However, directly applying this approach to few-shot text classification is challenging, since leveraging the statistics of known classes with sufficient samples to calibrate the distributions of novel classes may cause negative effects due to serious category difference in text domain. To alleviate this issue, we propose two simple yet effective strategies to estimate the distributions of the novel classes by utilizing unlabeled query samples, thus avoiding the potential negative transfer issue. Specifically, we first assume a class or sample follows the Gaussian distribution, and use the original support set and the nearest few query samples to estimate the corresponding mean and covariance. Then, we augment the labeled samples by sampling from the estimated distribution, which can provide sufficient supervision for training the classification model. Extensive experiments on eight few-shot text classification datasets show that the proposed method outperforms state-of-the-art baselines significantly.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Han Liu (340 papers)
  2. Feng Zhang (180 papers)
  3. Xiaotong Zhang (28 papers)
  4. Siyang Zhao (5 papers)
  5. Fenglong Ma (66 papers)
  6. Xiao-Ming Wu (91 papers)
  7. Hongyang Chen (61 papers)
  8. Hong Yu (114 papers)
  9. Xianchao Zhang (15 papers)
Citations (8)

Summary

We haven't generated a summary for this paper yet.