Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Self-Training with Differentiable Teacher (2109.07049v2)

Published 15 Sep 2021 in cs.CL and cs.LG

Abstract: Self-training achieves enormous success in various semi-supervised and weakly-supervised learning tasks. The method can be interpreted as a teacher-student framework, where the teacher generates pseudo-labels, and the student makes predictions. The two models are updated alternatingly. However, such a straightforward alternating update rule leads to training instability. This is because a small change in the teacher may result in a significant change in the student. To address this issue, we propose DRIFT, short for differentiable self-training, that treats teacher-student as a Stackelberg game. In this game, a leader is always in a more advantageous position than a follower. In self-training, the student contributes to the prediction performance, and the teacher controls the training process by generating pseudo-labels. Therefore, we treat the student as the leader and the teacher as the follower. The leader procures its advantage by acknowledging the follower's strategy, which involves differentiable pseudo-labels and differentiable sample weights. Consequently, the leader-follower interaction can be effectively captured via Stackelberg gradient, obtained by differentiating the follower's strategy. Experimental results on semi- and weakly-supervised classification and named entity recognition tasks show that our model outperforms existing approaches by large margins.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Simiao Zuo (25 papers)
  2. Yue Yu (343 papers)
  3. Chen Liang (140 papers)
  4. Haoming Jiang (52 papers)
  5. Siawpeng Er (6 papers)
  6. Chao Zhang (907 papers)
  7. Tuo Zhao (131 papers)
  8. Hongyuan Zha (136 papers)
Citations (14)

Summary

We haven't generated a summary for this paper yet.