Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Robust Fine-tuning via Perturbation and Interpolation from In-batch Instances (2205.00633v1)

Published 2 May 2022 in cs.CL

Abstract: Fine-tuning pretrained LLMs (PLMs) on downstream tasks has become common practice in natural language processing. However, most of the PLMs are vulnerable, e.g., they are brittle under adversarial attacks or imbalanced data, which hinders the application of the PLMs on some downstream tasks, especially in safe-critical scenarios. In this paper, we propose a simple yet effective fine-tuning method called Match-Tuning to force the PLMs to be more robust. For each instance in a batch, we involve other instances in the same batch to interact with it. To be specific, regarding the instances with other labels as a perturbation, Match-Tuning makes the model more robust to noise at the beginning of training. While nearing the end, Match-Tuning focuses more on performing an interpolation among the instances with the same label for better generalization. Extensive experiments on various tasks in GLUE benchmark show that Match-Tuning consistently outperforms the vanilla fine-tuning by $1.64$ scores. Moreover, Match-Tuning exhibits remarkable robustness to adversarial attacks and data imbalance.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Shoujie Tong (3 papers)
  2. Qingxiu Dong (39 papers)
  3. Damai Dai (38 papers)
  4. Tianyu Liu (177 papers)
  5. Baobao Chang (80 papers)
  6. Zhifang Sui (89 papers)
  7. Yifan Song (49 papers)
Citations (5)