Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Does QA-based intermediate training help fine-tuning language models for text classification? (2112.15051v1)

Published 30 Dec 2021 in cs.CL and cs.AI

Abstract: Fine-tuning pre-trained LLMs for downstream tasks has become a norm for NLP. Recently it is found that intermediate training based on high-level inference tasks such as Question Answering (QA) can improve the performance of some LLMs for target tasks. However it is not clear if intermediate training generally benefits various LLMs. In this paper, using the SQuAD-2.0 QA task for intermediate training for target text classification tasks, we experimented on eight tasks for single-sequence classification and eight tasks for sequence-pair classification using two base and two compact LLMs. Our experiments show that QA-based intermediate training generates varying transfer performance across different LLMs, except for similar QA tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Shiwei Zhang (179 papers)
  2. Xiuzhen Zhang (35 papers)
Citations (2)