Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Fine-tuning BERT for Low-Resource Natural Language Understanding via Active Learning (2012.02462v1)

Published 4 Dec 2020 in cs.CL

Abstract: Recently, leveraging pre-trained Transformer based LLMs in down stream, task specific models has advanced state of the art results in natural language understanding tasks. However, only a little research has explored the suitability of this approach in low resource settings with less than 1,000 training data points. In this work, we explore fine-tuning methods of BERT -- a pre-trained Transformer based LLM -- by utilizing pool-based active learning to speed up training while keeping the cost of labeling new data constant. Our experimental results on the GLUE data set show an advantage in model performance by maximizing the approximate knowledge gain of the model when querying from the pool of unlabeled data. Finally, we demonstrate and analyze the benefits of freezing layers of the LLM during fine-tuning to reduce the number of trainable parameters, making it more suitable for low-resource settings.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Daniel Grießhaber (4 papers)
  2. Johannes Maucher (12 papers)
  3. Ngoc Thang Vu (93 papers)
Citations (43)