Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Utilizing Bidirectional Encoder Representations from Transformers for Answer Selection (2011.07208v1)

Published 14 Nov 2020 in cs.CL and cs.IR

Abstract: Pre-training a transformer-based model for the LLMing task in a large dataset and then fine-tuning it for downstream tasks has been found very useful in recent years. One major advantage of such pre-trained LLMs is that they can effectively absorb the context of each word in a sentence. However, for tasks such as the answer selection task, the pre-trained LLMs have not been extensively used yet. To investigate their effectiveness in such tasks, in this paper, we adopt the pre-trained Bidirectional Encoder Representations from Transformer (BERT) LLM and fine-tune it on two Question Answering (QA) datasets and three Community Question Answering (CQA) datasets for the answer selection task. We find that fine-tuning the BERT model for the answer selection task is very effective and observe a maximum improvement of 13.1% in the QA datasets and 18.7% in the CQA datasets compared to the previous state-of-the-art.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Md Tahmid Rahman Laskar (30 papers)
  2. Enamul Hoque (26 papers)
  3. Jimmy Xiangji Huang (18 papers)
Citations (13)