Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Exploring Neural Net Augmentation to BERT for Question Answering on SQUAD 2.0 (1908.01767v3)

Published 4 Aug 2019 in cs.CL, cs.AI, and cs.LG

Abstract: Enhancing machine capabilities to answer questions has been a topic of considerable focus in recent years of NLP research. LLMs like Embeddings from LLMs (ELMo)[1] and Bidirectional Encoder Representations from Transformers (BERT) [2] have been very successful in developing general purpose LLMs that can be optimized for a large number of downstream language tasks. In this work, we focused on augmenting the pre-trained BERT LLM with different output neural net architectures and compared their performance on question answering task posed by the Stanford Question Answering Dataset 2.0 (SQUAD 2.0) [3]. Additionally, we also fine-tuned the pre-trained BERT model parameters to demonstrate its effectiveness in adapting to specialized language tasks. Our best output network, is the contextualized CNN that performs on both the unanswerable and answerable question answering tasks with F1 scores of 75.32 and 64.85 respectively.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (1)
  1. Suhas Gupta (1 paper)
Citations (1)