Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

BRENT: Bidirectional Retrieval Enhanced Norwegian Transformer (2304.09649v1)

Published 19 Apr 2023 in cs.CL

Abstract: Retrieval-based LLMs are increasingly employed in question-answering tasks. These models search in a corpus of documents for relevant information instead of having all factual knowledge stored in its parameters, thereby enhancing efficiency, transparency, and adaptability. We develop the first Norwegian retrieval-based model by adapting the REALM framework and evaluating it on various tasks. After training, we also separate the LLM, which we call the reader, from the retriever components, and show that this can be fine-tuned on a range of downstream tasks. Results show that retrieval augmented LLMing improves the reader's performance on extractive question-answering, suggesting that this type of training improves LLMs' general ability to use context and that this does not happen at the expense of other abilities such as part-of-speech tagging, dependency parsing, named entity recognition, and lemmatization. Code, trained models, and data are made publicly available.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Lucas Georges Gabriel Charpentier (8 papers)
  2. Sondre Wold (9 papers)
  3. David Samuel (23 papers)
  4. Egil Rønningstad (7 papers)
Citations (2)