Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

GiBERT: Introducing Linguistic Knowledge into BERT through a Lightweight Gated Injection Method (2010.12532v1)

Published 23 Oct 2020 in cs.CL

Abstract: Large pre-trained LLMs such as BERT have been the driving force behind recent improvements across many NLP tasks. However, BERT is only trained to predict missing words - either behind masks or in the next sentence - and has no knowledge of lexical, syntactic or semantic information beyond what it picks up through unsupervised pre-training. We propose a novel method to explicitly inject linguistic knowledge in the form of word embeddings into any layer of a pre-trained BERT. Our performance improvements on multiple semantic similarity datasets when injecting dependency-based and counter-fitted embeddings indicate that such information is beneficial and currently missing from the original model. Our qualitative analysis shows that counter-fitted embedding injection particularly helps with cases involving synonym pairs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Nicole Peinelt (4 papers)
  2. Marek Rei (52 papers)
  3. Maria Liakata (59 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.