Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Knowledge-Augmented Language Model and its Application to Unsupervised Named-Entity Recognition (1904.04458v2)

Published 9 Apr 2019 in cs.CL, cs.AI, and cs.LG

Abstract: Traditional LLMs are unable to efficiently model entity names observed in text. All but the most popular named entities appear infrequently in text providing insufficient context. Recent efforts have recognized that context can be generalized between entity names that share the same type (e.g., \emph{person} or \emph{location}) and have equipped LLMs with access to an external knowledge base (KB). Our Knowledge-Augmented LLM (KALM) continues this line of work by augmenting a traditional model with a KB. Unlike previous methods, however, we train with an end-to-end predictive objective optimizing the perplexity of text. We do not require any additional information such as named entity tags. In addition to improving LLMing performance, KALM learns to recognize named entities in an entirely unsupervised way by using entity type information latent in the model. On a Named Entity Recognition (NER) task, KALM achieves performance comparable with state-of-the-art supervised models. Our work demonstrates that named entities (and possibly other types of world knowledge) can be modeled successfully using predictive learning and training on large corpora of text without any additional information.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Angli Liu (4 papers)
  2. Jingfei Du (16 papers)
  3. Veselin Stoyanov (21 papers)
Citations (37)