Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
126 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

The Effectiveness of Masked Language Modeling and Adapters for Factual Knowledge Injection (2210.00907v1)

Published 3 Oct 2022 in cs.CL

Abstract: This paper studies the problem of injecting factual knowledge into large pre-trained LLMs. We train adapter modules on parts of the ConceptNet knowledge graph using the masked LLMing objective and evaluate the success of the method by a series of probing experiments on the LAMA probe. Mean P@K curves for different configurations indicate that the technique is effective, increasing the performance on subsets of the LAMA probe for large values of k by adding as little as 2.1% additional parameters to the original models.

Citations (3)

Summary

We haven't generated a summary for this paper yet.