Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

RuleBert: Teaching Soft Rules to Pre-trained Language Models (2109.13006v1)

Published 24 Sep 2021 in cs.AI, cs.CL, cs.LG, cs.LO, and cs.NE

Abstract: While pre-trained LLMs (PLMs) are the go-to solution to tackle many natural language processing problems, they are still very limited in their ability to capture and to use common-sense knowledge. In fact, even if information is available in the form of approximate (soft) logical rules, it is not clear how to transfer it to a PLM in order to improve its performance for deductive reasoning tasks. Here, we aim to bridge this gap by teaching PLMs how to reason with soft Horn rules. We introduce a classification task where, given facts and soft rules, the PLM should return a prediction with a probability for a given hypothesis. We release the first dataset for this task, and we propose a revised loss function that enables the PLM to learn how to predict precise probabilities for the task. Our evaluation results show that the resulting fine-tuned models achieve very high performance, even on logical rules that were unseen at training. Moreover, we demonstrate that logical notions expressed by the rules are transferred to the fine-tuned model, yielding state-of-the-art results on external datasets.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Mohammed Saeed (5 papers)
  2. Naser Ahmadi (3 papers)
  3. Preslav Nakov (253 papers)
  4. Paolo Papotti (22 papers)
Citations (28)

Summary

We haven't generated a summary for this paper yet.