FOLIO: Natural Language Reasoning with First-Order Logic (2209.00840v3)
Abstract: LLMs have achieved remarkable performance on a variety of natural language understanding tasks. However, existing benchmarks are inadequate in measuring the complex logical reasoning capabilities of a model. We present FOLIO, a human-annotated, logically complex and diverse dataset for reasoning in natural language (NL), equipped with first-order logic (FOL) annotations. FOLIO consists of 1,430 examples (unique conclusions), each paired with one of 487 sets of premises used to deductively reason for the validity of each conclusion. The logical correctness of the premises and conclusions is ensured by their FOL annotations, which are automatically verified by an FOL inference engine. In addition to the main NL reasoning task, NL-FOL pairs in FOLIO constitute a new NL-FOL translation dataset. Our experiments on FOLIO systematically evaluate the FOL reasoning ability of supervised fine-tuning on medium-sized LLMs. For both NL reasoning and NL-FOL translation, we benchmark multiple state-of-the-art LLMs. Our results show that a subset of FOLIO presents a challenge for one of the most capable {LLM} publicly available, GPT-4.
- Simeng Han (20 papers)
- Hailey Schoelkopf (22 papers)
- Yilun Zhao (59 papers)
- Zhenting Qi (19 papers)
- Martin Riddell (4 papers)
- Luke Benson (2 papers)
- Lucy Sun (1 paper)
- Ekaterina Zubova (1 paper)
- Yujie Qiao (4 papers)
- Matthew Burtell (2 papers)
- David Peng (3 papers)
- Jonathan Fan (3 papers)
- Yixin Liu (108 papers)
- Brian Wong (5 papers)
- Malcolm Sailor (1 paper)
- Ansong Ni (17 papers)
- Linyong Nan (17 papers)
- Jungo Kasai (38 papers)
- Tao Yu (282 papers)
- Rui Zhang (1138 papers)