Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

NaturalProver: Grounded Mathematical Proof Generation with Language Models (2205.12910v2)

Published 25 May 2022 in cs.CL and cs.AI

Abstract: Theorem proving in natural mathematical language - the mixture of symbolic and natural language used by humans - plays a central role in mathematical advances and education, and tests aspects of reasoning that are core to intelligence. Yet it has remained underexplored with modern generative models. We study large-scale LLMs on two new generation tasks: suggesting the next step in a mathematical proof, and full proof generation. We develop NaturalProver, a LLM that generates proofs by conditioning on background references (e.g. theorems and definitions that are either retrieved or human-provided), and optionally enforces their presence with constrained decoding. On theorems from the NaturalProofs benchmark, NaturalProver improves the quality of next-step suggestions and generated proofs over fine-tuned GPT-3, according to human evaluations from university-level mathematics students. NaturalProver is capable of proving some theorems that require short (2-6 step) proofs, and providing next-step suggestions that are rated as correct and useful over 40% of the time, which is to our knowledge the first demonstration of these capabilities using neural LLMs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Sean Welleck (54 papers)
  2. Jiacheng Liu (67 papers)
  3. Ximing Lu (52 papers)
  4. Hannaneh Hajishirzi (176 papers)
  5. Yejin Choi (287 papers)
Citations (54)

Summary

We haven't generated a summary for this paper yet.

Github Logo Streamline Icon: https://streamlinehq.com