Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Mathematical Reasoning via Self-supervised Skip-tree Training (2006.04757v3)

Published 8 Jun 2020 in cs.LG, cs.AI, cs.PL, and stat.ML

Abstract: We examine whether self-supervised LLMing applied to mathematical formulas enables logical reasoning. We suggest several logical reasoning tasks that can be used to evaluate LLMs trained on formal mathematical statements, such as type inference, suggesting missing assumptions and completing equalities. To train LLMs for formal mathematics, we propose a novel skip-tree task. We find that models trained on the skip-tree task show surprisingly strong mathematical reasoning abilities, and outperform models trained on standard skip-sequence tasks. We also analyze the models' ability to formulate new conjectures by measuring how often the predictions are provable and useful in other proofs.

Citations (2)

Summary

We haven't generated a summary for this paper yet.