Machine learning and information theory concepts towards an AI Mathematician (2403.04571v1)
Abstract: The current state-of-the-art in artificial intelligence is impressive, especially in terms of mastery of language, but not so much in terms of mathematical reasoning. What could be missing? Can we learn something useful about that gap from how the brains of mathematicians go about their craft? This essay builds on the idea that current deep learning mostly succeeds at system 1 abilities -- which correspond to our intuition and habitual behaviors -- but still lacks something important regarding system 2 abilities -- which include reasoning and robust uncertainty estimation. It takes an information-theoretical posture to ask questions about what constitutes an interesting mathematical statement, which could guide future work in crafting an AI mathematician. The focus is not on proving a given theorem but on discovering new and interesting conjectures. The central hypothesis is that a desirable body of theorems better summarizes the set of all provable statements, for example by having a small description length while at the same time being close (in terms of number of derivation steps) to many provable statements.
- A survey of exploration methods in reinforcement learning. arXiv preprint arXiv:2109.00157, 2021.
- Neural machine translation by jointly learning to align and translate. International Conference on Learning Representations, 2014.
- Yoshua Bengio. The consciousness prior. arXiv preprint arXiv:1709.08568, 2017.
- A neural probabilistic language model. Neural Information Processing Systems, 2001.
- Curriculum learning. International Conference on Machine Learning, 2009.
- Deep learning for ai. Communications of the ACM, 64(7):58–65, 2021.
- GFlowNet foundations. Journal of Machine Learning Research, (24):1–76, 2023.
- Interactive theorem proving and program development: Coq’Art: the calculus of inductive constructions. Springer, 2013.
- Léon Bottou. Large-scale machine learning with stochastic gradient descent. International Conference on Computational Statistics, 2010.
- Improving generalization with active learning. Machine learning, 15:201–221, 1994.
- A deep reinforcement learning approach to first-order logic theorem proving. AAAI Conference on Artificial Intelligence, 2021.
- Donald Davidson. Truth and meaning. Synthese, 17(1):304–323, 1967.
- Symbols and mental programs: a hypothesis about human singularity. Trends in Cognitive Sciences, 2022.
- DreamCoder: Bootstrapping inductive program synthesis with wake-sleep library learning. Programming Language Design and Implementation, 2021.
- Herbert Gelernter. Realization of a geometry theorem-proving machine. International Conference on Information Processing, 1959.
- Inductive biases for deep learning of higher-level cognition. Proceedings of the Royal Society A, 478(2266):20210068, 2022.
- Proof artifact co-training for theorem proving with language models. International Conference on Learning Representations, 2022.
- Steve Hanneke. Rates of convergence in active learning. The Annals of Statistics, 39(1):333–361, 2011.
- Using fast weights to deblur old memories. In Proceedings of the ninth annual conference of the Cognitive Science Society, pages 177–186, 1987.
- William A. Howard. The formulae-as-types notion of construction. 1969.
- Scientific reasoning: the Bayesian approach. Open Court Publishing, 2006.
- Aleatoric and epistemic uncertainty in machine learning: An introduction to concepts and methods. Machine Learning, 110:457–506, 2021.
- Sources of richness and ineffability for phenomenally conscious states. arXiv preprint arXiv:2302.06403, 2023.
- Daniel Kahneman. Thinking, Fast and Slow. Macmillan, 2011.
- Landmark-guided subgoal generation in hierarchical reinforcement learning. Neural Information Processing Systems, 2021.
- Deep learning. Nature, 521(7553):436–444, 2015.
- David JC MacKay. Information theory, inference and learning algorithms. Cambridge University Press, 2003.
- The mathlib community. The Lean mathematical library. International Conference on Certified Programs and Proofs, 2020.
- George A Miller. The magical number seven, plus or minus two: Some limits on our capacity for processing information. Psychological Review, 63(2):81, 1956.
- The use of machines to assist in rigorous proof. Philosophical Transactions of the Royal Society of London. Series A, Mathematical and Physical Sciences, 312(1522):411–422, 1984.
- The Lean 4 theorem prover and programming language. 2021.
- OpenAI. ChatGPT release notes, 2022.
- OpenAI. GPT-4 technical report, 2023.
- Formal mathematics statement curriculum learning. International Conference on Learning Representations, 2023.
- Beata Randrianantoanina and Narcisse Randrianantoanina, editors. Banach Spaces and their Applications in Analysis, 2007. De Gruyter.
- Jorma Rissanen. Coding and complexity. In Sequences: Combinatorics, Compression, Security, and Transmission, pages 312–325. Springer, 1990.
- Stuart Russell. Human compatible: Artificial intelligence and the problem of control. Penguin, 2019.
- Jürgen Schmidhuber. Learning complex, extended sequences using the principle of history compression. Neural Computation, 4(2):234–242, 1992.
- The IMO Grand Challenge. URL https://imo-grand-challenge.github.io/.
- Burr Settles. Active learning literature survey. Computer Sciences Technical Report 1648, University of Wisconsin–Madison, 2009.
- Mastering the game of go without human knowledge. Nature, 550(7676):354–359, 2017.
- Reinforcement learning: An introduction. MIT press, 2018.
- Alfred Tarski. The semantic conception of truth and the foundations of semantics. Philosophy and Phenomenological Research, 4(3):341–376, 1943.
- Vladimir N Vapnik. Estimation of Dependencies Based on Empirical Data. Springer, 1982.
- Attention is all you need. Neural Information Processing Systems, 2017.
- Automated proof compression by invention of new definitions. In Edmund M. Clarke and Andrei Voronkov, editors, Logic for Programming, Artificial Intelligence, and Reasoning, pages 447–462. Springer, 2010.
- Philip Wadler. Propositions as types. Communications of the ACM, 58(12):75–84, 2015.
- LeanDojo: Theorem proving with retrieval-augmented language models. Neural Information Processing Systems, 2023.
- MiniF2F: a cross-system benchmark for formal Olympiad-level mathematics. International Conference on Learning Representations, 2021.