Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 148 tok/s
Gemini 2.5 Pro 47 tok/s Pro
GPT-5 Medium 26 tok/s Pro
GPT-5 High 28 tok/s Pro
GPT-4o 94 tok/s Pro
Kimi K2 214 tok/s Pro
GPT OSS 120B 429 tok/s Pro
Claude Sonnet 4.5 36 tok/s Pro
2000 character limit reached

PatternBoost: Constructions in Mathematics with a Little Help from AI (2411.00566v1)

Published 1 Nov 2024 in math.CO and cs.LG

Abstract: We introduce PatternBoost, a flexible method for finding interesting constructions in mathematics. Our algorithm alternates between two phases. In the first local'' phase, a classical search algorithm is used to produce many desirable constructions. In the secondglobal'' phase, a transformer neural network is trained on the best such constructions. Samples from the trained transformer are then used as seeds for the first phase, and the process is repeated. We give a detailed introduction to this technique, and discuss the results of its application to several problems in extremal combinatorics. The performance of PatternBoost varies across different problems, but there are many situations where its performance is quite impressive. Using our technique, we find the best known solutions to several long-standing problems, including the construction of a counterexample to a conjecture that had remained open for 30 years.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (50)
  1. Sum of squares of sum of squares function r2⁢(n)subscript𝑟2𝑛r_{2}(n)italic_r start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT ( italic_n ). https://math.stackexchange.com/questions/322086/sum-of-squares-of-sum-of-squares-function-r-2n?rq=1. Accessed: 2024-09-27.
  2. Triangles in the integer grid [n]×[n]delimited-[]𝑛delimited-[]𝑛[n]\times[n][ italic_n ] × [ italic_n ]. https://mathweb.ucsd.edu/~asuk/solymosi.pdf. Accessed: 2024-09-27.
  3. Global Lyapunov functions: a long-standing open problem in mathematics, with symbolic transformers, 2024.
  4. On partitions of discrete boxes. Discrete mathematics, 257(2-3):255–258, 2002.
  5. Brandon Amos. Tutorial on amortized optimization, 2023.
  6. OptNet: Differentiable optimization as a layer in neural networks, 2021.
  7. Improved bounds for cross-Sperner systems. arXiv preprint arXiv:2302.02516, 2023.
  8. A note on the random greedy independent set algorithm. Random Structures & Algorithms, 49(3):479–502, 2016.
  9. Estimates for representation numbers of quadratic forms. 2006.
  10. An improvement to the Kelley-Meka bounds on three-term arithmetic progressions. arXiv preprint arXiv:2309.02353, 2023.
  11. Edge deletion preserving the diameter of the hypercube. Discrete Applied Mathematics, 63(1):91–95, 1995.
  12. Roger W Brockett. Dynamical systems that sort lists, diagonalize matrices, and solve linear programming problems. Linear Algebra and its applications, 146:79–91, 1991.
  13. Language models are few-shot learners. arXiv preprint arXiv:2005.14165, 2020.
  14. Pattern-avoiding (0, 1)-matrices. arXiv preprint arXiv:2005.00379, 2020.
  15. François Charton. Linear algebra with transformers, 2022.
  16. Neural optimization machine: A neural network approach for optimization, 2022.
  17. Neural ordinary differential equations, 2019.
  18. George Cybenko. Approximation by superpositions of a sigmoidal function. Mathematics of control, signals and systems, 2(4):303–314, 1989.
  19. Advancing mathematics by guiding human intuition with AI. Nature, 2021.
  20. A neural network based approach to mechanical design optimization. Engineering Optimization, 20(3):187–203, 1992.
  21. Convergence rates for ordinal embedding. arXiv preprint arXiv:1904.12994, 2019.
  22. Hypercube subgraphs with minimal detours. Journal of Graph Theory, 23(2):119–128, 1996.
  23. The history of degenerate (bipartite) extremal graph problems. In Erdős centennial, pages 169–264. Springer, 2013.
  24. Saturating Sperner families. Graphs and Combinatorics, 29(5):1355–1364, 2013.
  25. Timothy Gowers. Turán’s theorem for triangles. https://wtgowers.github.io/human-style-atp/2023/01/11/turan.html. Accessed: 2024-09-29.
  26. Changing and unchanging the diameter of a hypercube. Discrete Applied Mathematics, 37:265–274, 1992.
  27. Gurobi Optimization, LLC. Gurobi Optimizer Reference Manual, 2024.
  28. Gaussian Error Linear Units (GELUs), 2023.
  29. Computing with neural circuits: A model. Science, 233(4764):625–633, 1986.
  30. Kurt Hornik. Approximation capabilities of multilayer feedforward networks. Neural networks, 4(2):251–257, 1991.
  31. Andrej Karpathy. Makemore. https://github.com/karpathy/makemore. Accessed: 2024-09-29.
  32. Finite algebras of finite complexity. Discrete mathematics, 207(1-3):89–135, 1999.
  33. Deep learning for symbolic mathematics. arXiv preprint arXiv:1912.01412, 2019.
  34. Decomposing the complete r-graph. Journal of Combinatorial Theory, Series A, 154:21–31, 2018.
  35. Teaching arithmetic to small transformers, 2023.
  36. Saturation of k𝑘kitalic_k-chains in the Boolean lattice. arXiv preprint arXiv:2402.14113, 2024.
  37. Finding increasingly large extremal graphs with AlphaZero and Tabu search. arXiv preprint arXiv:2311.03583, 2023.
  38. On saturated k𝑘kitalic_k-Sperner systems. arXiv preprint arXiv:1402.5646, 2014.
  39. Bounded degree spanners of the hypercube. arXiv preprint arXiv:1910.09868, 2019.
  40. Investigating the limitations of transformers with simple arithmetic tasks. arXiv preprint arXiv:2102.13019, 2021.
  41. OEIS Foundation Inc. The On-Line Encyclopedia of Integer Sequences, 2024. Published electronically at http://oeis.org.
  42. Language models are unsupervised multitask learners. 2019.
  43. Mathematical discoveries from program search with large language models. Nature, 625(7995):468–475, 2024.
  44. Michael Saks. Kleitman and combinatorics. Discrete mathematics, 257(2-3):225–247, 2002.
  45. Torsten Thiele. Geometric selection problems and hypergraphs. PhD thesis, Citeseer, 1995.
  46. LLaMA: Open and Efficient Foundation Language Models. arXiv preprint arXiv:2302.13971, 2023.
  47. Solving olympiad geometry without human demonstrations. Nature, 625(7995):476–482, 2024.
  48. Attention is all you need. In Advances in Neural Information Processing Systems, pages 6000–6010, 2017.
  49. Adam Zsolt Wagner. Constructions in combinatorics via neural networks. arXiv preprint arXiv:2104.14516, 2021.
  50. Chai Wah Wu. Counting the number of isosceles triangles in rectangular regular grids. arXiv preprint arXiv:1605.00180, 2016.
Citations (2)

Summary

  • The paper demonstrates a novel iterative method that hybridizes greedy local search with transformer-based global pattern recognition to solve complex mathematical constructions.
  • The methodology leverages a feedback loop between local optimization and global pattern learning to produce near-optimal solutions in extremal combinatorics.
  • The research paves the way for versatile AI-assisted approaches that simplify addressing longstanding mathematical challenges across various domains.

Overview of "PatternBoost: Constructions in Mathematics with a Little Help from AI"

This paper presents "PatternBoost," an innovative approach that leverages machine learning, specifically transformer architectures, in conjunction with classical search algorithms to solve complex mathematical construction problems. The authors propose an iterative method that combines local optimization with global pattern recognition, utilizing transformers to generate and refine mathematical constructions. This method is particularly applied to extremal combinatorics, providing a new toolset for mathematicians seeking solutions to longstanding problems.

The PatternBoost Methodology

The main concept behind PatternBoost is a two-phase iterative process:

  1. Local Phase: A classical search algorithm, typically a greedy method, is employed to generate a set of initial constructions. These constructions are evaluated and filtered based on a bespoke scoring function appropriate to the mathematical problem at hand.
  2. Global Phase: A transformer neural network is trained on the top constructions from the local search. The trained model generates new candidate constructions by recognizing global patterns from the input dataset. The results of these suggestions are used as seeds for the subsequent local phase, creating a feedback loop that iteratively enhances the quality of solutions.

The strength of PatternBoost lies in its ability to exploit the rich, abstract representations offered by transformers, aiming to capture subtle patterns and structures that are challenging to identify through human intuition or isolated local search strategies.

Applications and Results

The paper details several applications of PatternBoost in extremal combinatorics. It successfully finds optimal or near-optimal solutions to longstanding open problems, including the construction of counterexamples to conjectures that remained unresolved for decades. Notably, the paper highlights its role in generating mathematical structures that produce competitive results compared to traditional handcrafted solutions. While the method does not always surpass human experts in all domains, it provides a robust, general-purpose tool that significantly narrows the gap.

Technical Implications

PatternBoost's hybrid approach can be seen as an evolution of genetic algorithms, where the neural network acts to enrich the generation pool based on an understanding of the problem's global constraints. This method's adaptability to various problems without the need for intricate machine learning expertise democratizes its utilization among mathematicians.

Notably, the paper underscores the model's resilience to operate "off the shelf," simplifying its deployment in mathematical research without necessitating extensive parameter tuning or computational resources beyond what's typically available in academic environments.

Future Directions

The research opens several avenues for future exploration:

  • Scalability to Other Domains: While demonstrated on combinatorial problems, PatternBoost could potentially apply to other mathematical or even non-mathematical domains where construction or optimization problems arise.
  • Enhanced Local Search Strategies: Further refinement of local search algorithms in tandem with transformer outputs could enhance results.
  • Tokenization and Data Representation: Exploring more sophisticated tokenization methods could improve model learning efficiencies, particularly for complex or higher-dimensional problems.
  • Interdisciplinary Applications: The applicability of PatternBoost in scientific domains where both pattern recognition and optimization are critical could lead to significant interdisciplinary breakthroughs.

In conclusion, this work represents a significant advancement in applying AI techniques to mathematical construction problems, offering new possibilities and insights into problem-solving methodologies that bridge human intuition and machine learning capabilities.

Dice Question Streamline Icon: https://streamlinehq.com
Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 7 tweets and received 166 likes.

Upgrade to Pro to view all of the tweets about this paper:

Youtube Logo Streamline Icon: https://streamlinehq.com