Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Omni-MATH: A Universal Olympiad Level Mathematic Benchmark For Large Language Models (2410.07985v2)

Published 10 Oct 2024 in cs.CL
Omni-MATH: A Universal Olympiad Level Mathematic Benchmark For Large Language Models

Abstract: Recent advancements in LLMs have led to significant breakthroughs in mathematical reasoning capabilities. However, existing benchmarks like GSM8K or MATH are now being solved with high accuracy (e.g., OpenAI o1 achieves 94.8% on MATH dataset), indicating their inadequacy for truly challenging these models. To bridge this gap, we propose a comprehensive and challenging benchmark specifically designed to assess LLMs' mathematical reasoning at the Olympiad level. Unlike existing Olympiad-related benchmarks, our dataset focuses exclusively on mathematics and comprises a vast collection of 4428 competition-level problems with rigorous human annotation. These problems are meticulously categorized into over 33 sub-domains and span more than 10 distinct difficulty levels, enabling a holistic assessment of model performance in Olympiad-mathematical reasoning. Furthermore, we conducted an in-depth analysis based on this benchmark. Our experimental results show that even the most advanced models, OpenAI o1-mini and OpenAI o1-preview, struggle with highly challenging Olympiad-level problems, with 60.54% and 52.55% accuracy, highlighting significant challenges in Olympiad-level mathematical reasoning.

Omni-MATH: A Universal Olympiad Level Mathematical Benchmark for LLMs

The paper "Omni-MATH: A Universal Olympiad Level Mathematic Benchmark for LLMs" introduces a sophisticated benchmark designed to rigorously evaluate the mathematical reasoning capabilities of LLMs at an Olympiad-level. This benchmark is positioned as a necessary progression from existing datasets, which are increasingly inadequate for assessing the upper echelons of LLM mathematical problem-solving abilities.

Key Contributions

  1. Benchmark Development: Omni-MATH presents an exhaustive set of 4,428 competition-level problems, all of which are meticulously annotated by humans. These problems are categorized into 33 sub-domains and over 10 distinct difficulty levels. This stratification allows for a nuanced evaluation of LLMs, providing insight into their strengths and weaknesses across various mathematical domains.
  2. Experimental Validation: Two advanced OpenAI models, o1-mini and o1-preview, were tested against the Omni-MATH benchmark, revealing significant performance gaps—accuracy rates of 60.54% and 52.55% respectively. These results indicate that even cutting-edge models face considerable challenges when navigating Olympiad-level problems.
  3. Omni-Judge: Proposed as an open-source verification tool, Omni-Judge achieved over 91% consistency with GPT-4o evaluations and 86% with human judgments. This verifier offers a robust mechanism to evaluate Olympiad-level problems efficiently, validating the reliability of model assessments against human standards.

Implications and Future Directions

The development of Omni-MATH is a pivotal advancement, establishing a new standard for LLM evaluation in the domain of complex mathematical reasoning. The distinct categorization of problem domains and difficulties facilitates a granular analysis of LLM capabilities, enriching our understanding of model performance. The identified deficiencies, particularly in discrete mathematics, underscore ongoing challenges that necessitate further research and development.

Key future directions include:

  • Enhanced Test-Time Scaling: As identified, the Best-of-N approach has limited efficacy at the Olympiad level, suggesting the need for more innovative scaling techniques that can better leverage model capabilities during inference.
  • Focused Domain Training: The difficulties faced by LLMs, especially in discrete mathematics, suggest potential benefits from targeted domain-specific training regimes, which may enhance generalization across unseen problem types.
  • Robust Data Leakage Analysis: The paper identifies nominal data leakage in the model datasets, which has minimal impact on overall performance metrics. Continued vigilance in this area will ensure the integrity and applicability of the benchmark results.

In conclusion, Omni-MATH provides a comprehensive framework to critically evaluate LLMs' mathematical reasoning skills at a high level of complexity. The benchmarks set forth by this work are poised to guide the next wave of research and development in LLM mathematical capabilities, ensuring these models can meet real-world technological and educational demands. The ongoing iterations and enhancements to models on benchmarks like Omni-MATH will drive forward our understanding and capabilities in artificial intelligence for mathematical reasoning.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (20)
  1. Bofei Gao (15 papers)
  2. Feifan Song (14 papers)
  3. Zhe Yang (60 papers)
  4. Zefan Cai (26 papers)
  5. Yibo Miao (24 papers)
  6. Qingxiu Dong (39 papers)
  7. Lei Li (1293 papers)
  8. Chenghao Ma (3 papers)
  9. Liang Chen (360 papers)
  10. Runxin Xu (30 papers)
  11. Zhengyang Tang (13 papers)
  12. Benyou Wang (109 papers)
  13. Daoguang Zan (24 papers)
  14. Shanghaoran Quan (12 papers)
  15. Ge Zhang (170 papers)
  16. Lei Sha (34 papers)
  17. Yichang Zhang (24 papers)
  18. Xuancheng Ren (59 papers)
  19. Tianyu Liu (177 papers)
  20. Baobao Chang (80 papers)
Citations (3)