Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Mathematical Language Models: A Survey (2312.07622v3)

Published 12 Dec 2023 in cs.CL
Mathematical Language Models: A Survey

Abstract: In recent years, there has been remarkable progress in leveraging LLMs (LMs), encompassing Pre-trained LLMs (PLMs) and Large-scale LLMs, within the domain of mathematics. This paper conducts a comprehensive survey of mathematical LMs, systematically categorizing pivotal research endeavors from two distinct perspectives: tasks and methodologies. The landscape reveals a large number of proposed mathematical LLMs, which are further delineated into instruction learning, tool-based methods, fundamental CoT techniques, and advanced CoT methodologies. In addition, our survey entails the compilation of over 60 mathematical datasets, including training datasets, benchmark datasets, and augmented datasets. Addressing the primary challenges and delineating future trajectories within the field of mathematical LMs, this survey is positioned as a valuable resource, poised to facilitate and inspire future innovation among researchers invested in advancing this domain.

Introduction

Recent advancements in LLMs have led to remarkable progress in their application within the field of mathematics. This survey concentrates on mathematical LLMs (LMs), which include Pre-trained LLMs (PLMs) and Large-scale LLMs. These models play a pivotal role in addressing various mathematical tasks such as performing calculations and reasoning, and this ability is transforming mathematical exploration and practical usage.

Mathematical Tasks

Mathematical tasks tackled by LMs fall into two main categories: mathematical calculation and mathematical reasoning. Mathematical calculation primarily involves arithmetic operations and the representation of numerical data. Initially, LMs portrayed basic computational skills using textual number representations; however, over time, they have evolved to handle arithmetic operations more proficiently. Approaches like GenBERT and NF-NSM have inserted numeric data directly into PLMs to enhance their mathematical performance.

Mathematical reasoning, on the other hand, involves solving complex problems that require logical thought processes. Recent studies show that LLMs can generate elaborate chains of thoughts when provided with exemplary reasoning examples, achieving higher success rates on various tasks.

LLM Methodologies

The methodologies employed in achieving mathematical proficiency can be categorized based on their underlying PLMs and LLMs structures. Among PLMs, autoregression (ALMs) and non-autoregression (NALMs) are two significant approaches used to comprehend and generate mathematical expressions. LLMs employ strategies such as instruction learning, tool-based methods, and chain-of-thought (CoT) techniques to improve mathematical reasoning capabilities. They draw on tools like symbolic solvers and computer programs to assist in problem-solving, while also leveraging training techniques like fine-tuning to bolster performance on certain arithmetic tasks.

Datasets and Challenges

Over 60 mathematical datasets have been compiled, which are divided into training datasets, benchmark datasets, and augmented datasets. These datasets are critical in both the training and evaluation of mathematical models and cover a wide range of complexity levels, from basic arithmetic to advanced theorem proving.

Despite the progress, challenges persist, including ensuring faithfulness in model output, enhancing multi-modal capabilities to handle non-textual mathematical information, managing uncertainty in calculations, devising robust evaluation metrics, and finding applications in educational settings as teaching aids.

Conclusion

The intersection of artificial intelligence and mathematical problem-solving is witnessing significant expansion, driven by the innovative capabilities of mathematical LLMs. By addressing existing hurdles and harnessing the potential of PLMs and LLMs, these models are poised to revolutionize the domain of mathematics and its numerous applications. This survey aims to ignite further research by providing a detailed account of current successes, areas for growth, and outlining prospective directions for advancements in this exciting field.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (11)
  1. Wentao Liu (87 papers)
  2. Hanglei Hu (4 papers)
  3. Jie Zhou (687 papers)
  4. Yuyang Ding (13 papers)
  5. Junsong Li (6 papers)
  6. Jiayi Zeng (3 papers)
  7. Mengliang He (3 papers)
  8. Qin Chen (57 papers)
  9. Bo Jiang (235 papers)
  10. Aimin Zhou (43 papers)
  11. Liang He (202 papers)
Citations (4)