Abstract

Large Language Models (LLMs) have made impressive strides in understanding and generating natural language, but they still struggle with mathematical problem-solving, especially when it comes to tasks that require multi-step reasoning and precise calculations. This review looks at recent advancements aimed at improving LLMs’ performance in math, focusing on two main approaches: refining inference methods and adding external tools. Techniques like Chain-of-Thought (CoT) prompting, Program-Aided Language Models (PAL), and Toolformer have helped improve LLMs’ ability to handle complex math problems. These models rely on external tools, such as Python interpreters or calculators, to perform precise calculations, which has proven effective for solving problems in algebra, calculus, and other areas. Models like Minerva and Llemma, which are pre-trained specifically on mathematical content, can solve more advanced problems like differential equations without needing additional tools. However, challenges still exist, such as the reliance on external tools for exact calculations, difficulties with multi-step reasoning, and limited transparency in the models’ decision-making processes. Looking ahead, the integration of multi-modal capabilities, autonomous computation, and human feedback could further enhance LLMs’ mathematical abilities. With continued improvements, LLMs could transform problem-solving in fields like education, research, and finance.

Details

Title
Advances in Mathematical Reasoning with Large Language Models
Author
Zhao, Zihao
Section
Machine Learning & Deep Learning Algorithms
Publication year
2025
Publication date
2025
Publisher
EDP Sciences
ISSN
24317578
e-ISSN
22712097
Source type
Conference Paper
Language of publication
English
ProQuest document ID
3284871302
Copyright
© 2025. This work is licensed under https://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and conditions, you may use this content in accordance with the terms of the License.