TL;DR: We present rStar-Math to demonstrate that small language models (SLMs, 1.5B-7B) can rival or even surpass the math reasoning capability of OpenAI o1
Abstract: We present rStar-Math to demonstrate that small language models (SLMs) can rival or even surpass the math reasoning capability of OpenAI o1, without distillation from superior models. rStar-Math achieves this by exercising ``deep thinking'' through Monte Carlo Tree Search (MCTS), where a math policy SLM performs test-time search guided by an SLM-based process reward model. rStar-Math introduces three innovations to tackle the challenges in training the two SLMs: (1) a novel code-augmented CoT data synthesis method, which performs extensive MCTS rollouts to generate step-by-step verified reasoning trajectories used to train the policy SLM; (2) a novel process reward model training method that avoids na\"ive step-level score annotation, yielding a more effective process preference model (PPM); (3) a self-evolution recipe in which the policy SLM and PPM are built from scratch and iteratively evolved to improve reasoning capabilities. Through 4 rounds of self-evolution with millions of synthesized solutions for 747k math problems, rStar-Math boosts SLMs' math reasoning to state-of-the-art levels. On MATH benchmark, it improves Qwen2.5-Math-7B from 58.8% to 90.0%, surpassing o1-preview by +4.5%. On the USA Math Olympiad (AIME), rStar-Math solves an average of 53.3% (8/15) of problems, ranking among the top 20% of the brightest high school math students. Code and data are available at https://github.com/microsoft/rStar.
Lay Summary: rStar-Math shows that small AI models can match or outperform larger ones like OpenAI’s o1 in solving challenging math problems, without learning from them directly. It uses a technique called Monte Carlo Tree Search (MCTS) to explore different solution paths before choosing the best one. rStar-Math trains two small models: one for generating step-by-step solutions and another for scoring them. It introduces three key innovations: high-quality data generation through code-augmented rollouts, a more effective scoring model without step-by-step labels, and a self-improvement loop where both models evolve together. After several rounds of this self-improvement process, rStar-Math reaches state-of-the-art results. On the difficult MATH benchmark, it reaches the same level as OpenAI’s o1. On real USA Math Olympiad (AIME) exams, it solves 8 out of 15 problems, placing it in the top 20% of top-performing high school students.
Link To Code: https://github.com/microsoft/rStar
Primary Area: Deep Learning->Large Language Models
Keywords: LLM, Reasoning, Self-evolution
Submission Number: 6558
Loading