HTML conversions sometimes display errors due to content that did not convert correctly from the source. This paper uses the following packages that are not yet supported by the HTML conversion tool. Feedback on these issues are not necessary; they are known and are being worked on.
Authors: achieve the best HTML results from your LaTeX submissions by following these best practices.
arXiv:2504.07086v1 [cs.LG] 09 Apr 2025
Report issue for preceding element
Andreas Hochlehnert1 āHardik Bhatnagar1ā āVishaal Udandarao1,2ā
Samuel AlbanieAmeya Prabhu1ā Matthias Bethge1ā
1Tübingen AI Center, University of Tübingen ā2 University of Cambridge equal contribution, ā core contributor, ā equal advising
Report issue for preceding element
Report issue for preceding element
Reasoning has emerged as the next major frontier for language models (LMs), with rapid advances from both academic and industrial labs. However, this progress often outpaces methodological rigor, with many evaluations relying on benchmarking practices that lack transparency, robustness, or statistical grounding. In this work, we conduct a comprehensive empirical study and find that current mathematical reasoning benchmarks are highly sensitive to subtle implementation choicesāincluding decoding parameters, random seeds, prompt formatting, and even hardware and software-framework configurations. Performance gains reported in recent studies frequently hinge on unclear comparisons or unreported sources of variance. To address these issues, we propose a standardized evaluation framework with clearly defined best practices and reporting standards. Using this framework, we reassess recent methods and find that reinforcement learning (RL) approaches yield only modest improvementsāfar below prior claimsāand are prone to overfitting, especially on small-scale benchmarks like AIMEā24. In contrast, supervised finetuning (SFT) methods show consistently stronger generalization. To foster reproducibility, we release all code, prompts, and model outputs, for reasoning benchmarks, establishing more rigorous foundations for future work.
Report issue for preceding element
Ā |
---|
\faGlobe Leaderboard\faDatabase [Eval Logs](https://huggingface.co/datasets/bethgelab/sober_reasoning/ āā) |
Figure 1: The Sombre State of LM Reasoning for Math.(left) when re-evaluating recent 1.5B reasoning-enhanced models on AIME-24 using a standardized framework (see SectionĀ 4), we find substantial drops to reported results in the original papers, (right) the observed improvements from recent methods (gray highlighted area) fall entirely within the variance range (orange box plots) of DeepSeek-R1 1.5B model performance.
This suggests that these methods do not significantly outperform the base modelāunderscoring the importance of rigorous, multi-seed evaluation protocols for obtaining reliable performance estimates.Report issue for preceding element
Report issue for preceding element
āThe first principle is that you must not fool yourself, and you are the easiest person to fool.ā
Report issue for preceding element
āRichard Feynman
Report issue for preceding element
Reasoning has become central to recent advances in large language models (LLMs), playing a key role in nearly all frontier systems (Jaech etĀ al., 2024; OpenAI, 2025; Meta-AI, 2025). Recent months have seen a surge of research focused on understanding and improving LLM reasoning, accompanied by several open-source tools and training strategies (seeĀ Li etĀ al. ( [2025b](https://arxiv.org/html/2504.07086v1#bib.bib55 āā)) for a survey). This momentum has sparked optimism that building capable, competitive reasoning models may soon be within reach.
Report issue for preceding element
However, as evaluation practices shape the direction and perceived progress of the field (Liao etĀ al., 2021; Marie etĀ al., 2021; Prabhu etĀ al., 2020; Colas etĀ al., 2018; Biderman etĀ al., [2024](https://arxiv.org/html/2504.07086v1#bib.bib11 āā)), their persistenceāespecially in reasoningācalls for renewed scrutiny and higher standards.
Report issue for preceding element
Motivated by a growing number of inconsistent empirical claims across the reasoning landscape, we conduct a rigorous investigation into the current state of reasoning benchmarks, focusing specifically on mathematical reasoningāone of the most widely used testbeds for evaluating algorithmic advances in this spaceĀ (HuggingFaceH4, 2024).
Report issue for preceding element
Our main finding is that many recent empirical conclusions may be overly optimistic and fail to generalize under careful re-evaluation. We identify a surprising degree of sensitivity in LLM-based reasoning pipelines to seemingly minor design choicesāranging from decoding parameters, prompt formatting, and random seeds to the hardware and software stacks used during evaluation (see TableĀ 1). Particularly concerning is the instability introduced by small benchmark sizes: for example, AIMEā24 and AMCā23 each contain only 30ā40 examples, making performance metrics highly volatileāwhere even one question can shift Pass@1 by over 3 percentage points. This leads to substantial variance across seeds, often resulting in double-digit performance swings that challenge the reliability of published results. In SectionĀ 3, we systematically analyze the root causes of this instability, including sampling variance, decoding configurations, evaluation frameworks, and hardware heterogeneity. We show that these factors can significantly distort conclusions if not carefully controlled.
Report issue for preceding element
In SectionĀ 4, we propose a set of best practices aimed at improving reproducibility and rigor in reasoning benchmarks. We also re-evaluate recent techniques using a standardized and reproducible evaluation stack. Our findings are soberingāreinforcement learning (RL) applied to distillation-based models such as DeepSeek-R1 yields little to no statistically significant gains. Some methods, such as OpenRS, show promising results in original reports, but fail to hold up under repeated evaluation. RL training on base models like Qwen2.5 Math does show stronger performance, but still often underperforms instruction-tuned counterparts.111We note that OpenReasoner-Zero is a consistent exception, achieving competitive performance. Furthermore, RL-trained models exhibit significant performance drops on newer benchmarks such as AIMEā25, echoing patterns of test set overfitting or āhill-climbingā observed in prior work (Golchin & Surdeanu, 2023; Dominguez-Olmedo etĀ al., [2024](https://arxiv.org/html/2504.07086v1#bib.bib26 āā)). In contrast, supervised fine-tuning (SFT) continues to deliver stable, generalizable improvements across benchmarks, underscoring its maturity as a training paradigm. These observations point to a critical need for more reliable and standardized evaluation protocols.
Report issue for preceding element
Taken together, in this work, we aim to provide not only a clearer assessment of where current methods stand, but also the tools and practices needed to make reasoning evaluation more transparent, robust, and reproducible. To this end, we open-source all code, prompts, and outputs to facilitate fair and accountable progress in this increasingly important area.
Report issue for preceding element
Report issue for preceding element
Language Model Reasoning (for Math). The recent releases of OpenAI-O1Ā (Jaech etĀ al., 2024) (in December 2024) and DeepSeek-R1Ā (DeepSeek-AI, 2025; Lightman etĀ al., 2023; Team, 2025), REINFORCE++Ā (Hu, 2025), DPO-VPĀ (Tu etĀ al., 2025), CPPOĀ (Lin etĀ al., 2025a) and GROĀ (Cai, 2025; Liu etĀ al., 2025b; Shao etĀ al., 2024), curriculumsĀ (Wen etĀ al., 2025b) and reward designĀ (Gao etĀ al., 2024a; Ma etĀ al., 2023; Xie etĀ al., 2025; Yu etĀ al., 2024; Sim & Chen, 2024; Yuan etĀ al., 2025). Further, some works also explored scaling up RL-based approaches to modalities beyond just language, including visionĀ (Ma etĀ al., 2025; Huang etĀ al., 2025; Chen etĀ al.,; Liu etĀ al., 2025c; Lin etĀ al., [2025b](https://arxiv.org/html/2504.07086v1#bib.bib60 āā)). In our work, we objectively re-evaluate the claims made by several of these recent works under a standardized lens, and find that many of the reported gains do not hold up strongly when pitted on a level-playing field against well-tuned baselines.
Report issue for preceding element
Sobering Studies on ML Progress. Machine learning is a field of rapid progress. Due to the lightning speed of papers coming out across the various sub-fields of machine learning, practitioners and researchers often fail to rigorously evaluate algorithmic progressĀ (Hutchinson etĀ al., 2022; Machado etĀ al., 2018; Balduzzi etĀ al., 2018; Cawley & Talbot, 2010; Prabhu etĀ al., 2024b; 2020) and test-time adaptationĀ (Press etĀ al., 2023; Gorsane etĀ al., 2022; Jordan etĀ al., 2020). Some works have even gone as far as suggesting that reliable benchmarking of RL-based methods is computationally infeasibleĀ (Jordan etĀ al., 2024; Nezhurina etĀ al., 2024), the type of tasks testedĀ (Yan etĀ al., 2025; Dominguez-Olmedo etĀ al., 2024), metrics usedĀ (Liu etĀ al., 2024; Kandpal etĀ al., 2023). Given such a volatile landscape, in this work, we aim to level the playing field across recent LM-methods that have been released and provide an objective look on the progress that the reasoning community has made. Our findings, which we discuss in the rest of the paper, are sobering at best.
Report issue for preceding element
Report issue for preceding element
Recent reasoning-focused language models are evaluated under highly heterogeneous conditionsāincluding differences in evaluation frameworks and hardware, number of random seeds, temperature, and nucleus sampling parameters (top_p) (see TableĀ 1). While prior work has examined the effect of sampling parameters in multiple-choiceĀ (Renze, 2024), the influence of these choices remains underexplored for open-ended reasoning modelsāparticularly those trained with reinforcement learning. In this section, we systematically assess how these evaluation design choices affect reported performance, and highlight the sources of variance that most impact the reliability of results.
Report issue for preceding element
Report issue for preceding element
We adopt a consistent experimental setup throughout this section, unless otherwise stated. Our analysis includes nine widely used models grouped into two commonly benchmarked size classes: 1.5B and 7B parameters. For the 1.5B class, we evaluate: DeepSeek-R1-Distill-1.5BĀ (DeepSeek-AI, 2025), II-1.5B-PreviewĀ (Intelligent Internet, 2025). Note that DeepScaleR-1.5B, II-1.5B-Preview, and the OpenRS models are all initialized from DeepSeek-R1-Distill-1.5B and subsequently finetuned via reinforcement learning (e.g., GRPOĀ (Shao etĀ al., 2024), and OpenThinker-7BĀ (Team, 2025), trained using supervised learning on reasoning traces derived from DeepSeek-R1. All models are benchmarked on three widely used datasets: AIMEā24Ā ( AI-MO,), and MATH500Ā (Hendrycks etĀ al., 2021) with the vllm backendĀ (Kwon etĀ al., [2023](https://arxiv.org/html/2504.07086v1#bib.bib52 āā)).
Report issue for preceding element
Table 1: Taxonomy of current open-weight reasoning models. For each model, we report the base model it was post-trained from and the exact type of post-training algorithm applied (RL vs SFT). Further, we note the evaluation framework that the original paper uses for reporting results along with the exact temperature, generation sequence length, and top_p sampling parameters used for AIME-24 evaluation, with the number of generations used for computing Pass@1 (K). It is evident that there is no clear standardization across different models with respect to evaluation frameworks used and the sampling parameters. This motivates the need to closely scrutinize the evaluations of current reasoning models.
Ā | Ā | Ā | Ā | Ā | Ā | Ā | Ā |
---|---|---|---|---|---|---|---|
Model | Algorithm | Base | Framework | Temp | Top_p | Seq. Len | K |
DeepSeek-R1-Distill-1.5B | SFT | Qwen2.5-Math-1.5B | ā | 0.6 | 0.95 | 32,768 | 64 |
DeepSeek-R1-Distill-7B | SFT | Qwen2.5-Math-7B | ā | 0.6 | 0.95 | 32,768 | 64 |
DeepSeek-R1-Distill-14B | SFT | Qwen2.5-14B | ā | 0.6 | 0.95 | 32,768 | 64 |
DeepSeek-R1-Distill-32B | SFT | Qwen2.5-32B | ā | 0.6 | 0.95 | 32,768 | 64 |
OpenThinker-32B | SFT | Qwen2.5-32B-Instruct | evalchemy | 0.7 | 0.8 | 32,768 | 5 |
Bespoke-Stratos-32B | SFT | Qwen2.5-32B-Instruct | evalchemy | 0.7 | 0.8 | 32,768 | 5 |
Bespoke-Stratos-7B | SFT | Qwen2.5-7B-Instruct | evalchemy | 0.7 | 0.8 | 32,768 | 5 |
s1.1-7B | SFT | Qwen2.5-7B-Instruct | lm-eval-harness | 0 | ā | 32,768 | 64 |
s1.1-32B | SFT | Qwen2.5-32B-Instruct | lm-eval-harness | 0 | ā | 32,768 | 64 |
LIMO | SFT | Qwen2.5-32B-Instruct | math-eval-harness | 0 | 1 | 32,768 | 1 |
MiniMath-R1-1.5B | SFT | DeepSeek-R1-Distill-1.5B | oumi-ai | ā | ā | ā | ā |
DeepScaleR-1.5B-Preview | RL | DeepSeek-R1-Distill-1.5B | verl | 0.6 | 0.95 | 32,768 | 16 |
Open-RS1 | RL | DeepSeek-R1-Distill-1.5B | lighteval | 0.6 | 0.95 | 32,768 | 32 |
Open-RS2 | RL | DeepSeek-R1-Distill-1.5B | lighteval | 0.6 | 0.95 | 32,768 | 32 |
Open-RS3 | RL | DeepSeek-R1-Distill-1.5B | lighteval | 0.6 | 0.95 | 32,768 | 32 |
II-Thought-1.5B-Preview | RL | DeepSeek-R1-Distill-1.5B | evalscope | 0.6 | 0.95 | 32,768 | 64 |
Oat-Zero-1.5B | RL | Qwen2.5-Math-1.5B | custom | 0 | 1 | 3,000 | 1 |
Oat-Zero-7B | RL | Qwen2.5-Math-7B | custom | 0 | 1 | 3,000 | 1 |
STILL-3-1.5B-preview | RL | DeepSeek-R1-Distill-1.5B | custom | 0.6 | 0.95 | 32,768 | 5 |
FastCurl-1.5B-Preview | RL | DeepSeek-R1-Distill-1.5B | verl | 0.6 | 0.95 | 32,768 | 16 |
LIMR | RL | Qwen2.5-Math-7B | custom | 0.4 | 0.95 | 3,072 | 4 |
Report issue for preceding element
Sampling Parameters: To systematically compare the impact of sampling parameters on accuracy, all experiments in this section were performed with a standardized configuration: temperature=0.8, top_p=0.9, and both max_model_len and max_new_tokens set to 32,768 tokens. This context length matches the limits of models such as OpenThinker-7B and S1.1-7B, although certain models (e.g., DeepSeek) support longer sequences of up to 131,072 tokens. We chose this standardized evaluation length to ensure comparability, with a detailed analysis of the influence of completion length presented in FigureĀ 9. Unless otherwise specified, results in this section are averaged over 10 random seeds for AIMEā24 and AMCā23, and 3 seeds for MATH500, following the recommendations from SectionĀ 3.2.1.
Report issue for preceding element
Report issue for preceding elementFigure 2: Accuracy varies significantly across random seeds. We find significantly high Pass@1 variation across 20 different random seeds for nine models on AIMEā24, AMCā23, and MATH500. Variance is particularly high on AIMEā24 (upto 15%) and AMCā23 (upto 13%) due to the small number of test samples, highlighting instability of single-seed evaluations.Report issue for preceding element
We begin by analyzing the variance induced purely by the random seed used during evaluationāan aspect often neglected in benchmarking practices. While recent work calls for statistical rigor (e.g., using error bars and multiple runs)Ā (Bowyer etĀ al., 2025; Madaan etĀ al.,.
Report issue for preceding element
Key Insight. Pass@1 values show surprisingly high standard deviationāranging from 5 to 15 percentage points across seeds. This issue is particularly severe for AIMEā24 and AMCā23, which have only 30 and 40 test samples respectively. A change in just one question shifts Pass@1 by 2.5ā3.3 percentage points.
Report issue for preceding element
Takeawayā1Ā Ā Single-seed evaluations on small datasets are highly unstable. Accurate reporting requires averaging over multiple seeds.Report issue for preceding element
Takeawayā2Ā Ā Small datasets such as AIME24 (30 samples) make model comparisons unreliable, as solving just one extra question already shifts pass@1 by Ā 3%. Variance from sampling parameters or random seeds can easily cause fluctuations of 1ā2 correct answers, leading to unstable rankings ā especially when models cluster around 30% performance.Report issue for preceding element
Report issue for preceding elementFigure 3: Bootstrapped seed averaging is reliable only beyond a threshold. We plot the variance of Mean Pass@1 scores on AIMEā24 when averaging over K=1š¾1K=1italic_K = 1 to K=10š¾10K=10italic_K = 10 seed runs, finding that the variance is extremely high for small Kš¾Kitalic_K and significantly reduced by K=10š¾10K=10italic_K = 10. This suggests that using multi-seed evaluations (Kā„10š¾10K\geq 10italic_K ā„ 10) would yield more stable estimates. For results on AMC23 and MATH500 see Figures 12 and 13 respectively.Report issue for preceding element
To mitigate high variance, recent work has adopted bootstrappingāaveraging multiple evaluation runs to stabilize results. For example, DeepSeek reports Pass@1 over 64 runs, while DeepScaleR uses 16. We study the effectiveness of this approach by bootstrapping estimates for AIMEā24 using 1 to 10 evaluation runs. FigureĀ 3 shows that while variance is extreme for K=1š¾1K=1italic_K = 1 and still large for K=2š¾2K=2italic_K = 2, it reduces sharply for Kā„10š¾10K\geq 10italic_K ā„ 10. Further analysis of variance across additional datasets is presented Figures 12 and 13.
Report issue for preceding element
Takeawayā3Ā Ā Bootstrapping over 10 runs substantially stabilizes Pass@1 estimates and should be considered a minimal standard for reliable evaluation.Report issue for preceding element
Report issue for preceding elementFigure 4: Higher temperatures yield better accuracies. We find across all three datasets, higher temperatures produce better peak accuracy but introduce instability, revealing a tradeoff between performance and reproducibility. Results obtained by varying temperature from 0 to 1 in increments of 0.1, while keeping top_p fixed at 0.9.Report issue for preceding element
Figure 5: Higher top_p values improve performance at no cost to stability. Across all datasets, we find that higher top_p values generally improve performance while preserving similar amounts of variance as lower top_p values. Results were obtained by varying top_p from 0 to 1 in increments of 0.1, while holding the temperature constant at 0.8.Report issue for preceding element
Reducing the temperature or increasing the nucleus sampling parameter (top_p) improves the accuracy of performance estimates without incurring additional computational cost. Figure 4 illustrate the impact of temperature and Figure 5 show that of top_p across multiple models and datasets. Notably, a more reproducible estimate is associated with significant drops in measured performance, highlighting a consistent tradeoff between reproducibility and high performance. We recommend optimizing the temperature for performance, and comparing the best parameter per model.
Report issue for preceding element
Additionally, we investigate the impact of the temperature and top_p hyperparameter as prior works often employ different temperature and top_p settings when comparing the same model. To isolate the impact of varying temperature and top_p, we averaged pass@1 across seeds and compute variation of this estimate across temperature and top_p in a boxplot. Figure 6 and 7 show the performance variation. We see that temperature-induced and top_p-induced fluctuations not only affect performance estimates but also introduce substantial variability in performance itself, which can lead to unfair comparisons when evaluating the same model across different temperatures.
Report issue for preceding element
Figure 6: Accuracies vary significantly across temperature values. Across nine different models and three datasets, we observe consistently large variations in performance (upto 15%) induced by changing the temperature. Results were obtained by varying the temperature from 0 to 1 in increments of 0.1, while holding top_p constant at 0.9.Report issue for preceding element
Figure 7: Accuracies vary significantly across top_p values. Across nine different models and three datasets, we observe consistently large variations in performance (upto 8%) induced by changing the top_p value. Results were obtained by varying top_p from 0 to 1 in increments of 0.1, while holding the temperature constant at 0.8.Report issue for preceding element
Takeawayā4Ā Ā Temperature and top_p can introduce substantial performance variationāespecially on small benchmarksāand should be set to each modelās optimal values to ensure fair and stable evaluation.Report issue for preceding element
Report issue for preceding element
(a) AIME24. Significant differences are observed in model performance across compute clusters.Report issue for preceding element
(b) AMC23. Similar variability is seen across hardware in AMC23 results.Report issue for preceding element
Figure 8: Performance variation across compute clusters. Accuracy differences emerge when the same models are evaluated across compute clusters for both AIME24 and AMC23 datasetsāthese large differences in performance also persist when evaluating 7B models.Report issue for preceding element
Performance can also vary due to non-obvious factors like hardware and evaluation frameworkāyet this is rarely acknowledged. Models are often tested on heterogeneous systems and evaluated using different toolchains.For example, S-1.1Ā (Muennighoff etĀ al., 2025), the OpenRS model suite uses lighteval(Fourrier etĀ al., 2023) for evaluation.
Report issue for preceding element
Hardware Variation. We evaluated the same model across five different compute clusters, each with varying GPU types and memory configurations. As shown in FigureĀ 8, performance varied by up to 8% for OpenRS-1.5B and 6% for DeepSeek-R1-Distill-7B on AIMEā24, with similar trends observed on AMCā23. While it is known that inference engines such as vLLM can be sensitive to hardware differencesĀ (vLLM Contributors, 2024) may introduce non-determinismāour results demonstrate that these effects can measurably impact benchmark accuracy, even when averaging over multiple seeds.
Report issue for preceding element
Evaluation across different Python frameworks. Evaluation results can vary based on the framework used, due to differences in prompt templates, inference engines (e.g., vLLMĀ (Kwon etĀ al., 2023), evalchemy(Guha etĀ al., 2024) and evalscope( [Alibaba ModelScope Community,](https://arxiv.org/html/2504.07086v1#bib.bib5 āā)).
Report issue for preceding element
To assess this impact, we compare lighteval and evalchemy, keeping all other variables fixed: model, dataset, hardware, decoding parameters, and random seeds (3 per model). For a fair comparison, we evaluated two models, DeepSeek-R1-Distill-1.5B and S1.1-7B, at their default temperature and top_p parameter values on a single GPU. We present results averaged over three seeds for higher robustness. As shown in TableĀ 2, framework-induced differences are generally small (1ā2pp) but can still affect model rankings in tightly clustered scenarios.
Report issue for preceding element
Model | lighteval | evalchemy |
---|---|---|
R1-Distill-1.5B | 26.6 | 26.6 |
S1.1-7B | 22.2 | 17.7 |
Table 2: AIME24 across frameworks.Report issue for preceding element
Overall, our findings underscore that significant performance variations can arise solely from differences in hardware and software configurations, emphasizing the need to standardize for reliable evaluations.
Report issue for preceding element
Takeawayā5Ā Ā Re-running the exact same experimental configurations across compute clusters and evaluation frameworks yields notably different results.Report issue for preceding element
Report issue for preceding elementFigure 9: Models are extremely sensitive to output token lengths. We sweep across different max_new_tokens (number of tokens that models are allowed to generate) for DeepScaleR-1.5B and DeepSeek-R1-Distill-1.5B/7B on three datasets and find that they are heavily sensitive to output length limits, with premature truncation degrading the performance.Report issue for preceding element
Maximum Output Tokens. FigureĀ 9 shows that reducing max_new_tokens harms performanceāespecially on long-form problems. This sensitivity varies by model and dataset. Although reducing this setting lowers cost, it may induce premature stopping, leading to incorrect answers.
Report issue for preceding element
Figure 10: Using no prompt templates yields worse performance. We compare Pass@1 scores across three prompt formats: (1) math-specific prompt with chat template, (2) default chat template only, and (3) no template. Instruction-tuned models perform best with structured prompts and templates; omitting templates leads to consistent performance drops.Report issue for preceding element
Prompt Format. Prompt formatting has a measurable impact on accuracy. As shown in FigureĀ 10, models perform best when using math-specific prompts and their native chat templates. Omitting templates leads to performance drops, particularly for instruction-tuned models. We compare accuracy under three different prompt settings (see TableĀ 5): (1) a math-specific prompt formatted using the modelās chat template, (2) only the modelās chat template with no additional prompt, and (3) no template at all, i.e., the question without any special tokens or instructions. Interestingly, while base models like Qwen2.5-Math may benefit from prompt-free setupsĀ (Liu etĀ al., [2025b](https://arxiv.org/html/2504.07086v1#bib.bib65 āā)), instruction-tuned models rely heavily on format alignment. Thus, maintaining consistent and format-aware prompting is essential for maximizing instruction-tuned model performance.
Report issue for preceding element
Takeawayā6Ā Ā It is critical to use large generation context lengths to avoid output truncation which can degrade performance; further, using correct prompt formats and chat templates is important for extracting best model performance.Report issue for preceding element
Report issue for preceding element
In this section, we standardize evaluation frameworks, propose best practices, and comprehensively evaluate existing methods.
Report issue for preceding element
Report issue for preceding element
We propose a set of best practices informed by our experiments and guided with current research insights:
Report issue for preceding element
Hardware and Software Stack Standardization: To promote reproducibility and facilitate future work, we release all code within a Docker container, along with step-by-step instructions for running experiments on Runpodās publicly accessible, on-demand GPU instances. This setup allows any researcher to replicate and extend our results under identical conditions.
Report issue for preceding element
Variance Estimates: For small benchmarks (e.g., AIMEā24), run evaluations with at least ten random seeds. Report the mean and standard deviation to quantify uncertainty and assess the statistical significance of performance differences.
Report issue for preceding element
Model-Specific Hyperparameter Optimization: Tune hyperparameters (such as temperature and top_p) separately for each model, then fix them across tasks to ensure consistency and fair comparisons.
Report issue for preceding element
Context Length and Prompt Template Selection: Ensure the context length is sufficiently largeāespecially for models with long reasoning chainsāto avoid premature truncation and under-reported accuracy. For instruction-tuned models, always use the appropriate chat template to match the expected input format.
Report issue for preceding element
Robust Answer Matching: We strongly recommend using a resilient answer extraction pipeline that handles parsing issues and evaluates expression equivalence, rather than relying on exact string matching. This reduces the likelihood of spurious gains from formatting artifacts.
Report issue for preceding element
Transparent Evaluation Protocols: We recommend to release code, prompts, and model outputs, and clearly document the evaluation stack. Report uncertainties (e.g., via standard deviations) and include both quantitative and qualitative analyses to enable thorough and reproducible comparisons.
Report issue for preceding element
Report issue for preceding element
We adopt a largely consistent experimental setup with prior work, with the key difference being our use of publicly accessible cloud instances from Runpod222 https://www.runpod.io/pricing across six benchmarks: AIMEā24Ā ( AI-MO,), AMCā23 (Knovel Engineering, 2025), Minerva (Lewkowycz etĀ al., 2022). All experiments are conducted using the LightEval framework (Fourrier etĀ al., [2023](https://arxiv.org/html/2504.07086v1#bib.bib29 āā)) (0.8.1) with a vLLM backend, repeated across ten random seeds for AIMEā24, AIMEā25, AMCā23 and three random seeds for the rest. Depending on the base model architecture, we set the maximum number of new tokens (e.g., 4096 for QwenMath-based models), apply optimal hyperparameters, and use the appropriate chat template. LightEvalās LaTeX-based answer extraction and evaluation pipeline ensures reliable and consistent result parsing and correctness matching, similar to math-verify.
Report issue for preceding element
Report issue for preceding element
Model | AIMEā24 | AIMEā25 | AMCā23 | MATH500 | Minerva | Olympiad |
Based on: Deepseek R1 Distill Qwen 1.5B (RL) | Ā | Ā | Ā | Ā | Ā | Ā |
R1-Distill (DeepSeek-AI, [2025](https://arxiv.org/html/2504.07086v1#bib.bib23 āā)) | 28.7±plus-or-minus\pm±4.8 | 22.3±plus-or-minus\pm±5.2 | 71.5±plus-or-minus\pm±3.9 | 84.9±plus-or-minus\pm±0.3 | 30.5±plus-or-minus\pm±1.0 | 52.4±plus-or-minus\pm±0.4 |
L1-Exact (Aggarwal & Welleck, [2025](https://arxiv.org/html/2504.07086v1#bib.bib2 āā)) | 24.4±plus-or-minus\pm±3.3 | 22.3±plus-or-minus\pm±4.2 | 70.5±plus-or-minus\pm±3.7 | 86.6±plus-or-minus\pm±0.8 | 31.5±plus-or-minus\pm±1.7 | 52.5±plus-or-minus\pm±1.3 |
L1-Max (Aggarwal & Welleck, [2025](https://arxiv.org/html/2504.07086v1#bib.bib2 āā)) | 27.7±plus-or-minus\pm±4.2 | 21.0±plus-or-minus\pm±5.0 | 73.2±plus-or-minus\pm±6.0 | 84.7±plus-or-minus\pm±0.1 | 33.3±plus-or-minus\pm±0.9 | 52.3±plus-or-minus\pm±0.6 |
Open-RS1 (Dang & Ngo, [2025](https://arxiv.org/html/2504.07086v1#bib.bib20 āā)) | 28.9±plus-or-minus\pm±6.0 | 21.3±plus-or-minus\pm±4.2 | 75.0±plus-or-minus\pm±3.3 | 85.1±plus-or-minus\pm±0.8 | 30.4±plus-or-minus\pm±0.2 | 53.2±plus-or-minus\pm±1.9 |
Open-RS2 (Dang & Ngo, [2025](https://arxiv.org/html/2504.07086v1#bib.bib20 āā)) | 31.3±plus-or-minus\pm±7.7 | 22.7±plus-or-minus\pm±5.6 | 73.0±plus-or-minus\pm±5.7 | 84.1±plus-or-minus\pm±0.2 | 29.2±plus-or-minus\pm±1.1 | 53.7±plus-or-minus\pm±0.6 |
Open-RS3 (Dang & Ngo, [2025](https://arxiv.org/html/2504.07086v1#bib.bib20 āā)) | 29.7±plus-or-minus\pm±4.6 | 24.7±plus-or-minus\pm±6.5 | 69.2±plus-or-minus\pm±5.5 | 84.2±plus-or-minus\pm±1.1 | 28.6±plus-or-minus\pm±2.3 | 51.8±plus-or-minus\pm±0.8 |
STILL-3 (Min etĀ al., [2024](https://arxiv.org/html/2504.07086v1#bib.bib76 āā)) | 34.7±plus-or-minus\pm±5.5 | 24.0±plus-or-minus\pm±6.4 | 72.5±plus-or-minus\pm±5.4 | 86.6±plus-or-minus\pm±1.9 | 30.0±plus-or-minus\pm±0.6 | 53.9±plus-or-minus\pm±1.5 |
II-Thought (Intelligent Internet, [2025](https://arxiv.org/html/2504.07086v1#bib.bib44 āā)) | 32.0±plus-or-minus\pm±5.9 | 24.0±plus-or-minus\pm±4.1 | 79.5±plus-or-minus\pm±5.1 | 86.6±plus-or-minus\pm±0.6 | 31.7±plus-or-minus\pm±0.6 | 54.9±plus-or-minus\pm±0.4 |
FastCuRL (Song etĀ al., [2025](https://arxiv.org/html/2504.07086v1#bib.bib99 āā)) | 36.3±plus-or-minus\pm±4.3 | 27.0±plus-or-minus\pm±3.7 | 78.8±plus-or-minus\pm±4.1 | 87.9±plus-or-minus\pm±1.2 | 30.8±plus-or-minus\pm±1.4 | 56.5±plus-or-minus\pm±0.6 |
DeepScaleR (Luo etĀ al., [2025](https://arxiv.org/html/2504.07086v1#bib.bib67 āā)) | 37.0±plus-or-minus\pm±6.6 | 30.3±plus-or-minus\pm±4.3 | 76.2±plus-or-minus\pm±4.6 | 87.8±plus-or-minus\pm±1.0 | 31.0±plus-or-minus\pm±1.5 | 55.5±plus-or-minus\pm±1.1 |
Based on: Deepseek R1 Distill Qwen 7B (RL) | Ā | Ā | Ā | Ā | Ā | Ā |
R1-Distill (DeepSeek-AI, [2025](https://arxiv.org/html/2504.07086v1#bib.bib23 āā)) | 52.3±plus-or-minus\pm±6.3 | 39.0±plus-or-minus\pm±5.9 | 91.5±plus-or-minus\pm±2.7 | 94.1±plus-or-minus\pm±0.3 | 40.1±plus-or-minus\pm±0.4 | 67.3±plus-or-minus\pm±0.1 |
Light-R1 (Wen etĀ al., [2025a](https://arxiv.org/html/2504.07086v1#bib.bib109 āā)) | 53.0±plus-or-minus\pm±4.8 | 41.0±plus-or-minus\pm±3.5 | 90.0±plus-or-minus\pm±3.1 | 93.5±plus-or-minus\pm±0.5 | 41.3±plus-or-minus\pm±1.3 | 68.0±plus-or-minus\pm±1.2 |
Based on: Qwen2.5 Math 1.5B (RL) | Ā | Ā | Ā | Ā | Ā | Ā |
Math (Base) (Yang etĀ al., [2024b](https://arxiv.org/html/2504.07086v1#bib.bib115 āā)) | 11.3±plus-or-minus\pm±3.6 | 5.7±plus-or-minus\pm±2.7 | 44.0±plus-or-minus\pm±4.9 | 51.7±plus-or-minus\pm±5.5 | 11.3±plus-or-minus\pm±2.2 | 26.0±plus-or-minus\pm±0.6 |
Oat-Zero (Liu etĀ al., [2025a](https://arxiv.org/html/2504.07086v1#bib.bib64 āā)) | 16.0±plus-or-minus\pm±3.2 | 6.7±plus-or-minus\pm±3.4 | 52.5±plus-or-minus\pm±2.9 | 73.5±plus-or-minus\pm±1.7 | 26.3±plus-or-minus\pm±0.8 | 37.2±plus-or-minus\pm±1.3 |
Math (Instruct) (Yang etĀ al., [2024b](https://arxiv.org/html/2504.07086v1#bib.bib115 āā)) | 12.0±plus-or-minus\pm±1.7 | 11.7±plus-or-minus\pm±5.7 | 54.8±plus-or-minus\pm±5.3 | 74.7±plus-or-minus\pm±0.5 | 26.7±plus-or-minus\pm±1.8 | 37.9±plus-or-minus\pm±0.2 |
Based on: Qwen2.5 Math 7B (RL) | Ā | Ā | Ā | Ā | Ā | Ā |
Math (Base) (Yang etĀ al., [2024b](https://arxiv.org/html/2504.07086v1#bib.bib115 āā)) | 20.7±plus-or-minus\pm±3.8 | 8.7±plus-or-minus\pm±3.9 | 56.2±plus-or-minus\pm±5.7 | 64.3±plus-or-minus\pm±0.5 | 17.3±plus-or-minus\pm±1.9 | 29.0±plus-or-minus\pm±0.5 |
SimpleRL-Zoo (Zeng etĀ al., [2025b](https://arxiv.org/html/2504.07086v1#bib.bib121 āā)) | 22.7±plus-or-minus\pm±5.2 | 10.7±plus-or-minus\pm±3.4 | 62.2±plus-or-minus\pm±3.6 | 76.9±plus-or-minus\pm±1.8 | 30.1±plus-or-minus\pm±2.8 | 39.3±plus-or-minus\pm±0.6 |
LIMR (Li etĀ al., [2025a](https://arxiv.org/html/2504.07086v1#bib.bib54 āā)) | 30.7±plus-or-minus\pm±3.2 | 7.8±plus-or-minus\pm±3.3 | 62.2±plus-or-minus\pm±3.4 | 76.5±plus-or-minus\pm±0.4 | 34.9±plus-or-minus\pm±1.3 | 39.3±plus-or-minus\pm±0.9 |
Oat-Zero (Liu etĀ al., [2025a](https://arxiv.org/html/2504.07086v1#bib.bib64 āā)) | 28.0±plus-or-minus\pm±3.1 | 8.8±plus-or-minus\pm±2.5 | 66.2±plus-or-minus\pm±3.6 | 79.4±plus-or-minus\pm±0.3 | 34.4±plus-or-minus\pm±1.4 | 43.8±plus-or-minus\pm±1.1 |
Math (Instruct) (Yang etĀ al., [2024b](https://arxiv.org/html/2504.07086v1#bib.bib115 āā)) | 15.7±plus-or-minus\pm±3.9 | 10.7±plus-or-minus\pm±3.8 | 67.0±plus-or-minus\pm±3.9 | 82.9±plus-or-minus\pm±0.1 | 35.0±plus-or-minus\pm±0.6 | 41.3±plus-or-minus\pm±0.9 |
Based on: Qwen2.5 1.5B (RL) | Ā | Ā | Ā | Ā | Ā | Ā |
Qwen (Base) (Yang etĀ al., [2024a](https://arxiv.org/html/2504.07086v1#bib.bib114 āā)) | 0.0±plus-or-minus\pm±0.0 | 0.0±plus-or-minus\pm±0.0 | 2.5±plus-or-minus\pm±2.5 | 3.3±plus-or-minus\pm±1.5 | 1.8±plus-or-minus\pm±0.4 | 1.5±plus-or-minus\pm±0.5 |
SimpleRL-Zoo (Zeng etĀ al., [2025b](https://arxiv.org/html/2504.07086v1#bib.bib121 āā)) | 0.3±plus-or-minus\pm±1.1 | 0.3±plus-or-minus\pm±1.1 | 13.2±plus-or-minus\pm±4.7 | 12.0±plus-or-minus\pm±6.5 | 4.0±plus-or-minus\pm±2.4 | 4.2±plus-or-minus\pm±2.0 |
Qwen (Instruct) (Yang etĀ al., [2024a](https://arxiv.org/html/2504.07086v1#bib.bib114 āā)) | 1.3±plus-or-minus\pm±1.7 | 0.7±plus-or-minus\pm±1.4 | 26.2±plus-or-minus\pm±4.8 | 57.5±plus-or-minus\pm±1.1 | 19.4±plus-or-minus\pm±1.3 | 20.3±plus-or-minus\pm±1.1 |
Based on: Qwen2.5 7B (RL) | Ā | Ā | Ā | Ā | Ā | Ā |
Qwen (Base) (Yang etĀ al., [2024a](https://arxiv.org/html/2504.07086v1#bib.bib114 āā)) | 3.3±plus-or-minus\pm±3.3 | 0.0±plus-or-minus\pm±0.0 | 30.0±plus-or-minus\pm±9.0 | 64.6±plus-or-minus\pm±1.0 | 25.7±plus-or-minus\pm±0.9 | 30.1±plus-or-minus\pm±1.2 |
SimpleRL-Zoo (Zeng etĀ al., [2025b](https://arxiv.org/html/2504.07086v1#bib.bib121 āā)) | 14.0±plus-or-minus\pm±2.1 | 4.3±plus-or-minus\pm±2.7 | 58.0±plus-or-minus\pm±1.6 | 77.9±plus-or-minus\pm±0.8 | 33.0±plus-or-minus\pm±0.2 | 39.0±plus-or-minus\pm±0.1 |
Open Reasoner Zero (Hu etĀ al., [2025](https://arxiv.org/html/2504.07086v1#bib.bib40 āā)) | 19.7±plus-or-minus\pm±2.9 | 15.7±plus-or-minus\pm±2.7 | 59.5±plus-or-minus\pm±4.5 | 83.9±plus-or-minus\pm±1.1 | 31.6±plus-or-minus\pm±1.3 | 47.6±plus-or-minus\pm±1.7 |
Qwen (Instruct) | 12.3±plus-or-minus\pm±3.2 | 7.3±plus-or-minus\pm±3.4 | 52.8±plus-or-minus\pm±4.8 | 77.1±plus-or-minus\pm±1.2 | 34.9±plus-or-minus\pm±1.0 | 38.7±plus-or-minus\pm±1.0 |
Based on: Qwen2.5 7B (SFT) | Ā | Ā | Ā | Ā | Ā | Ā |
Qwen (Instruct) (Yang etĀ al., [2024a](https://arxiv.org/html/2504.07086v1#bib.bib114 āā)) | 12.3±plus-or-minus\pm±3.2 | 7.3±plus-or-minus\pm±3.4 | 52.8±plus-or-minus\pm±4.8 | 77.1±plus-or-minus\pm±1.2 | 34.9±plus-or-minus\pm±1.0 | 38.7±plus-or-minus\pm±1.0 |
Eurus2 Prime (Cui etĀ al., [2025](https://arxiv.org/html/2504.07086v1#bib.bib19 āā)) | 17.8±plus-or-minus\pm±2.2 | 14.0±plus-or-minus\pm±1.7 | 63.0±plus-or-minus\pm±3.9 | 80.1±plus-or-minus\pm±0.1 | 37.5±plus-or-minus\pm±1.0 | 43.9±plus-or-minus\pm±0.3 |
s1.1 (Muennighoff etĀ al., [2025](https://arxiv.org/html/2504.07086v1#bib.bib78 āā)) | 19.0±plus-or-minus\pm±3.2 | 21.0±plus-or-minus\pm±5.5 | 59.5±plus-or-minus\pm±3.7 | 80.8±plus-or-minus\pm±0.6 | 37.5±plus-or-minus\pm±1.1 | 48.2±plus-or-minus\pm±1.4 |
Bespoke Stratos (Bespoke Labs, [2024](https://arxiv.org/html/2504.07086v1#bib.bib10 āā)) | 20.3±plus-or-minus\pm±4.3 | 18.0±plus-or-minus\pm±4.8 | 60.2±plus-or-minus\pm±4.9 | 84.7±plus-or-minus\pm±0.5 | 39.1±plus-or-minus\pm±1.3 | 51.9±plus-or-minus\pm±1.1 |
OpenThinker (Team, [2025](https://arxiv.org/html/2504.07086v1#bib.bib102 āā)) | 30.5±plus-or-minus\pm±6.2 | 26.0±plus-or-minus\pm±4.4 | 71.4±plus-or-minus\pm±3.9 | 88.3±plus-or-minus\pm±1.4 | 37.9±plus-or-minus\pm±3.8 | 55.6±plus-or-minus\pm±1.4 |
OpenR1 (Face, [2025](https://arxiv.org/html/2504.07086v1#bib.bib27 āā)) | 48.3±plus-or-minus\pm±8.9 | 35.5±plus-or-minus\pm±4.2 | 86.0±plus-or-minus\pm±4.5 | ā | ā | ā |
OpenThinker2 (Team, [2025](https://arxiv.org/html/2504.07086v1#bib.bib102 āā)) | 53.0±plus-or-minus\pm±4.6 | 41.0±plus-or-minus\pm±5.0 | 87.0±plus-or-minus\pm±3.5 | 81.6±plus-or-minus\pm±0.7 | 33.9±plus-or-minus\pm±0.2 | 46.9±plus-or-minus\pm±1.3 |
Table 3: A Standardized and Sober Compilation of LM-Reasoning Results. We report Pass@1 accuracy (mean ±plus-or-minus\pm± std) of all models across six math reasoning benchmarks under a standardized evaluation setupāresults are averaged over ten seeds for AIME and AMC, and three seeds for the rest, using the LightEval framework with best hyperparameters tuned per method, 32,768 context lengths for all except 4,096 for Math models, and appropriate prompt templates. RL- and SFT-based variants are evaluated relative to their respective base or instruction-tuned models. Main takeawaysā(1) RL-trained methods do not yield meaningful performance gains, (2) SFT on reasoning traces yields significant generalization.Report issue for preceding element
We present experimental results in TableĀ 3, and analyze different aspects of the results.
Report issue for preceding element
RL-training on R1-Distill We evaluated several reinforcement learning (RL) approaches (e.g., GRPO) using the DeepSeek R1-Distill-1.5B model. We first observe that none of the L1 models (Aggarwal & Welleck, 2025) reported strong gains (10ā15%) on AIME, AMC, and OlympiadBench. However, our replication showed no statistically significant improvements over the R1 - Distill baseline. Same case held for Still-3 and Light-R1 model, which showed no significant improvement over the R1-Distill baseline. II-Thought and FastCurl yield modest improvements across benchmarks, especially over AIMEā24 but the observed gains did not carry over significantly to AIMEā25 indicating overfitting to existing benchmarks. Only DeepscaleR demonstrated robust, significant improvements across benchmarks.
Report issue for preceding element
Takeawayā1Ā Ā Most RL-trained variants of the DeepSeek R1-Distill model do not yield meaningful performance improvements (except DeepscaleR), suggesting that a reliable and scalable RL training recipes are still lacking.Report issue for preceding element
RL Training on Qwen2.5 Math and Base Models: We next analyze RL training applied to the Qwen2.5 Base and Qwen2.5 Math Base models, a trend trying to replicate gains by Deepseek-R1 Zero. Unlike the R1-Distill results, RL training with Oat-Zero, LIMR, and SimpleRL-Zoo consistently produced statistically significant gains over the base model, especially across Math500, Minerva and OlympiadBench benchmarks. This indicates that RL-based approaches can indeed offer substantial improvements given a base model instead of a distilled R1 model. However, these gains remained smaller than those achieved via instruction tuning in the original Qwen papers, suggesting that instruction tuning alone may be sufficient to far surpass current gains from RL methods in this setting. We also observed that the improvements on AIMEā24 were also significant, but did not carry over to AIMEā25 indicating a troubling overfitting trend. Notably, Open Reasoner-Zero-7B was the only RL-trained model to consistently outperform the instruct-tuned baseline by large margins across all benchmarks.
Report issue for preceding element
Takeawayā2Ā Ā While RL-trained methods can often substantially improve base model performance, instruction tuning remains superior (except Open Reasoner Zero), suggesting again that a reliable and scalable RL training recipes are still lacking.Report issue for preceding element
Effectiveness of Supervised Finetuning. We assessed supervised finetuning methods like s1.1, Eurus2 Prime, Bespoke Stratos, OpenR1 and OpenThinker models, which further refine instruction-tuned models using reasoning traces. Supervised methods consistently outperformed the instruct-tuned baseline across all benchmarks (even Minerva) and generalized comparatively well to AIMEā25. The performance improvements from OpenThinker were especially notable. These results underscore the maturity and effectiveness of SFT when training recipes are scaled to large datasets.
Report issue for preceding element
Takeawayā3Ā Ā Supervised finetuning on reasoning traces from larger models yields significant, generalizable gains across benchmarks with progress over time successfully replicated ā highlighting its robustness and maturity as a training paradigm.Report issue for preceding element
Overfitting and Generalization We now examine the overfitting by comparing performance on AIMEā24 versus the more challenging AIMEā25. RL-trained models showed a pronounced performance drop between the two, indicating overfitting to the training distribution. In contrast, supervised fine-tuning (SFT) models maintained consistent improvements, suggesting better generalization. Openthinker2 showed significant degradation compared to Openthinker across benchmarks not provided in their blogpost, indicating overfitting via data-curation. This highlights a gap in current evaluation protocols, and a need to assess out-of-distribution generalization for reasoning models.
Report issue for preceding element
Takeawayā4Ā Ā Current RL-based approaches are very susceptible to overfitting, emphasizing the need for more rigorous out-of-distribution benchmarks. By comparison, SFT models exhibit stronger generalization and resilience.Report issue for preceding element
Report issue for preceding element
We further investigate two recently noted phenomena to see if they replicate in our experiments: (1) how response length correlates with performance, and (2) the decline in response diversity following reasoning-focused training.
Report issue for preceding element
Report issue for preceding elementFigure 11: Response Length vs. Accuracy. Histogram of correct vs. incorrect responses by response length, averaged over random seeds across AIME24, AIME25, AMC23, MATH500, Minerva and OlympiadBench benchmarks. Longer outputs tend to be more error-prone, even in complete responses not close to the maximum sequence length.Report issue for preceding element
Recent research (Wang etĀ al., [2025](https://arxiv.org/html/2504.07086v1#bib.bib108 āā)) suggests that incorrect answers often have disproportionately long reasoning chains. We first verify whether this finding holds in our setting, and then we explore possible explanations behind the observed variations.
Report issue for preceding element
Do longer responses indicate a higher likelihood of an incorrect answer? We compare the distribution of response lengths for correct and incorrect answers across 6 datasets (AIME24, AIME25, AMC23, MATH500, Minerva and OlympiadBench) averaged across random seeds for each model. FigureĀ 11 shows histograms of the average number of responses per seed, binned by response length. A clear trend emerges: shorter responses are significantly more likely to be correct, while longer responses become progressively more error-prone. This pattern is consistent across all seeds and is especially pronounced for responses exceeding 10,000 tokens. We now address two questions:
Report issue for preceding element
Report issue for preceding element
Yes. We find the trend is consistent across both RL- and SFT-trained models (additional figures provided in Appendix figuresĀ 17 and Ā 18 ). We consistently observe that the effect is more pronounced in RL-trained models (displayed on the left) than in SFT-trained models (displayed on the right). As detailed in the Appendix, both the Qwen 2.5 Math base exhibit a slight shift in length, though this shift is notably more evident in R1-distill and subsequent RL-trained models.
Report issue for preceding element
Report issue for preceding element
Although responses nearing the 32,000-token limit are almost always incorrect (due to limited context-length), this trend persists even for complete responses which are shorterā Longer responses are associated with a higher likelihood of being incorrect.
Report issue for preceding element
Takeawayā5Ā Ā Longer responses correlate with a greater chance of error, response length is a practical heuristic for consensus@k, identifying low-confidence or failed generations.Report issue for preceding element
Report issue for preceding element
Ā | Ā | AIMEā24 | AIMEā25 | AMCā23 | Ā | Ā | Ā | Ā | Ā | Ā |
---|---|---|---|---|---|---|---|---|---|---|
Model | Baseline | Ī“ā¢@ā¢1šæ@1\delta@1italic_Ī“ @ 1 | Ī“ā¢@ā¢5šæ@5\delta@5italic_Ī“ @ 5 | Ī“ā¢@ā¢10šæ@10\delta@10italic_Ī“ @ 10 | Ī“ā¢@ā¢1šæ@1\delta@1italic_Ī“ @ 1 | Ī“ā¢@ā¢5šæ@5\delta@5italic_Ī“ @ 5 | Ī“ā¢@ā¢10šæ@10\delta@10italic_Ī“ @ 10 | Ī“ā¢@ā¢1šæ@1\delta@1italic_Ī“ @ 1 | Ī“ā¢@ā¢5šæ@5\delta@5italic_Ī“ @ 5 | Ī“ā¢@ā¢10šæ@10\delta@10italic_Ī“ @ 10 |
Open-RS3 | R1-Distill | +1.51.5{\color[rgb]{0,0.5,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.5,0}+1.5}+ 1.5 | -0.6 | +1.5 | -1.0 | +0.4 | +1.4 | -0.9 | -0.2 | +0.4 |
DeepScaleR | R1-Distill | +9.09.0{\color[rgb]{0,0.5,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.5,0}+9.0}+ 9.0 | +0.6 | +0.2 | +4.44.4{\color[rgb]{0,0.5,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.5,0}+4.4}+ 4.4 | +3.6 | +2.2 | +6.16.1{\color[rgb]{0,0.5,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.5,0}+6.1}+ 6.1 | -1.8 | -1.7 |
S1.1-7B | Qwen-Instruct | +5.75.7{\color[rgb]{0,0.5,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.5,0}+5.7}+ 5.7 | +10.9 | +13.5 | +11.911.9{\color[rgb]{0,0.5,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.5,0}+11.9}+ 11.9 | +10.5 | +10.4 | +5.85.8{\color[rgb]{0,0.5,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.5,0}+5.8}+ 5.8 | +9.6 | +9.7 |
II-Thought | R1-Distill | +2.52.5{\color[rgb]{0,0.5,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.5,0}+2.5}+ 2.5 | -3.5 | -3.6 | +0.80.8{\color[rgb]{0,0.5,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.5,0}+0.8}+ 0.8 | +0.5 | +1.2 | +6.36.3{\color[rgb]{0,0.5,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0.5,0}+6.3}+ 6.3 | +0.7 | +0.2 |
Table 4: RL-trained models do not show a diversity collapseĀ ( [Dang etĀ al.,](https://arxiv.org/html/2504.07086v1#bib.bib21 āā)). We report the delta between Pass@k of RL-trained models and their corresponding baselines. Unlike reported in prior work, we observe no significant phenomenon of diversity collapse: Ī“ā¢@ā¢5šæ@5\delta@5italic_Ī“ @ 5 and Ī“ā¢@ā¢10šæ@10\delta@10italic_Ī“ @ 10 are largely positive, and are negative at similar rates as Ī“ā¢@ā¢1šæ@1\delta@1italic_Ī“ @ 1.Report issue for preceding element
[Dang etĀ al.](https://arxiv.org/html/2504.07086v1#bib.bib21 āā) has reported a counterintuitive phenomenon in reasoning models: improvements in Pass@1 achieved through supervised fine-tuning or RL can reduce Pass@k performance due to diminished output diversityāa phenomenon termed diversity collapse. Theoretical analyses attribute this collapse to the model concentrating too much probability mass on a single reasoning path, while current decoding strategies fail to recover the lost diversity.
Report issue for preceding element
To examine these claims, we compare the Pass@k performance (for kā1,5,10š1510k\in{1,5,10}italic_k ā 1 , 5 , 10) of RL-trained models against their corresponding base models (e.g., DeepSeek-R1-Distill-Qwen-1.5B) across all datasets. TableĀ 4 shows the delta in Pass@k relative to each methodās base model.
Report issue for preceding element
Findings. We do not observe a consistent diversity collapse. Gains in Pass@1 generally come with improvements in Pass@k, though the magnitude of these gains varies. When Pass@k performance does drop, it does so alongside (rather than independently of) occasional declines in Pass@1, providing no support for the diversity collapse hypothesis.
Report issue for preceding element
Takeawayā6Ā Ā Standard decoding strategies appear sufficient to capture the modelās full distribution over valid reasoning paths, counter to the diversity collapse hypothesis.Report issue for preceding element
Report issue for preceding element
Our study shows that much of the perceived progress in LLM-based reasoning, particularly in mathematical benchmarks, rests on unstable and often non-reproducible foundations. We find that minor differences in sampling parameters, prompt formatting, hardware, and software configurations can lead to major shifts in reported performanceācasting doubt on many recent empirical claims. Reinforcement learning methods, while promising in theory, offer at best modest gains in practice and are prone to overfitting, especially on small benchmarks like AIMEā24. In contrast, supervised finetuning continues to deliver consistent, generalizable improvements across a wide range of benchmarks and model sizes.
Report issue for preceding element
To address these challenges, we advocate for standardized, transparent evaluation protocols. Our open-sourced framework, complete with Dockerized environments, seed-averaged metrics, and robust answer matching, provides reproducible foundations for future research. We hope this work shifts the focus from leaderboard chasing to methodological rigorāensuring that future claims of progress in reasoning are both meaningful and measurable.
Report issue for preceding element
Report issue for preceding element
Andreas, Vishaal and Ameya conceived the project. Andreas and Hardik co-led the experiments, with Vishaal and Ameya advising the experimental design. The manuscript was written by Andreas, Hardik, Vishaal and Ameya. Matthias and Samuel provided helpful feedback and advice throughout the project.
Report issue for preceding element
Report issue for preceding element
The authors would like to thank (in alphabetical order): Matteo Farina, Shyamgopal Karthik, Nikhil Parthasarathy, Shiven Sinha, Joschka Strüber, Thaddäus Wiedemer for helpful feedback on the draft. AH acknowledges funding by the Federal Ministry of Education and Research (BMBF), FKZ: 01IS24079A. HB has received funding from the Digital Europe Programme under grant agreement No 101195233 (OpenEuroLLM). AH, HB and VU thank the International Max Planck Research School for Intelligent Systems (IMPRS-IS) for support. VU also thanks the European Laboratory for Learning and Intelligent Systems (ELLIS) PhD program for support. VU was supported by a Google PhD Fellowship in Machine Intelligence. AP and MB acknowledge financial support by the Federal Ministry of Education and Research (BMBF), FKZ: 011524085B and Open Philanthropy Foundation funded by the Good Ventures Foundation. This work was supported by the Digital Europe Programme under grant agreement No 101195233 (OpenEuroLLM).
Report issue for preceding element
Report issue for preceding element
Deep reinforcement learning at the edge of the statistical precipice.
Advances in neural information processing systems, 34:29304ā29320, 2021.
L1: Controlling how long a reasoning model thinks with reinforcement learning.
arXiv preprint arXiv:2503.04697, 2025.
AIMO Validation AIME Dataset.
AIMO Validation AMC Dataset.
[https://huggingface.co/datasets/AI-MO/aimo-validation-amc](https://huggingface.co/datasets/AI-MO/aimo-validation-amc āā), 2024.
Accessed: 2025-03-29.
Evalscope documentation.
[https://evalscope.readthedocs.io/en/latest/](https://evalscope.readthedocs.io/en/latest/ āā).
Accessed: 2025-03-29.
What matters in on-policy reinforcement learning? a large-scale empirical study.
arXiv preprint arXiv:2006.05990, 2020.
Claude 3.7 Sonnet System Card, 2025.
URL [https://assets.anthropic.com/m/785e231869ea8b3b/original/claude-3-7-sonnet-system-card.pdf](https://assets.anthropic.com/m/785e231869ea8b3b/original/claude-3-7-sonnet-system-card.pdf āā).
Accessed: 2025-03-29.
Optimizing large language model hyperparameters for code generation.
arXiv preprint arXiv:2408.10577, 2024.
Re-evaluating evaluation.
Advances in Neural Information Processing Systems, 31, 2018.
Bespoke-stratos-7b.
[https://huggingface.co/bespokelabs/Bespoke-Stratos-7B](https://huggingface.co/bespokelabs/Bespoke-Stratos-7B āā), 2024.
Accessed: 2025-03-29.
Lessons from the trenches on reproducible evaluation of language models.
arXiv preprint arXiv:2405.14782, 2024.
Position: Donāt use the CLT in LLM evals with fewer than a few hundred datapoints.
arXiv preprint arXiv:2503.01747, 2025.
One framework to rule them all: Unifying rl-based and rl-free methods in rlhf.
arXiv preprint arXiv:2503.19523, 2025.
Baseline methods for active learning.
In Active Learning and Experimental Design workshop In conjunction with AISTATS 2010, pp.Ā 47ā57. JMLR Workshop and Conference Proceedings, 2011.
On over-fitting in model selection and subsequent selection bias in performance evaluation.
The Journal of Machine Learning Research, 11:2079ā2107, 2010.
Measuring the reliability of reinforcement learning algorithms.
arXiv preprint arXiv:1912.05663, 2019.
Vinci. r1-v: Reinforcing super generalization ability in vision-language models with less than 3 dollars.
How many random seeds? statistical power analysis in deep reinforcement learning experiments.
arXiv preprint arXiv:1806.08295, 2018.
Process reinforcement through implicit rewards.
arXiv preprint arXiv:2502.01456, 2025.
Reinforcement learning for reasoning in small llms: What works and what doesnāt, 2025.
URL [https://arxiv.org/abs/2503.16219](https://arxiv.org/abs/2503.16219 āā).
Assessing diversity collapse in reasoning.
In Scaling Self-Improving Foundation Models without Human Supervision.
Gemini 2.5: Our most intelligent ai model, 2025.
URL [https://blog.google/technology/google-deepmind/gemini-model-thinking-updates-march-2025/](https://blog.google/technology/google-deepmind/gemini-model-thinking-updates-march-2025/ āā).
Accessed: 2025-04-07.
DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning, 2025.
URL [https://arxiv.org/abs/2501.12948](https://arxiv.org/abs/2501.12948 āā).
The benchmark lottery.
arXiv preprint arXiv:2107.07002, 2021.
Openvlthinker: An early exploration to complex vision-language reasoning via iterative self-improvement.
arXiv preprint arXiv:2503.17352, 2025.
Training on the test task confounds evaluation and emergence.
arXiv preprint arXiv:2407.07890, 2024.
Open r1: A fully open reproduction of deepseek-r1, January 2025.
URL [https://github.com/huggingface/open-r1](https://github.com/huggingface/open-r1 āā).
Video-r1: Reinforcing video reasoning in mllms.
arXiv preprint arXiv:2503.21776, 2025.
LightEval: A lightweight framework for LLM evaluation, 2023.
URL [https://github.com/huggingface/lighteval](https://github.com/huggingface/lighteval āā).
On designing effective rl reward at training time for llm reasoning.
arXiv preprint arXiv:2410.15115, 2024a.
A framework for few-shot language model evaluation, 07 2024b.
URL [https://zenodo.org/records/12608602](https://zenodo.org/records/12608602 āā).
Onebench to test them all: Sample-level benchmarking over open-ended capabilities.
arXiv preprint arXiv:2412.06745, 2024.
Time travel in llms: Tracing data contamination in large language models.
arXiv preprint arXiv:2308.08493, 2023.
Towards a standardised performance evaluation protocol for cooperative marl.
Advances in Neural Information Processing Systems, 35:5510ā5521, 2022.
Guha etĀ al. (2024)ā Etash Guha, Negin Raoof, Jean Mercat, Ryan Marten, Eric Frankel, Sedrick Keh, Sachin Grover, George Smyrnis, Trung Vu, Jon Saad-Falcon, Caroline Choi, Kushal Arora, Mike Merrill, Yichuan Deng, Ashima Suvarna, Hritik Bansal, Marianna Nezhurina, Yejin Choi, Reinhard Heckel, Seewong Oh, Tatsunori Hashimoto, Jenia Jitsev, Vaishaal Shankar, Alex Dimakis, Mahesh Sathiamoorthy, and Ludwig Schmidt, November 2024.
He etĀ al. (2024)ā Chaoqun He, Renjie Luo, Yuzhuo Bai, Shengding Hu, ZhenĀ Leng Thai, Junhao Shen, Jinyi Hu, XuĀ Han, Yujie Huang, Yuxiang Zhang, Jie Liu, Lei Qi, Zhiyuan Liu, and Maosong Sun.
Olympiadbench: A challenging benchmark for promoting agi with olympiad-level bilingual multimodal scientific problems, 2024.
URL [https://arxiv.org/abs/2402.14008](https://arxiv.org/abs/2402.14008 āā).
Deep reinforcement learning that matters.
In Proceedings of the AAAI conference on artificial intelligence, volumeĀ 32, 2018.
Measuring mathematical problem solving with the math dataset.
arXiv preprint arXiv:2103.03874, 2021.
Reinforce++: A simple and efficient approach for aligning large language models.
arXiv preprint arXiv:2501.03262, 2025.
Open-reasoner-zero: An open source approach to scaling reinforcement learning on the base model.
[https://github.com/Open-Reasoner-Zero/Open-Reasoner-Zero](https://github.com/Open-Reasoner-Zero/Open-Reasoner-Zero āā), 2025.
Vision-r1: Incentivizing reasoning capability in multimodal large language models.
arXiv preprint arXiv:2503.06749, 2025.
Math-500 dataset.
[https://huggingface.co/datasets/HuggingFaceH4/MATH-500/blob/main/README.md](https://huggingface.co/datasets/HuggingFaceH4/MATH-500/blob/main/README.md āā), 2024.
Accessed: 2025-03-29.
Evaluation gaps in machine learning practice.
In Proceedings of the 2022 ACM conference on fairness, accountability, and transparency, pp.Ā 1859ā1876, 2022.
II-Thought : A Large-Scale, High-Quality Reasoning Dataset, 2025.
Openai o1 system card.
arXiv preprint arXiv:2412.16720, 2024.
Rlsf: Reinforcement learning via symbolic feedback.
arXiv preprint arXiv:2405.16661, 2024.
Evaluating the performance of reinforcement learning algorithms.
In International Conference on Machine Learning, pp.Ā 4962ā4973. PMLR, 2020.
Position: Benchmarking is limited in reinforcement learning research.
arXiv preprint arXiv:2406.16241, 2024.
Large language models struggle to learn long-tail knowledge.
In International Conference on Machine Learning, pp.Ā 15696ā15707. PMLR, 2023.
Vineppo: Unlocking rl potential for llm reasoning through refined credit assignment.
arXiv preprint arXiv:2410.01679, 2024.
Amc-23 dataset, 2025.
URL [https://huggingface.co/datasets/knoveleng/AMC-23](https://huggingface.co/datasets/knoveleng/AMC-23 āā).
Efficient memory management for large language model serving with pagedattention.
In Proceedings of the ACM SIGOPS 29th Symposium on Operating Systems Principles, 2023.
Solving quantitative reasoning problems with language models.
In S.Ā Koyejo, S.Ā Mohamed, A.Ā Agarwal, D.Ā Belgrave, K.Ā Cho, and A.Ā Oh (eds.), Advances in Neural Information Processing Systems, volumeĀ 35, pp.Ā 3843ā3857. Curran Associates, Inc., 2022.
URL [https://proceedings.neurips.cc/paper_files/paper/2022/file/18abbeef8cfe9203fdf9053c9c4fe191-Paper-Conference.pdf](https://proceedings.neurips.cc/paper_files/paper/2022/file/18abbeef8cfe9203fdf9053c9c4fe191-Paper-Conference.pdf āā).
LIMR: Less is More for RL Scaling.
arXiv preprint arXiv:2502.11886, 2025a.
From system 1 to system 2: A survey of reasoning large language models.
arXiv preprint arXiv:2502.17419, 2025b.
Are we learning yet? A meta review of evaluation failures across machine learning.
In Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2), 2021.
Letās verify step by step.
In The Twelfth International Conference on Learning Representations, 2023.
Aime 2025 dataset, 2025.
URL [https://huggingface.co/datasets/yentinglin/aime_2025](https://huggingface.co/datasets/yentinglin/aime_2025 āā).
Accessed: 2025-03-29.
Cppo: Accelerating the training of group relative policy optimization-based reasoning models.
arXiv preprint arXiv:2503.22342, 2025a.
Mind with eyes: from language reasoning to multimodal reasoning.
arXiv preprint arXiv:2503.18071, 2025b.
Troubling trends in machine learning scholarship: Some ml papers suffer from flaws that could mislead the public and stymie future research.
Queue, 17(1):45ā77, 2019.
Code-r1: Reproducing r1 for code with reliable rewards.
2025.
Are your llms capable of stable reasoning?
arXiv preprint arXiv:2412.13147, 2024.
There may not be aha moment in r1-zero-like training ā a pilot study.
[https://oatllm.notion.site/oat-zero](https://oatllm.notion.site/oat-zero āā), 2025a.
Notion Blog.
Understanding r1-zero-like training: A critical perspective, 2025b.
URL [https://arxiv.org/abs/2503.20783](https://arxiv.org/abs/2503.20783 āā).
Visual-rft: Visual reinforcement fine-tuning.
arXiv preprint arXiv:2503.01785, 2025c.
DeepScaleR: Surpassing O1-Preview with a 1.5B Model by Scaling RL, 2025.
Notion Blog.
Exploring the limit of outcome reward for learning mathematical reasoning.
arXiv preprint arXiv:2502.06781, 2025.
Rethinking rl scaling for vision language models: A transparent, from-scratch framework and comprehensive evaluation scheme.
arXiv preprint arXiv:2504.02587, 2025.
Eureka: Human-level reward design via coding large language models.
arXiv preprint arXiv:2310.12931, 2023.
Revisiting the arcade learning environment: Evaluation protocols and open problems for general agents.
Journal of Artificial Intelligence Research, 61:523ā562, 2018.
Quantifying variance in evaluation benchmarks, 2024.
URL https://arxiv. org/abs/2406.10229.
Scientific credibility of machine translation research: A meta-evaluation of 769 papers.
arXiv preprint arXiv:2106.15195, 2021.
Mm-eureka: Exploring visual aha moment with rule-based large-scale reinforcement learning.
arXiv preprint arXiv:2503.07365, 2025.
The llama 4 herd: The beginning of a new era of natively multimodal ai innovation, 2025.
URL [https://ai.meta.com/blog/llama-4-multimodal-intelligence/](https://ai.meta.com/blog/llama-4-multimodal-intelligence/ āā).
Accessed: 2025-04-07.
Imitate, explore, and self-improve: A reproduction report on slow-thinking reasoning systems, 2024.
URL [https://arxiv.org/abs/2412.09413](https://arxiv.org/abs/2412.09413 āā).
Gsm-symbolic: Understanding the limitations of mathematical reasoning in large language models.
arXiv preprint arXiv:2410.05229, 2024.
s1: Simple test-time scaling, 2025.
URL [https://arxiv.org/abs/2501.19393](https://arxiv.org/abs/2501.19393 āā).
A metric learning reality check.
In Computer VisionāECCV 2020: 16th European Conference, Glasgow, UK, August 23ā28, 2020, Proceedings, Part XXV 16, pp.Ā 681ā699. Springer, 2020.
Alice in wonderland: Simple tasks showing complete reasoning breakdown in state-of-the-art large language models.
arXiv preprint arXiv:2406.02061, 2024.
OpenAI o3-mini System Card, January 2025.
URL [https://cdn.openai.com/o3-mini-system-card-feb10.pdf](https://cdn.openai.com/o3-mini-system-card-feb10.pdf āā).
The neglected tails in vision-language models.
In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.Ā 12988ā12997, 2024.
Empirical design in reinforcement learning.
Journal of Machine Learning Research, 25(318):1ā63, 2024.
Lmm-r1: Empowering 3b lmms with strong reasoning abilities through two-stage rule-based rl.
arXiv preprint arXiv:2503.07536, 2025.
Proof or bluff? evaluating llms on 2025 usa math olympiad.
arXiv preprint arXiv:2503.21934, 2025.
Gdumb: A simple approach that questions our progress in continual learning.
In Computer VisionāECCV 2020: 16th European Conference, Glasgow, UK, August 23ā28, 2020, Proceedings, Part II 16, pp.Ā 524ā540. Springer, 2020.
Randumb: A simple approach that questions the efficacy of continual representation learning.
arXiv e-prints, pp.Ā arXivā2402, 2024a.
Efficient lifelong model evaluation in an era of rapid progress.
arXiv preprint arXiv:2402.19472, 2024b.
Rdumb: A simple approach that questions our progress in continual test-time adaptation.
Advances in Neural Information Processing Systems, 36:39915ā39935, 2023.
Reproducibility ā pytorch documentation.
[https://pytorch.org/docs/stable/notes/randomness.html](https://pytorch.org/docs/stable/notes/randomness.html āā), 2024.
Accessed: 2025-04-09.
The effect of sampling temperature on problem solving in large language models.
In Findings of the Association for Computational Linguistics: EMNLP 2024, pp.Ā 7346ā7356, 2024.
BetterBench: Assessing AI Benchmarks, Uncovering Issues, and Establishing Best Practices.
arXiv preprint arXiv:2411.12990, 2024.
Zerobench: An impossible visual benchmark for contemporary large multimodal models.
arXiv preprint arXiv:2502.09696, 2025.
To the cutoff⦠and beyond? a longitudinal perspective on llm data contamination.
In The Twelfth International Conference on Learning Representations, 2023.
Tapered off-policy reinforce: Stable and efficient reinforcement learning for llms.
arXiv preprint arXiv:2503.14286, 2025.
Deepseekmath: Pushing the limits of mathematical reasoning in open language models.
arXiv preprint arXiv:2402.03300, 2024.
Exploring data scaling trends and effects in reinforcement learning from human feedback.
arXiv preprint arXiv:2503.22230, 2025.
Critique of impure reason: Unveiling the reasoning behaviour of medical large language models.
arXiv preprint arXiv:2412.15748, 2024.
FastCuRL: Curriculum Reinforcement Learning with Progressive Context Extension for Efficient Training R1-like Reasoning Models, 2025.
URL [https://arxiv.org/abs/2503.17287](https://arxiv.org/abs/2503.17287 āā).
Functional benchmarks for robust evaluation of reasoning performance, and the reasoning gap.
arXiv preprint arXiv:2402.19450, 2024.
Expanding rl with verifiable rewards across diverse domains.
arXiv preprint arXiv:2503.23829, 2025.
Open Thoughts.
https://open-thoughts.ai, January 2025.
Kimi k1. 5: Scaling reinforcement learning with llms.
arXiv preprint arXiv:2501.12599, 2025.
Enhancing llm reasoning with iterative dpo: A comprehensive empirical investigation.
arXiv preprint arXiv:2503.12854, 2025.
Noā zero-shotā without exponential data: Pretraining concept frequency determines multimodal model performance.
In The Thirty-eighth Annual Conference on Neural Information Processing Systems, 2024.
Solving math word problems with process-and outcome-based feedback.
arXiv preprint arXiv:2211.14275, 2022.
Inference reproducibility script.
[https://github.com/vllm-project/vllm/blob/098900d7c2b53324687977eece400f634755cf51/examples/offline_inference/reproduciblity.py](https://github.com/vllm-project/vllm/blob/098900d7c2b53324687977eece400f634755cf51/examples/offline_inference/reproduciblity.py āā), 2024.
Accessed: 2025-04-09.
Thoughts are all over the place: On the underthinking of o1-like llms.
arXiv preprint arXiv:2501.18585, 2025.
Light-r1: Curriculum sft, dpo and rl for long cot from scratch and beyond, 2025a.
URL [https://arxiv.org/abs/2503.10460](https://arxiv.org/abs/2503.10460 āā).
Light-r1: Curriculum sft, dpo and rl for long cot from scratch and beyond.
arXiv preprint arXiv:2503.10460, 2025b.
Grok 3 beta ā the age of reasoning agents.
February 2025.
URL [https://x.ai/news/grok-3](https://x.ai/news/grok-3 āā).
Accessed: 2025-03-29.
Logic-rl: Unleashing llm reasoning with rule-based reinforcement learning.
arXiv preprint arXiv:2502.14768, 2025.
Recitation over reasoning: How cutting-edge language models can fail on elementary school-level reasoning problems?
arXiv preprint arXiv:2504.00509, 2025.
Qwen2. 5 technical report.
arXiv preprint arXiv:2412.15115, 2024a.
Qwen2.5-math technical report: Toward mathematical expert model via self-improvement.
arXiv preprint arXiv:2409.12122, 2024b.
Codepmp: Scalable preference model pretraining for large language model reasoning.
arXiv preprint arXiv:2410.02229, 2024.
Dapo: An open-source llm reinforcement learning system at scale.
arXiv preprint arXiv:2503.14476, 2025.
Naturalreasoning: Reasoning in the wild with 2.8 m challenging questions.
arXiv preprint arXiv:2502.13124, 2025.
Vapo: Efficient and reliable reinforcement learning for advanced reasoning tasks.
arXiv preprint arXiv:2504.05118, 2025.
Versaprm: Multi-domain process reward model via synthetic reasoning data.
arXiv preprint arXiv:2502.06737, 2025a.
SimpleRL-Zoo: Investigating and Taming Zero Reinforcement Learning for Open Base Models in the Wild.
arXiv preprint arXiv:2503.18892, 2025b.
Med-rlvr: Emerging medical reasoning from a 3b base model via reinforcement learning.
arXiv preprint arXiv:2502.19655, 2025.
Report issue for preceding element
Report issue for preceding element
To complement our analysis in 3, we present bootstrapped variance results on two additional datasets: AMCā23 and MATH500. As shown in FiguresĀ 12 andĀ 13, high variance in Pass@1 persists even when averaging over multiple seeds (K=5š¾5K=5italic_K = 5), mirroring the trends observed on AIMEā24. These results reinforce our conclusion that small benchmark sizes yield unstable estimates and that robust performance reporting requires multiple seed runs.
Report issue for preceding element
Figure 12: Variance of mean Pass@1 on AMCā23. Bootstrapped estimates show substantial variance even with K=5š¾5K=5italic_K = 5 evaluation runs, highlighting the instability of single-seed evaluations.Report issue for preceding element
Figure 13: Variance of mean Pass@1 on MATH500. Similar to AIMEā24 and AMCā23, the estimates remain volatile across seeds. Even K=5š¾5K=5italic_K = 5 runs do not eliminate variance, underscoring the need for larger Kš¾Kitalic_K.Report issue for preceding element
Report issue for preceding element
In FigureĀ 14, we show that the model performance variation due to hardware configuration is not limited to AIMEā24 and AMCā23. Similar discrepancies are observed on MATH500, where different compute clusters yield different accuracy scoresāeven when model, seeds, and decoding parameters are held constant. This further emphasizes the need for hardware and software standardization when reporting benchmark results.
Report issue for preceding element
Figure 14: Performance variation across compute clusters on MATH500. Differences in GPU type and environment lead to non-trivial shifts in performance, reinforcing the importance of hardware standardization.Report issue for preceding element
Report issue for preceding element
We provide the exact templates used for our three prompt settings in TableĀ 5: Math, Default, and No Template. These formats are based on the DeepSeek tokenizer but adapted for each modelās specific chat template. Our results (in 3.4) indicate that instruction-tuned models are highly sensitive to prompt formatting, with performance degrading significantly when prompts deviate from their training-time structure.
Report issue for preceding element
Prompt | Example |
---|---|
Math | <|begin_of_sentence|><|User|>Solve the following math problem efficiently and clearly. The last line of your response should be of the following format: āTherefore, the final answer is: $\boxed{ANSWER}$. I hope it is correctā (without quotes) where ANSWER is just the final number or expression that solves the problem. Think step by step before answering.\n <|Assistant|> |
Default | <|begin_of_sentence|><|User|>{Question} <|Assistant|> |
No Template | {Question} |
Table 5: Prompt templates used in our evaluation. The inclusion or exclusion of structured prompt tokens significantly impacts performance for instruction-tuned models.Report issue for preceding element
Report issue for preceding element
We further explore how varying max_new_tokens impacts model accuracy. Figures below compare OpenRS-series models (with 131,072-token context windows) and OpenThinker/S1.1 models (with 32,768-token limits).
Report issue for preceding element
FigureĀ 15 shows that OpenRS models are highly sensitive to this parameterāshortening outputs results in clear accuracy drops. Similarly, FigureĀ 16 reveals the same pattern for OpenThinker-7B and S1.1-7B, despite their smaller context lengths. In both cases, premature truncation leads to incomplete reasoning chains and incorrect answers, confirming the importance of setting appropriate generation limits.
Report issue for preceding element
Figure 15: Impact of max_new_tokens on OpenRS models. Models with long context support (131,072 tokens) experience degraded performance when max_new_tokens is set too low.Report issue for preceding element
Figure 16: Impact of max_new_tokens on OpenThinker and S1.1 models. Despite shorter context limits (32,768 tokens), performance still degrades noticeably when output length is constrained.Report issue for preceding element
Report issue for preceding element
To supplement the aggregated results shown in FigureĀ 11, we include detailed histograms for each individual model in the appendix. These plots show the distribution of correct and incorrect responses across response lengths, averaged over random seeds. Due to the number of models analyzed, we split the results into two figures for clarity.
Report issue for preceding element
FiguresĀ 17 andĀ 18 reveal that the overall trend observed in the main paper holds consistently across nearly all models: incorrect responses tend to be longer than correct ones.
Report issue for preceding element
These results reinforce the idea that excessively long outputs often indicate failure modes such as hallucinated reasoning, verbose overthinking, or degenerate loops. Importantly, this correlation persists well below the maximum sequence length, ruling out truncation as the sole cause.
Report issue for preceding element
Figure 17: Response Length vs. Correctness ā Models (1/2). Average number of correct and incorrect responses across response length bins for a subset of models. Longer responses consistently correlate with incorrect predictions.Report issue for preceding element
Figure 18: Response Length vs. Correctness ā Models (2/2). Continuation of model-wise response length analysis. The same trend holds across the remaining models, with incorrect answers being disproportionately long.Report issue for preceding element
Across all models, longer responses are a consistent marker of incorrect outputs, making response length a useful signal for detecting low-confidence or erroneous reasoning chains.
Report issue for preceding element
Report IssueReport Issue for Selection
Generated by
L
A
T
Exml