Extracted Text


2504.13837v4.pdf

October 24, 2025
Does Reinforcement Learning Really Incentivize Reasoning
Capacity in LLMs Beyond the Base Model?
Yang Yue
1
,Zhiqi Chen
1 ,Rui Lu
1 ,Andrew Zhao
1 ,Zhaokai Wang
2 ,Yang Yue
1 ,
Shiji Song
1
, and Gao Huang
1
1
LeapLab, Tsinghua University
2
Shanghai Jiao Tong University

Equal Contribution

Project Lead
B
Corresponding Author
Reinforcement Learning with Verifiable Rewards (RLVR) has recently demonstrated notable
success in enhancing the reasoning performance of large language models (LLMs), particularly
in mathematics and programming tasks. It is widely believed that, similar to how traditional
RL helps agents to explore and learn new strategies, RLVR enables LLMs to continuously self-
improve, thus acquiring novel reasoning abilities that exceed the capacity of the corresponding
base models. In this study, we take a critical look at
probing the reasoning capability boundaries of RLVR-trained LLMs across various model families,
RL algorithms, and math/coding/visual reasoning benchmarks, using pass@k
as the evaluation metric. While RLVR improves sampling efficiency towards correct paths, we
surprisingly find that current training does
observe that while RLVR-trained models outperform their base models at smaller values ofk
(e.g.,k=1), base models achieve higher pass@kscore whenkis large. Moreover, we observe that
the reasoning capability boundary of LLMs often narrows as RLVR training progresses. Further
coverage and perplexity analysis shows that the reasoning paths generated by RLVR models
are already included in the base models’ sampling distribution, suggesting that their reasoning
abilities originate from and are
the base model as an upper bound, our quantitative analysis shows that six popular RLVR
algorithms perform similarly and remain far from optimal in fully leveraging the potential of the
base model. In contrast, we find that distillation can introduce new reasoning patterns from the
teacher and genuinely expand the model’s reasoning capabilities. Taken together, our findings
suggest that current RLVR methods have not fully realized the potential of RL to elicit genuinely
novel reasoning abilities in LLMs. This underscores the need for improved RL paradigms–such
as continual scaling and multi-turn agent-environment interaction–to unlock this potential.
Project Page
1. Introduction
The development of reasoning-centric large language models (LLMs), such as OpenAI-o1 (Jaech et al.,
2024), DeepSeek-R1 (Guo et al.,), and Kimi-1.5 (Team et al.,), has significantly advanced
the frontier of LLM capabilities, particularly in solving complex logical tasks involving mathematics
and programming. In contrast to traditional instruction-tuned approaches that rely on human-curated
annotations (Achiam et al.,;,), the key driver behind this leap forward is
The first author 乐) 乐)
same English name but different Chinese names.
huang@tsinghua.edu.cn.
arXiv:2504.13837v4 [cs.AI] 23 Oct 2025

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?More
Efficient Sampling
0
0
000+1
0
+1
Base Model
0+1
0
+1
RLVR Model
0
0
00
Reduced Scope of
Reasoning Capacity
+1
0
00
Problem A
Problem B
+1
0
00 1248163264128256
Number of Samples k
0.1
0.2
0.3
0.4
0.5
0.6
0.7
Coverage (pass@k)
Omni-MATH-Train
Qwen2.5-7B
GRPO-step150
GRPO-step300
GRPO-step450
Figure 1:
from the base and RLVR-trained models for a given problem.
the model, while
rewards. Our key finding is that all reasoning paths in the RLVR model are already present in the base model. For certain
problems like Problem A, RLVR training biases the distribution toward rewarded paths, improving sampling efficiency.
However, this comes at the cost of reduced scope of reasoning capacity: For other problems like Problem B, the base
model contains the correct path, whereas that of the RLVR model does not.
average performance (i.e., pass@1) improves, but the coverage of solvable problems (i.e., pass@256) decreases, indicating a
reduction in LLM’s reasoning boundary.
large-scaleLambert et al.,;,
2025). RLVR starts with a pretrained base model or one fine-tuned on long chains of thought (CoT)
data, optimizing it via reinforcement learning based on simple, automatically computable rewards. These
rewards are determined by whether the model’s output matches a ground-truth solution in mathematics
or passes unit tests in code, thus enabling scalability without human labeling. This framework has
gained significant attention due to its simplicity and practical effectiveness. In traditional RL settings
such as game playing (e.g., Atari, Go), agents often autonomously discover new strategies and surpass
even human-level performance through self-improvement (Mnih et al.,;,). Inspired
by this success, it is widely believed that RLVR similarly enables LLMs to autonomously develop
novel reasoning patterns, including enumeration, self-reflection, and iterative refinement, surpassing
the capabilities of their base models (Guo et al.,). Consequently, RLVR has been considered a
promising path toward continuously self-evolving LLMs, potentially bringing us closer to more powerful
intelligence (,).
However, despite its empirical success, the underlying effectiveness of current RLVR remains underexam-
ined. This raises a fundamental question:
novel reasoning abilities–similar to how traditional RL discovers new strategies through
exploration–or does it simply utilize reasoning patterns already in the base model?
To rigorously answer this question, we must first assess the reasoning capability boundaries of both base
and RLVR-trained models. Traditional evaluation metrics rely on average score from greedy decoding
or nucleus sampling (Holtzman et al.,), which reflects average-case behavior. However, these
metrics risk underestimating the true potential of a model, especially if it fails on difficult problems
after limited attempts, despite being capable of solving them with more sampling. To overcome this
limitation, we adopt the pass@kBrown et al.,), where a problem is considered solved
if any of theksampled outputs is correct. By allowing multiple attempts, pass@k
a model has the potential to solve a problem. The average pass@k
proportion of problems a model can potentially solve withinktrials, offering a more robust view of its
reasoning boundary. This provides a rigorous test on whether the RLVR training yields fundamentally
transcending capacity, enabling the model to solve problems that the base model cannot.
Using the pass@k
LLM families, model sizes, and RLVR algorithms to compare base models with their RLVR-trained
counterparts. We uncover several surprising findings that offer a more comprehensive assessment of the
2

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
effectiveness of current RLVR training and reveal the gap between existing RLVR methods and the ideal
goals of RL-discovering genuinely new reasoning strategies:

Current RLVR models exhibit narrower reasoning coverage than their base models.
pass@kcurves, although RLVR models outperform their base models at smallk, it is surprising that
base models consistently surpass RLVR models across all benchmarks and LLM families askincreases.
This suggests that current RLVR training does
solvable problems. Manual inspection of model responses shows that, for most problems, the base model
can produce
problems that were previously considered only solvable for RLVR models.

Reasoning paths generated by current RLVR model already exist in its base model.
To further investigate this phenomenon, we analyze the accuracy distribution. The results show that
although RLVR improves average performance (i.e., pass@1) by sampling more efficiently on problems
already solvable by the base model, it does not enable the model to solve new problems. Further
perplexity analysis reveals that the reasoning paths produced by RLVR models already exist within
the output distribution of the base model. These findings indicate that RLVR does not introduce
fundamentally new reasoning capabilities and that the reasoning capacity of current RLVR models
remains bounded by that of its base model. This effect of RLVR is illustrated in Figure

Current RLVR algorithms perform similarly and remain far from optimal.
base model as an upper bound, we define the∆ SE), shown in Figure
the difference between an RL model’s pass@1 and the base model’s pass@k(withk= 256
upper-bound performance). This metric quantifies how closely an RL algorithm approaches the optimal
bound. Across all algorithms (e.g., PPO, GRPO, Reinforce++), SEshows only minor variation yet
remains consistently large, suggesting that current RLVR methods, while improving sampling efficiency,
are still far from optimal.

RLVR and distillation are fundamentally different.
more efficiently sampling high-reward outputs, it does not elicit new reasoning capabilities and remains
constrained within the base model’s capacity. In contrast, distillation can transfer new reasoning patterns
from a stronger teacher to the student. As a result, distilled models often demonstrate an expanded
reasoning scope beyond that of the base model.
In conclusion, our findings show that current RLVR methods, while improving sampling efficiency, do not
elicit novel reasoning beyond the base model’s capabilities. This highlights a gap between existing RLVR
methods and the goals of reinforcement learning, underscoring the need for improved RL paradigms
such as continual scaling, better exploration, and multi-turn agent interaction.
2. Preliminaries
In this section, we first outline the fundamentals of RLVR, then introduce the pass@kmetric to evaluate
reasoning boundaries, and explain why it is preferred over alternatives like best-of-N
2.1. Reinforcement Learning with Verifiable Rewards
Verifiable Rewards. πθbe an LLM with parametersθthat generates a token sequence
(y1, . . . , yT ) x. A deterministic Vreturns a binary
reward:r=V(x,y∈ {0,1},wherer= 1
A format reward may also be added to encourage the model to explicitly separate the reasoning
process from the final answer. The goal of RL is to learn a policy to maximize the expected reward:
Jθ x∼D
ˆ
E
yπ θ(·|x)[
˜
, where
RLVR Algorithms.Schulman et al.,) proposed using the
following clipped surrogate to maximize the objective:
LCLIP=r t(θA t,r t(θ,)A t)]
wherert(θ) =πθ(yt|x,y<t)/πθold(yt|x,y<t), andAtis the advantage estimated by a value networkVϕ.
KL divergence term is optionally applied, to constrain the model from deviating too far from the original
3

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?1416642561024
Number of Samples k
0
0.2
0.4
0.6
0.8
1
Coverage (pass@k)
Base
RL
1416642561024
0
0.2
0.4
0.6
0.8
1
1416642561024
0
0.2
0.4
0.6
0.8
1
1416642561024
0
0.2
0.4
0.6
0.8
1
1248163264128
0
0.2
0.4
0.6
0.8
1
1248163264128
0
0.2
0.4
0.6
0.8
1
1248163264128
0
0.2
0.4
0.6
0.8
1
1248163264128
0
0.2
0.4
0.6
0.8
1
1248163264128
0
0.2
0.4
0.6
0.8
1248163264128
0
0.2
0.4
0.6
0.8
1248163264128
0
0.2
0.4
0.6
0.8
1248163264128
0
0.2
0.4
0.6
0.8
1248163264128
0
0.2
0.4
0.6
0.8
1248163264128
0
0.2
0.4
0.6
0.8
1248163264128
0
0.2
0.4
0.6
0.8
1416642561024
0
0.2
0.4
0.6
0.8
AIME24
MATH500
Minerva
Olympiad
Qwen-2.5-7B Qwen-2.5-14B Qwen-2.5-32B LLaMA-3.1-8B
Figure 2: Pass@kcurves of base models and their RLVR-trained counterparts across multiple mathematical benchmarks.
Whenkis small, RL-trained models outperform their base versions. However, askincreases to the tens or hundreds,
base models consistently catch up and surpass RL-trained models. More results on GSM8K and AMC23 can be found
at Figure.
policy. More algorithms are introduced in Section.
Policy Gradient.Williams,;
Sutton et al.,). These methods learn exclusively from,.e., samples generated
by the current LLM. In the context of verifiable rewards, the training objective generally
the log-likelihood of samples with correct answers and minimizes the likelihood of those with incorrect
answers.
Zero RL TrainingGuo
et al.,). To clearly study the effect of RLVR, we follow this zero-RL setting for all math tasks
using pretrained models as start model. However, for coding and visual reasoning tasks, open-source
work typically uses instruction-tuned models as starting points, primarily due to the training instability
and limited effectiveness of using a pure zero-RL setting. Following this convention, we compare the
finetuned model with its RLVR-trained counterpart to focus solely on the effect of RLVR.
4

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
2.2. Metrics for LLM Reasoning Capacity Boundary
Pass@kMetrics.
challenging, as methods like greedy decoding or the average of nucleus samplings (Holtzman et al.,)
only reflect average-case performance. To accurately measure the reasoning ability boundary, we extend
the commonly used pass@kChen et al.,) to all tasks with verifiable
rewards. Given a problem, we samplekoutputs from the model. The pass@kvalue for this question is 1
if at least one of theksamples passes verification; otherwise, it is 0. The average pass@kvalue over
the dataset reflects the proportion of problems in the dataset that the model can solve withinktrials,
providing a rigorous evaluation of the reasoning capacity coverage of LLMs. We adopt an unbiased,
low-variance estimator for computing to calculate pass@k.
Comparison with Best-of-NNCobbe et al.,) and majority
voting are practical methods for selecting correct answers, but they may overlook a model’s full reasoning
potential. In contrast, we use pass@knot to assess practical utility
reasoning capacity. If a model produces a correct solution in any of theksamples, we treat the problem
as within its potential scope. Thus, if RL enhances reasoning, the RL-trained model should succeed in
more such problems than the base model. Methods like Best-of-Nor majority voting may miss these
successes if the correct answer is not selected by the verifier or voting.
Random Guessing Issue. For
used as verifiers, the pass@kvalue can accurately reflect whether the model can solve the problem. In
mathematics, the issue of “guessing” can become pronounced as kincreases, where a model may generate
an incorrect CoT but still accidentally arrive at the correct answer. To address this, we manually check
the correctness of CoT for a subset of model outputs as detailed in Section. By combining reuslts on
math with manually checking and coding, we rigorously evaluate the scope of LLM’s reasoning capacity.
Another caveat is that, with an astronomically largek, even uniform sampling over the token dictionary
would stumble upon the correct reasoning path–though this is infeasible within today’s time and compute
budgets. Crucially, we find that the base model already produces correct outputs at realistic values
(k), well within practical resource limits.
Table 1: Experimental setup for assessing RLVR’s effect on the reasoning boundaries of LLMs.
Task Start Model RL Framework RL Algorithm(s) Benchmark(s)
Mathematics
LLaMA-3.1-8B
Qwen2.5-7B/14B/32B-Base
Qwen2.5-Math-7B
SimpleRLZoo
Oat-Zero
DAPO
GRPO
GSM8K, MATH500
Minerva, Olympiad
AIME24, AMC23
Code Generation
Qwen2.5-7B-Instruct
DeepSeek-R1-Distill-Qwen-14B
Code-R1
DeepCoder
GRPO
LiveCodeBench
HumanEval+
Visual Reasoning
MathVista
MathVision
Deep Analysis
Qwen2.5-7B-Base
Qwen2.5-7B-Instruct
DeepSeek-R1-Distill-Qwen-7B
VeRL
PPO, GRPO
Reinforce++
RLOO, ReMax, DAPO
Omni-Math-Rule
MATH500
3. RLVR’s Effect on Reasoning Capacity Boundary
With the evaluation metrics for reasoning boundaries established, we now conduct a comprehensive
evaluation of the base and RLVR models through extensive experiments. Our analysis is organized
by task category, covering three representative domains: mathematics, code generation, and visual
reasoning. The overall experimental setup is summarized in Table.
Evaluation Protocol.
of 0.6 and a top-pvalue of 0.95, allowing a maximum generation of 16,384 tokens. We also show the
effect of different temperature settings in Figure. For evaluation of the base model, a common practice
is to include few-shot examples in the prompt to guide the output (Grattafiori et al.,;,
2024;,). However, to ensure a fair and unbiased comparison, we deliberately avoid using
few-shot prompts for base models, eliminating any potential confounding effects on reasoning that might
5

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
be introduced by in-context examples. For evaluating both the base and RLVR models, we use the same
zero-shot prompt as in RLVR training, or the default prompt provided by the benchmark, ensuring a
consistent setup across both models. Interestingly, although base models often produce unformatted or
non-sensical responses without few-shot guidance, we observe that with sufficient sampling, they are still
capable of generating correctly formatted outputs and successfully solving complex problems. Prompt
templates for training and evaluation are provided in Section.
3.1. RLVR for Mathematical Reasoning
Models and Benchmarks.i.e.,
CoT) along with the final answer. To ensure the robustness of conclusions, we experiment with multiple
LLM families, primarily Qwen2.5 (7B/14B/32B base variants) (Yang et al.,) and additionally
LLaMA-3.1-8B (Grattafiori et al.,). We adopt RLVR models released by SimpleRLZoo (Zeng et al.,
2025), which train zero-RL models using GRPO on GSM8K and the MATH training set, with correctness
reward only, excluding any format-based reward. We compare the pass@kcurves of base and zero-RL
models on benchmarks of varying difficulty: GSM8K (Cobbe et al.,), MATH500 (Hendrycks et al.,
2021), Minerva (Lewkowycz et al.), Olympiad (He et al.,), AIME24, and AMC23. Additionally,
we include the RLVR model Oat-Zero-7B and DAPO-32B (Liu et al.,;,). These two
models are characterized by strong performance on the challenging AIME24 benchmark.
The Effect of RLVR: Increased Likelihood of Correct Samples, Decreased Coverage of
Solvable Problems., we consistently observe a contrasting trend between small
and largekvalues. Whenkis small (e.g., k= 1, equivalent to average-case accuracy), RL-trained
models outperform their base counterparts. This aligns with the common observation that RL improves
performance, suggesting that RLVR makes models significantly more likely to sample correct responses.
However, askincreases, with steeper curves, base models consistently catch up to and eventually surpass
RL-trained models across all benchmarks, indicating their broader coverage of solvable problems. For
example, on the Minerva benchmark with a 32B-sized model, the base model outperforms the RL-trained
model by approximately 9% atk= 128, implying that it can solve around 9% more problems in the
validation set.
We further examine RL models trained with Oat-Zero and DAPO. As shown in Figure, although the
RL model initially demonstrates a strong performance, nearly 30% higher than the base model, it is
eventually surpassed by the base model. Based on these results, we conclude that RLVR increases the
likelihood of sampling correct responses at lowk, but narrows the model’s overall coverage. We further
analyze the root cause of this phenomenon in Section.
CoT Case Analysis.
and Figure, manually selected from 2048 samplings for the hardest questions in AIME24. The
responses from the base model tend to be long CoTs and exhibit reflective behavior, highlighting the
strong reasoning ability inherent in the base model.
Validity of Chain-of-Thought.
on the correctness of the final answer, with the risk of “hacking”. To accurately reflect the reasoning
ability boundary using pass@k, it is important to assess how many solved problems result from sampling
genuinely correct CoTs, rather than from lucky guesses. Following (Brown et al.,), we manually
inspect all CoTs that led to correct answers to the most challenging solvable problems in the GSM8k
dataset – those with an average accuracy below 5% but above 0%. The base model answered 25 such
questions, with 24 containing
questions, 23 of which included
in the challenging AIME24 benchmark with an average accuracy below 5%. Details can be found
in Section. The base model answered 7 such questions, with 5 out of 6 containing
correct CoT (excluding one ambiguous case of correctness due to skipped reasoning steps). Similarly,
the RL-trained model answered 6 questions, 4 of which included
suggest that the base model can sample valid reasoning paths to solve the problems.
6

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?1248163264128
Number of Samples k
0.75
0.8
0.85
0.9
0.95
Coverage (pass@k)
HumanEval+
Qwen-2.5-7B-Instruct-1M
Coder-R1-Zero-Qwen-2.5-7B
1248163264128
Number of Samples k
0.7
0.8
0.9
MBPP+ 1248163264128
Number of Samples k
0.9
0.8
0.7
0.6
0.5Coverage (pass@k)
MathVista
Qwen2.5-VL-7b-Instruct
Qwen2.5-VL-7b-Instruct-RL
1248163264128
Number of Samples k
0.2
0.4
0.6
0.8
MathVision
Figure 4: Pass@k
3.2. RLVR for Code Generation1 4 16 64
Number of Samples (k)
0.5
0.55
0.6
0.65
0.7
0.75
0.8
Coverage (pass@k)
LiveCodeBench (2024.8-2025.1)
DeepSeek-R1-Distill-Qwen-14B (Base)
Deepcoder-14B-Preview (RL)
Figure 3: RLVR for Coding.
Models and Benchmarks.
trained model, CodeR1-Zero-Qwen2.5-7B (Liu & Zhang,), which
trains zero-RL models on 12K LeetCode and TACO samples over 832
steps, based on Qwen2.5-7B-Instruct-1M (,
evaluation, models are assessed on LiveCodeBench v5, comprising 279
problems that span from August 2024 to January 2025 (Jain et al.,
2025), as well as HumanEval+ and MBPP+ (Liu et al.,). We
also evaluate the most powerful open-source RLVR-trained coding
LLM, DeepCoder-14B (Luo et al.,), built on DeepSeek-R1-Distill-
Qwen-14B. Here both models take 32k response length. Due to their
high computational cost, we evaluate them only on LiveCodeBench
as a representative benchmark.
The Effect of RLVR.
to achieve by guesswork, pass@kprovides a reliable measure of a model’s reasoning boundary. As shown
in Figure, Figure, and Figure
exhibit trends that are highly consistent with those observed in mathematical benchmarks.
3.3. RLVR for Visual Reasoning
Models and Benchmarks.
inputs to solve complex reasoning problems. This has gained significant attention in the multimodal
community since the rise of LLM reasoning (Chen et al.,;,;,).
For our experiments, we select math within visual contexts as a representative task. We use the EasyR1
framework (Zheng et al.,) to train Qwen2.5-VL-7B (Bai et al.) on Geometry3K (Lu et al.,
2021), and evaluate its visual reasoning capabilities on filtered MathVista-TestMini (Lu et al.,) and
MathVision-TestMini (Wang et al.,), where multiple-choice questions are removed.
The Effect of RLVR.
consistent with those observed in math and coding benchmarks. This suggests that the original model
has broader coverage of solvable questions even in multimodal tasks.
Validity of Chain-of-Thought.
problems,.e.those with an average accuracy below 5%. We find that for both the original and RL
models, 7 out of 8 problems have
4. Deep Analysis
In this section, we conduct a deeper analysis of the effects of current RLVR training. We also highlight
the distinct characteristics of distillation in comparison to RLVR. In addition, we design controlled
experiments to examine the impact of different RL algorithms and design choices.
7

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
4.1. Reasoning Paths Already Present in Base Models0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
Accuracy Interval
0
20
40
60
80
100
120
Frequency
Base
RL
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0.0 0.2 0.4 0.6 0.8 1.0
0.0
0.2
0.4
0.6
0.8
1.0
0
10
20
30
40
Percentage (%)
AIME24
AMC23
MATH500
Minerva
GSM8K
Olympiad
Qwen-2.5-7B Qwen-2.5-14B Qwen-2.5-32B LLaMA-3.1-8B
Figure 5: Qwen2.5-7B Accuracy His-
togram on Minerva.
Accuracy Distribution Analysis.
reveal a surprising trend: the base model covers a wider range
of solvable problems than the RLVR-trained model. To better
understand this, we analyze how the accuracy distribution changes
before and after RLVR training. As shown in Figure, RLVR
increases the frequency of high accuracies near 1.0 and reduces the
frequency of low accuracies (e.g., 0.1, 0.2). However, a deviation
from this trend is the
cating that RLVR leads to more unsolvable problems. This also
explains the improvement of RLVR in average scores, driven not
by solving new problems but by improving sampling efficiency on
problems already solvable by the base model. Additional accuracy
histograms are provided in Figure.
Table 2: We evaluate on AIME24 (k= 1024) and
MATH500 (k= 128). The table reports the solvable/un-
solvable fraction of problems falling into four categories.
Base SimpleRLZoo AIME24 MATH500
✓ ✓ 63.3% 92.4%
✓ ✗ 13.3% 3.6%
✗ ✓ 0.0% 1.0%
✗ ✗ 23.3% 3.0%
Solvable-Problem Coverage Analysis.
investigate, we compare the set of solvable questions
for both the base model and its corresponding RL-
trained version on AIME24 and MATH500. We find
that there are many cases where the base model solves
a problem but the RLVR model fails, and very few
where RLVR succeeds while the base model does not,
as shown in Table. Details can be found at Sec-
tion. As shown in Table, the set of problems
solved by the RL-trained model is nearly a subset of those solvable by the base model. A similar trend
is observed in coding tasks as shown in Table. This raises the natural question: Do all reasoning paths
generated by RL-trained models already exist within the output distribution of their base models?
Perplexity Analysis. To answer this question, we utilize the metric. Given a model m, a
problemx, and a response y1, . . . , yT )
humans), the perplexity is defined as the exponentiated average negative log-likelihood of a sequence:
PPLm(Y) = exp


1
T
T
X
t=1
logy t| 1, . . . , yt−1)
!
,
which reflects the model’s ability to predict the given response x. Lower
perplexity indicates that the model has a higher likelihood of generating this response.PPLBase(YBase)
PPLBase(YRL)
PPLRL(YRL)
PPLBase(YGT)
1.00
1.05
1.10
1.15
1.20
1.25
1.30
Perplexity
PPLBase(YBase)
PPLBase(YRL)
PPLRL(YRL)
PPLBase(YGT)
1.0
1.2
1.4
1.6
1.8
2.0
2.2
Figure 6: Perplexity distribution of responses. The con-
ditioning problem
We randomly sample two problems from AIME24
and employ Qwen2.5-7B-Base and SimpleRL-
Qwen2.5-7B-Base to generate 16 responses for each
problem, denoted as baseand RL, respectively.
We also let OpenAI-o1 (Jaech et al.,) generate
8 responses, denoted as GT. As shown in Figure,
the distribution ofPPLBase (YRL|x)
the lower portion of thePPLBase (YBase|x)
bution, corresponding to responses that the base
model tends to generate. This suggests that the
responses from RL-trained models are highly likely
to be generated by the base model. In Section,
we show thatPPLBase (YRL|x)
as RL training progresses, indicating that RLVR
mainly sharpens the distribution within the base
model’s prior rather than expanding beyond it.
Summary.
rive at three key observations. First, problems solved by the RLVR model are also solvable by the base
8

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
model; the observed improvement in average scores stems from more efficient sampling on these already
solvable problems, rather than learning to solve new problems. Second, after RLVR training, the model
often exhibits narrower reasoning coverage compared to its base model. Third, all the reasoning paths
exploited by the RLVR model are already present in the sampling distribution of the base model. These
findings indicate that RLVR does not introduce fundamentally new reasoning capabilities and that the
reasoning capacity of the trained model remains bounded by that of its base model.
4.2. Distillation Expands the Reasoning Boundary1248163264128
Number of Samples k
0
0.2
0.4
0.6
Coverage (pass@k)
Minerva
Base
RL
Instruct
Distill
Figure 7: pass@kof base, In-
struct, RLVR, and distilled
models.
In addition to direct RL training, another effective approach to improving
the reasoning ability of small base models is distillation from a powerful
reasoning model (Guo et al.,). This process is analogous to instruction-
following fine-tuning in post-training. However, instead of using short
instruction-response pairs, the training data consist of long CoT reasoning
traces generated by the teacher model. Given the limitations of current
RLVR in expanding reasoning capabilities, it is natural to ask whether
distillation exhibits similar behavior. We focus on a representative model,
DeepSeek-R1-Distill-Qwen-7B, which distills DeepSeek-R1 into Qwen2.5-
Math-7B. We compare it with the base model Qwen2.5-Math-7B and its
RL-trained counterpart Qwen2.5-Math-7B-Oat-Zero and include Qwen2.5-
Math-7B-Instruct as an additional baseline. As shown in Figure, the
pass@kcurve of the distilled model is consistently and significantly above
that of the base model. This indicates that, unlike RL that is fundamentally
bounded by the reasoning capacity of the base model, distillation introduces
new reasoning patterns learned from a stronger teacher model. As a result,
the distilled model is capable of surpassing the reasoning boundary of the
base model.
4.3. Effects of Different RL Algorithms
As discussed previously, the primary effect of RL is to enhance sampling efficiency rather than to expand
a model’s reasoning capacity. To quantify this, we propose the∆ SE), defined
as the difference between the RL-trained model’s pass@1 and the base model’s pass@k(we usek= 256
in our evaluation). Lower SEis better. Here we conduct clean experiments to study the effect of
different RL algorithms in enhancing sampling efficiency.
Experiment Setup.
et al.,) for fair comparison, including PPO (Schulman et al.,), GRPO (Shao et al.,),
Reinforce++ (Hu,), RLOO (Ahmadian et al.,), ReMax (Li et al.,), and DAPO (Yu
et al.,). Following DAPO (Yu et al.,) and Oat-Zero (Liu et al.,), we remove the KL
term to avoid constraining model learning. During training, we use the AdamW optimizer (Loshchilov
& Hutter,) with a constant learning rate of
−6
. For rollout, we employ a prompt batch size of
256 and generate 8 responses per prompt. The maximum rollout length is set to 8,192 tokens, and the
sampling temperature is set as 1.0. We use a PPO mini-batch size of 256.
To assess in-domain and out-of-domain generalization under RLVR, we split Omni-MATH-Rule, a subset
of Omni-MATH (Gao et al.,) containing verifiable problems, into a training set (2,000 samples)
and an in-domain test set (821 samples), and use MATH500 as the out-of-domain benchmark.
Results.
pass@1 and pass@256, these differences are not fundamental. Different RL algorithms yield slightly
different SEvalues (i.e., ranging from GRPO’s 43.9 to RLOO’s best 42.6 on the in-domain test set).
Furthermore, we observe that SEremains consistently above 40 points across different algorithms,
highlighting that existing RL methods are still far from achieving optimal sampling efficiency. This
suggests that novel RL algorithms or entirely new paradigms may be necessary to approach the upper
bound. Additional observations can be found at Section.
9

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?1248163264128256
Number of Samples k
0.1
0.2
0.3
0.4
0.5
0.6
0.7
SE = 0.359
Omni-MATH-Train
1248163264128256
Number of Samples k
0.1
0.2
0.3
0.4
0.5
0.6
0.7
SE = 0.410
Omni-MATH-Test (In Domain)
1248163264128256
Number of Samples k
0.4
0.6
0.8
1.0
SE = 0.206
MATH500 (Out of Domain)
Qwen2.5-7B
GRPO
PPO
ReMax
RLOO
Reinforce++
DAPO 1248163264128256
Number of Samples k
0.1
0.2
0.3
0.4
0.5
0.6
0.7
SE = 0.247
Omni-MATH-Train
1248163264128256
Number of Samples k
0.1
0.2
0.3
0.4
0.5
0.6
0.7
SE = 0.408
Omni-MATH-Test (In Domain)
1248163264128256
Number of Samples k
0.4
0.6
0.8
1.0
SE = 0.199
MATH500 (Out of Domain)
Qwen2.5-7B
GRPO-step150
GRPO-step300
GRPO-step450
Figure 8:
for each point at pass@1 and pass@256 are provided in Table.
4.4. Effects of RL Training
Asymptotic Effects., we investigate the effect of the training
steps on the asymptotic performance of the model. As shown in Figure
progresses, pass@1 on the training set consistently improves from 26.1 to 42.5. However, as RLVR
training progresses, pass@256 progressively decreases, indicating a reduced reasoning boundary.
Effect of Number of Rolloutsn. n, the number of responses per
prompt, can affect pass@kby enabling broader exploration during training. We increasenfrom 8 to 32.
As shown in Figure, pass@ kimproves slightly overn= 8, but the RL-trained model is still eventually
outperformed by the base model. We leave the question of whether scaling RLVR training can eventually
surpass the base model to future investigation.
Effect of KL Loss.
applying a KL term with coefficient 0.001. As shown in Figure, the KL-regularized model achieves
similar pass@1 to GRPO without KL, but with a much lower pass@128.
4.5. Effects of Entropy
As RL training progresses, the model’s output entropy typically decreases (Yu et al.,), which may
contribute to a reduced reasoning boundary due to less diverse output. To investigate this factor, we
increase the generation temperature of the RLVR-trained model to match the output entropy of the base
model atT= 0.6. As shown in Figure, although the RLVR model performs slightly better pass@ kat
higher temperatures compared to its own performance atT= 0.6, it still underperforms the base model
across pass@k. This suggests that while reduced entropy contributes to the narrowing of the reasoning
boundary, it alone does not fully account for the reduction.
4.6. Effects of Model Size Scaling
Scaling plays a central role in the capabilities of contemporary LLMs. It remains an important question
whether the conclusions drawn continue to hold as model size increases. For many large models, isolating
the effect of RLVR is not feasible. For example, in the case of GPT-o1, the base model is not publicly
10

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
accessible. Qwen3-235B (Yang et al.,) is trained through multiple stages, including RLVR and
long-context CoT supervised fine-tuning, which makes it impossible to disentangle the impact of RLVR
alone. For Deepseek-R1-Zero, the absence of a publicly hosted API forced us to self-host the model,
but throughput was limited to around 50 tokens per second at a maximum sequence length of 32k,
rendering pass@kevaluation currently impractical. As a more tractable alternative, we selected the
Magistral-Medium-2506 API to conduct a preliminary set of experiments. This model is trained using
pure RL, with Mistral-Medium-3-2505 as the starting model (Rastogi et al.,). Although the model
size is not disclosed, Magistral-Medium performs comparably to Deepseek-R1 and is positioned near the
frontier in terms of reasoning capability.1 4 16642561024
Number of Samples (k)
0
0.2
0.4
0.6
0.8
1
Coverage (pass@k)
AIME24
Mistral-Medium-3 (Base)
Magistral-Medium (RLVR) 1 4 16642561024
Number of Samples (k)
0
0.2
0.4
0.6
0.8
1
Coverage (pass@k)
AIME25
Mistral-Medium-3 (Base)
Magistral-Medium (RLVR)
Figure 9: pass@k
We queried the models using a maximum
context length of 40k as the original pa-
per does. Once again, we observed that
RLVR provides significant gains at low
but little or no improvement at higherk.
Specifically, atk= 1, the RLVR-enhanced
model solves approximately 7 more prob-
lems on AIME24 and 8 more on AIME25
compared to its base version. However, as
kincreases, the performance gap steadily
narrows. These observations suggest that
our conclusion continues to hold even for
current, highly capable, near-frontier rea-
soning models. Whether this trend per-
sists as more compute, such as pre-training
scale budgets, is dedicated to RL training remains a critical question for the future of LLM reasoning.
5. Related Work
We summarize key related works on the analysis of RLVR here and provide a more comprehensive
discussion in Appendix. While recent RLVR methods have achieved impressive empirical results (Guo
et al.,;,), their fundamental impact on reasoning remains underexplored.
Several studies (Liu et al.,;,;,) suggest that reflective behaviors in
RLVR models originate from the base models rather than being learned through reinforcement learning.
Dang. (Dang et al.,) observed a decline in pass@ kperformance post-RLVR training, but their
analysis was limited in scope. More importantly, they did not explore the relationship between the base
model and the RL model. Deepseek-Math (Shao et al.,) also observed similar trends, but their
study was limited to a single instruction-tuned model and two math benchmarks. In contrast, our work
systematically investigates a wide range of models, tasks, and RL algorithms to accurately assess the
effects of current RLVR methods and models. We further provide in-depth analyses, including accuracy
distributions, reasoning coverage, perplexity trends, and comparison against distilled models, offering a
comprehensive understanding of RLVR’s capabilities and limitations.
6. Conclusion and Discussion
RLVR is widely regarded as a promising approach to enable LLMs to continuously self-improve and
acquire novel reasoning capabilities. In this paper, we systematically investigate the effect of current
RLVR methods on the reasoning capacity boundaries of LLMs. Surprisingly, our findings reveal that
current RLVR does not elicit fundamentally new reasoning patterns; instead, the reasoning capabilities of
RLVR-trained models remain bounded by those of their base models. These results indicate that current
RLVR methods have not fully realized the potential of reinforcement learning to elicit novel reasoning
abilities in LLMs through exploration and exploitation. This limitation may stem from the lack of
effective exploration strategies in the vast language space or the absence of multi-turn agent-environment
interactions needed to generate novel experience. We provide a more detailed discussion of the possible
causes of this gap and promising future directions in Section.
11

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
Acknowledgements
This work is supported in part by the National Key R&D Program of China under Grant 2022ZD0114903,
the National Natural Science Foundation of China under Grants 42327901 and U24B20173, and
the Scientific Research Innovation Capability Support Project for Young Faculty under Grant
ZYGXQNJSKYCXNLZCXM-I20.
Author Contributions
All authors
to the writing, editing, and overall management of the project.

Yang Yue (乐乐乐洋洋洋)
by the base model, and proposed the idea; designed the experiments and partially conducted
experiments; took primary responsibility for writing the manuscript.

Zhiqi Chen
benchmarks, and the perplexity analysis; contributed to discussions, figure creation, and manuscript
review.

Rui Lu
and manual check of AI reasoning trajectory.

Andrew Zhao
analysis, and contributed to the early implementation of the RL training code.

Zhaokai Wang
review.

Yang Yue (乐乐乐阳阳阳)
and figure refinement.

References
Achiam, J., Adler, S., Agarwal, S., Ahmad, L., Akkaya, I., Aleman, F. L., Almeida, D., Altenschmidt,
J., Altman, S., Anadkat, S., et al. Gpt-4 technical report., 2023.
Ahmadian, A., Cremer, C., Gallé, M., Fadaee, M., Kreutzer, J., Pietquin, O., Üstun, A., and Hooker, S.
Back to basics: Revisiting reinforce style optimization for learning from human feedback in llms.,
2024.,
Bai, S., Chen, K., Liu, X., Wang, J., Ge, W., Song, S., Dang, K., Wang, P., Wang, S., Tang, J., et al.
Qwen2.5-vl technical report., 2025.
Brown, B., Juravsky, J., Ehrlich, R., Clark, R., Le, Q. V., Ré, C., and Mirhoseini, A. Large language
monkeys: Scaling inference compute with repeated sampling., 2024.
2,
Chen, L., Li, L., Zhao, H., Song, Y., and Vinci. R1-v: Reinforcing super generalization ability in
vision-language models with less than $3.https://github.com/Deep-Agent/R1-V, 2025. Accessed:
2025-02-02.,
Chen, M., Tworek, J., Jun, H., Yuan, Q., Pinto, H. P. D. O., Kaplan, J., Edwards, H., Burda, Y.,
Joseph, N., Brockman, G., et al. Evaluating large language models trained on code.
arXiv:2107.03374, 2021.,
Cobbe, K., Kosaraju, V., Bavarian, M., Chen, M., Jun, H., Kaiser, L., Plappert, M., Tworek, J., Hilton,
J., Nakano, R., Hesse, C., and Schulman, J. Training verifiers to solve math word problems.
preprint arXiv:2110.14168, 2021.,
12

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
Dang, X., Baek, C., Kolter, J. Z., and Raghunathan, A. Assessing diversity collapse in reasoning.
In, 2025. URL https://
openreview.net/forum?id=AMiKsHLjQh.,
Gao, B., Song, F., Yang, Z., Cai, Z., Miao, Y., Dong, Q., Li, L., Ma, C., Chen, L., Xu, R., Tang,
Z., Wang, B., Zan, D., Quan, S., Zhang, G., Sha, L., Zhang, Y., Ren, X., Liu, T., and Chang, B.
Omni-math: A universal olympiad level mathematic benchmark for large language models, 2025.
Grattafiori, A., Dubey, A., Jauhri, A., Pandey, A., Kadian, A., Al-Dahle, A., Letman, A., Mathur, A.,
Schelten, A., Vaughan, A., et al. The llama 3 herd of models., 2024.
1,,
Gulcehre, C., Paine, T. L., Srinivasan, S., Konyushkova, K., Weerts, L., Sharma, A., Siddhant, A.,
Ahern, A., Wang, M., Gu, C., et al. Reinforced self-training (rest) for language modeling.
preprint arXiv:2308.08998, 2023.
Guo, D., Yang, D., Zhang, H., Song, J., Zhang, R., Xu, R., Zhu, Q., Ma, S., Wang, P., Bi, X., et al.
Deepseek-r1: Incentivizing reasoning capability in llms via reinforcement learning.
arXiv:2501.12948, 2025.,,,,,
He, C., Luo, R., Bai, Y., Hu, S., Thai, Z. L., Shen, J., Hu, J., Han, X., Huang, Y., Zhang, Y., et al.
Olympiadbench: A challenging benchmark for promoting agi with olympiad-level bilingual multimodal
scientific problems., 2024.
Hendrycks, D., Burns, C., Kadavath, S., Arora, A., Basart, S., Tang, E., Song, D., and Steinhardt, J.
Measuring mathematical problem solving with the math dataset., 2021.
Holtzman, A., Buys, J., Du, L., Forbes, M., and Choi, Y. The curious case of neural text degeneration.
ICLR, 2020.,
Hu, J. Reinforce++: A simple and efficient approach for aligning large language models.
arXiv:2501.03262, 2025.
Jaech, A., Kalai, A., Lerer, A., Richardson, A., El-Kishky, A., Low, A., Helyar, A., Madry, A., Beutel,
A., Carney, A., et al. Openai o1 system card., 2024.,,
Jain, N., Han, K., Gu, A., Li, W.-D., Yan, F., Zhang, T., Wang, S., Solar-Lezama, A., Sen, K., and
Stoica, I. Livecodebench: Holistic and contamination free evaluation of large language models for code.
ICLR, 2025.
Lambert, N., Morrison, J., Pyatkin, V., Huang, S., Ivison, H., Brahman, F., Miranda, L. J. V., Liu,
A., Dziri, N., Lyu, S., et al. Tulu 3: Pushing frontiers in open language model post-training.
preprint arXiv:2411.15124, 2024.,,
Lewkowycz, A., Andreassen, A., Dohan, D., Dyer, E., Michalewski, H., Ramasesh, V., Slone, A., Anil,
C., Schlag, I., Gutman-Solo, T., et al. Solving quantitative reasoning problems with language models.
NeurIPS, 2022.
Li, Z., Xu, T., Zhang, Y., Lin, Z., Yu, Y., Sun, R., and Luo, Z.-Q. Remax: A simple, effective, and
efficient reinforcement learning method for aligning large language models., 2024.
Liu, A., Feng, B., Xue, B., Wang, B., Wu, B., Lu, C., Zhao, C., Deng, C., Zhang, C., Ruan, C., et al.
Deepseek-v3 technical report., 2024.
Liu, J. and Zhang, L. Code-r1: Reproducing r1 for code with reliable rewards.https://github.com/
ganler/code-r1, 2025. GitHub repository.,
Liu, J., Xia, C. S., Wang, Y., and Zhang, L. Is your code generated by chatGPT really correct? rigorous
evaluation of large language models for code generation. In
13

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
Liu, Z., Chen, C., Li, W., Pang, T., Du, C., and Lin, M. There may not be aha moment in r1-zero-like
training – a pilot study., 2025a. Notion Blog.,
Liu, Z., Chen, C., Li, W., Qi, P., Pang, T., Du, C., Lee, W. S., and Lin, M. Understanding r1-zero-like
training: A critical perspective., 2025b.,,
Loshchilov, I. and Hutter, F. Decoupled weight decay regularization. In, 2017.
Lu, P., Gong, R., Jiang, S., Qiu, L., Huang, S., Liang, X., and Zhu, S.-C. Inter-gps: Interpretable
geometry problem solving with formal language and symbolic reasoning. In, 2021.
Lu, P., Bansal, H., Xia, T., Liu, J., Li, C., Hajishirzi, H., Cheng, H., Chang, K.-W., Galley, M., and
Gao, J. Mathvista: Evaluating mathematical reasoning of foundation models in visual contexts. In
ICLR, 2024.
Luo, M., Tan, S., Huang, R., Patel, A., Ariyak, A., Wu, Q., Shi, X., Xin, R., Cai, C., Weber, M., Zhang,
C., Li, L. E., Popa, R. A., and Stoica, I. Deepcoder: A fully open-source 14b coder at o3-mini level,
2025. Notion Blog.
Mnih, V., Kavukcuoglu, K., Silver, D., Rusu, A. A., Veness, J., Bellemare, M. G., Graves, A., Riedmiller,
M., Fidjeland, A. K., Ostrovski, G., et al. Human-level control through deep reinforcement learning.
nature, 518(7540):529–533, 2015.,
Ouyang, L., Wu, J., Jiang, X., Almeida, D., Wainwright, C., Mishkin, P., Zhang, C., Agarwal, S., Slama,
K., Ray, A., et al. Training language models to follow instructions with human feedback.,
2022.
Rafailov, R., Sharma, A., Mitchell, E., Manning, C. D., Ermon, S., and Finn, C. Direct preference
optimization: Your language model is secretly a reward model., 2023.
Ramamurthy, R., Ammanabrolu, P., Brantley, K., Hessel, J., Sifa, R., Bauckhage, C., Hajishirzi, H., and
Choi, Y. Is reinforcement learning (not) for natural language processing: Benchmarks, baselines, and
building blocks for natural language policy optimization. In, 2023.
Rastogi, A., Jiang, A. Q., Lo, A., Berrada, G., Lample, G., Rute, J., Barmentlo, J., Yadav, K.,
Khandelwal, K., Chandu, K. R., et al. Magistral., 2025.
Schulman, J., Wolski, F., Dhariwal, P., Radford, A., and Klimov, O. Proximal policy optimization
algorithms., 2017.,
Shah, D. J., Rushton, P., Singla, S., Parmar, M., Smith, K., Vanjani, Y., Vaswani, A., Chaluvaraju, A.,
Hojel, A., Ma, A., et al. Rethinking reflection in pre-training., 2025.
11,
Shao, Z., Wang, P., Zhu, Q., Xu, R., Song, J., Bi, X., Zhang, H., Zhang, M., Li, Y., Wu, Y., et al.
Deepseekmath: Pushing the limits of mathematical reasoning in open language models.
arXiv:2402.03300, 2024.,,
Shen, H., Zhang, Z., Zhao, K., Zhang, Q., Xu, R., and Zhao, T. Vlm-r1: A stable and generalizable
r1-style large vision-language model.https://github.com/om-ai-lab/VLM-R1, 2025. Accessed:
2025-02-15.,
Sheng, G., Zhang, C., Ye, Z., Wu, X., Zhang, W., Zhang, R., Peng, Y., Lin, H., and Wu, C. Hybridflow:
A flexible and efficient rlhf framework., 2024.
Silver, D. and Sutton, R. S. Welcome to the era of experience., 2025.
Silver, D., Schrittwieser, J., Simonyan, K., Antonoglou, I., Huang, A., Guez, A., Hubert, T., Baker, L.,
Lai, M., Bolton, A., et al. Mastering the game of go without human knowledge., 550(7676):
354–359, 2017.,
14

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
Sutton, R. S., Barto, A. G., et al., volume 1. MIT press
Cambridge, 1998.
Team, K., Du, A., Gao, B., Xing, B., Jiang, C., Chen, C., Li, C., Xiao, C., Du, C., Liao, C., et al. Kimi
k1. 5: Scaling reinforcement learning with llms., 2025.
Wang, K., Pan, J., Shi, W., Lu, Z., Ren, H., Zhou, A., Zhan, M., and Li, H. Measuring multimodal
mathematical reasoning with math-vision dataset. In, 2024.
7
Wang, Y., Ivison, H., Dasigi, P., Hessel, J., Khot, T., Chandu, K., Wadden, D., MacMillan, K., Smith,
N. A., Beltagy, I., et al. How far can camels go? exploring the state of instruction tuning on open
resources., 2023.
Williams, R. J. Simple statistical gradient-following algorithms for connectionist reinforcement learning.
Machine learning, 8:229–256, 1992.
Yang, A., Yang, B., Zhang, B., Hui, B., Zheng, B., Yu, B., Li, C., Liu, D., Huang, F., Wei, H., et al.
Qwen2.5 technical report., 2024.,
Yang, A., Li, A., Yang, B., Zhang, B., Hui, B., Zheng, B., Yu, B., Gao, C., Huang, C., Lv, C., et al.
Qwen3 technical report., 2025a.
Yang, A., Yu, B., Li, C., Liu, D., Huang, F., Huang, H., Jiang, J., Tu, J., Zhang, J., Zhou, J., et al.
Qwen2.5-1m technical report., 2025b.
Yu, Q., Zhang, Z., Zhu, R., Yuan, Y., Zuo, X., Yue, Y., Fan, T., Liu, G., Liu, L., Liu, X., et al. Dapo:
An open-source llm reinforcement learning system at scale., 2025.,
9,,
Yue, Y., Kang, B., Xu, Z., Huang, G., and Yan, S. Value-consistent representation learning for
data-efficient reinforcement learning. In, 2023.
Zelikman, E., Wu, Y., Mu, J., and Goodman, N. Star: Bootstrapping reasoning with reasoning.,
2022.
Zeng, W., Huang, Y., Liu, Q., Liu, W., He, K., Ma, Z., and He, J. Simplerl-zoo: Investigating and
taming zero reinforcement learning for open base models in the wild.,
2025.,
Zhao, R., Meterez, A., Kakade, S., Pehlevan, C., Jelassi, S., and Malach, E. Echo chamber: Rl
post-training amplifies behaviors learned in pretraining., 2025.,
Zheng, Y., Lu, J., Wang, S., Feng, Z., Kuang, D., and Xiong, Y. Easyr1: An efficient, scalable,
multi-modality rl training framework., 2025.,
15

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
Appendix
Appendix Contents
A Implementation Details
A.1 RLVR Algorithms
A.2 Low-Variancek
B More Related Works
C Discussion 18
D Detailed Experimental Results
D.1 More Results on Mathematics and Coding
D.2 Validity of Chain-of-Thought on AIME24
D.3 Accuracy Distribution Visulization
D.4 Perplexity Analysis
D.5 Different RLVR Algorithms
D.6 Effects of KL and Rollout Number
D.7 Solvable Problem Coverage Analysis
D.8 Temperature and Entropy Analysis
D.9 Training Dynamics
D.10 CoT Case Analysis
E Prompt Templates
F Broader Impacts 30
16

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
A. Implementation Details
A.1. RLVR Algorithms
To reduce memory and computational overhead, several critic-free variants have been proposed.
GRPO (Shao et al.,) estimates the advantage with a normalized reward within a group of responses
to the same question:Ai= [ri− (r)]/std(r) , wherer={r1, . . . , rG} denotes the set of rewards for
a group ofGsampled responses. RLOO (Ahmadian et al.,) instead adopts a leave-one-out baseline
within each batch i= i−
1
|B|−1
P
j=i
rj.
A.2. Low-Variancek
Directly computing pass@kusing onlyksampled outputs per problem can lead to high variance. To
mitigate this, we follow the unbiased estimation method proposed by Chen. (Chen et al.,).
Specifically, for each problemxifrom the evaluation datasetD, we generatensamples (n ) and
count the number of correct samples asci. The unbiased estimator of pass@kover the dataset is given
by:
pass@k xi∼D
"
1
`
n−ci
k
´
`
n
k
´
#
(2)
With this formulation, we can easily estimate pass@k.
In our experiments, we setnto the largest (i.e., rightmost) kvalue in the pass@kcurves, typically 128,
256, or 1024. For example, in Figure, we use n= 128
n= 1024 n= 128
and
B. More Related Works
Reinforcement Learning for LLM Reasoning.
phase has proven crucial to enhance problem solving and reasoning abilities (Ouyang et al.,). This
stage typically falls into three main categories: supervised fine-tuning using human-curated or distilled
data (Wang et al.,), self-improvement iteration (,;,),
and reinforcement learning (Ouyang et al.,). Previously, a reward model or preferences between
responses were employed for reward modeling (Ouyang et al.;,). Recently,
Reinforcement Learning with Verifiable Rewards (RLVR) has gained significant traction as a method to
improve the reasoning capabilities of LLMs in domains such as mathematics and programming (Lambert
et al.,;,). An encouraging landmark work is OpenAI’s o1 model (Jaech et al.,),
which was among the first large-scale applications of RL for reasoning, achieving state-of-the-art results
at the time of its release. Following this, Deepseek-R1 (Guo et al.,) became the first open-weight
model to match or surpass the performance of o1. A significant innovation introduced with R1 is
the “zero” setting, where reinforcement learning is applied directly to the base LLM, bypassing any
intermediate supervised tuning. This approach inspired a wave of open-source efforts to replicate or
extend R1’s methodology and improve RL algorithms (Zeng et al.;,;,;
Liu & Zhang,). In parallel, reinforcement learning has also gained attention in the multimodal
domain, driving advancements in multimodal reasoning (Chen et al.,;,;
et al.,).
Analysis of RLVR. Although there are many excellent open-source works and algorithmic designs in
the field of RLVR, there remains a lack of deep understanding regarding the root effects of RLVR on
LLM reasoning abilities and its limitations when starting from the base model. Several studies (Liu
et al.,;,;,) highlight that the reflective behaviors observed in
R1-like models actually emerge from the base models, rather than being introduced by RLVR training.
Dang. (Dang et al.,
rapidly and fails to recover with reinforcement learning, but this was seen only in a limited experimental
setup with Qwen-2.5-0.5B on GSM8K. More importantly, they did not explore the relationship between
17

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
the base model and the RL model. In contrast, our paper conducts systematic and rigorous experiments
to show that not only reflective behaviors but all reasoning paths are already embedded in the base
model. We further demonstrate that RLVR does not elicit any new reasoning abilities beyond the base
model.
C. Discussion
In Section, we identified key limitations of RLVR in enhancing LLM reasoning capabilities.
In this section, we explore possible underlying factors that may explain why RLVR remains bounded by
the reasoning capacity of the base model.
Discussion 1: Key Differences Between Traditional RL and RLVR for LLMs are Vast Action
Space and Pretrained Priors.Silver
et al.,;,;,) can continuously improve the performance of a policy in
environments like Go and Atari games. There are two key differences
between traditional RL and RLVR for LLMs. First, the action space in language models is exponentially
larger than that of Go or Atari games (Ramamurthy et al.,
designed to handle such a vast action space, which makes it nearly impossible to explore the reward
signal effectively if training starts from scratch. Therefore, the second distinction is that RLVR for LLMs
starts with a pretrained base model with useful prior, whereas traditional RL in Atari and GO games
often begins from scratch. This pretrained prior guides the LLM in generating reasonable responses,
making the exploration process significantly easier, and the policy can receive positive reward feedback.
Discussion 2: Priors as a Double-Edged Sword in This Vast Action Space.
of responses is guided by the pretrained prior, the policy may struggle to explore new reasoning patterns
beyond what the prior already provides. Specifically, in such a complex and highly combinatorial
space, most responses generated during training are constrained by the base model’s prior. Any sample
deviating from the prior is highly likely to produce invalid or non-sensical outputs, leading to negative
reward. As discussed in Section, policy gradient algorithms aim to maximize the log-likelihood of
responses within the prior that receive positive rewards, while minimizing the likelihood of responses
outside the prior that receive negative rewards. As a result, the trained policy tends to produce responses
already present in the prior, constraining its reasoning ability within the boundaries of the base model.
From this perspective, training RL models from a distilled model may temporarily provide a beneficial
solution, as distillation helps inject a better prior.
Possible Future Work.
the vast action space, facilitating the discovery of out-of-prior reasoning patterns and the development
of more capable reasoning models. Furthermore, current RLVR frameworks are limited to single-turn
interactions with the verifier, lacking the ability to iteratively revise or improve based on environmen-
tal feedback. A multi-turn RL agent paradigm—with richer, ongoing interaction with a grounded
environment—could enable models to generate novel experiences and learn from them. This agent
paradigm has been described as the beginning of an “era of experience” (Silver & Sutton,).
18

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
D. Detailed Experimental Results
D.1. More Results on Mathematics and Coding1416642561024
Number of Samples k
0
0.2
0.4
0.6
0.8
1
Coverage (pass@k)
Base
RL
1416642561024
0
0.2
0.4
0.6
0.8
1
1416642561024
0
0.2
0.4
0.6
0.8
1
1416642561024
0
0.2
0.4
0.6
0.8
1
1248163264128
0.8
0.85
0.9
0.95
1
1248163264128
0.8
0.85
0.9
0.95
1
1248163264128
0.8
0.85
0.9
0.95
1
1248163264128
0
0.2
0.4
0.6
0.8
1
AMC23
GSM8K
Qwen-2.5-7B Qwen-2.5-14B Qwen-2.5-32B LLaMA-3.1-8B
Figure 10: More results of SimpleRLZoo on GSM8K and AMC23.1416642561024
Number of Samples k
0
0.2
0.4
0.6
0.8
1.0
Coverage (pass@k)
AIME24
Qwen-2.5-Math-7B
Qwen-2.5-Math-7B-Oat-Zero 1 4 16642561024
Number of Samples (k)
0
0.2
0.4
0.6
0.8
1
Coverage (pass@k)
AIME24
Qwen-2.5-Base-32B
DAPO-32B Figure 11: Oat-Zero-7B and DAPO-32B are evaluated on AIME24 and compared against their respective
base models.1248163264128
Number of Samples k
0.1
0.2
0.3
Coverage (pass@k)
LiveCodeBench (2024.8-2025.1)
Qwen-2.5-7B-Instruct-1M
Coder-R1-Zero-Qwen-2.5-7B 1248163264128
Number of Samples k
0.1
0.2
0.3
0.4
0.5
Coverage (pass@k)
LiveCodeBench (2023.5-2025.1)
Qwen-2.5-7B-Instruct-1M
Coder-R1-Zero-Qwen-2.5-7B
Figure 12: Coder-R1 on LiveCodeBench.
19

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
D.2. Validity of Chain-of-Thought on AIME24
We manually check the CoTs for the most challenging AIME24 benchmark. To begin, we introduce a
filtering mechanism designed to eliminate easily guessable problems. Specifically, we prompt Qwen2.5-
7B-Base to answer questions directly, without using chain-of-thought reasoning, and sample answers
multiple times. If a problem can be answered correctly with a low but non-zero probability (e.g., <5%),
we consider it to be guessable and remove it. Problems that can be directly answered correctly with a
high probability are retained, as they are likely easier and solvable using valid CoTs.1 4 16 642561024
Number of Samples k
0
0.2
0.4
0.6
0.8
1
Coverage (pass@k)
AIME24
Base
RL
Figure 13: Pass@k
The base and RL model pass@kcurves on this filtered AIME24 can be found in Figure, showing a
similar trending to previous results. Although this filtering method is heuristic, it proves to be effective.
Applying it to AIME24 (30 questions) results in a subset of 18 questions. We then prompt the models to
answer these filtered questions using CoT reasoning. Then we perform a manual inspection of all CoTs
that led to correct answers on the hardest problems – those with an average accuracy below 5%. The
base model answered 7 such questions, with 5 out of 6 containing
one ambiguous case of correctness due to skipped reasoning steps). Similarly, the RL-trained model
answered 6 questions, 4 of which included
the hardest questions in the challenging AIME24, base model can sample valid reasoning paths to solve
the problems.
20

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
D.3. Accuracy Distribution Visulization0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
Accuracy Interval
0
2
4
6
8
10
12
14
16
18
Frequency
Base
RL
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
3
6
9
12
15
18
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
3
6
9
12
15
18
21
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
3
6
9
12
15
18
21
24
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
2
4
6
8
10
12
14
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
2
4
6
8
10
12
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
2
4
6
8
10
12
14
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
5
10
15
20
25
30
35
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
50
100
150
200
250
300
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
60
120
180
240
300
360
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
60
120
180
240
300
360
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
40
80
120
160
200
240
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
20
40
60
80
100
120
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
20
40
60
80
100
120
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
20
40
60
80
100
120
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
30
60
90
120
150
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
180
360
540
720
900
1080
1260
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
200
400
600
800
1000
1200
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
200
400
600
800
1000
1200
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
70
140
210
280
350
420
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
40
80
120
160
200
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
40
80
120
160
200
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
40
80
120
160
200
0.0
(0.0,0.1](0.1,0.2](0.2,0.3](0.3,0.4](0.4,0.5](0.5,0.6](0.6,0.7](0.7,0.8](0.8,0.9](0.9,1.0)
1.0
0
60
120
180
240
300
360
0
10
20
30
40
50
60
Percentage (%)
0
10
20
30
40
0
10
20
30
40
50
60
0
10
20
30
40
0
10
20
30
40
50
60
70
80
90
0
10
20
30
0
10
20
30
40
50
60
70
0
10
20
30
0
10
20
30
40
50
60
70
0
10
20
30
40
0
10
20
30
40
50
60
70
80
90
100
0
10
20
30
0
10
20
30
40
50
60
70
0
10
20
30
40
0
10
20
30
40
50
60
70
80
0
10
20
30
40
0
10
20
30
40
50
60
70
80
90
100
0
10
20
30
0
10
20
30
40
50
60
70
80
0
10
20
30
40
50
60
70
80
90
100
0
10
20
30
40
50
0
10
20
30
40
50
60
0
10
20
30
0
10
20
30
40
50
AIME24
AMC23
MATH500
Minerva
GSM8K
Olympiad
Qwen-2.5-7B Qwen-2.5-14B Qwen-2.5-32B LLaMA-3.1-8B
Figure 14: Accuracy histogram before and after RLVR with SimpleRLZoo models.
21

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
D.4. Perplexity Analysis
To analyze how perplexity evolves over the course of RLVR training, we evaluated three RLVR checkpoints–
early, middle, and final in Section
problem, computed the median among 32 perplexity values, and reported the average over the first
10 problems in the table. As expected, we observed thatPPLBase (YRL|x)
training progresses, indicating that RLVR mainly sharpens the distribution within the base model’s
prior rather than expanding beyond it.
Figure 15: Perplexity Evolution during RL Training.
D.5. Different RLVR Algorithms
We report several additional observations on different RLVR algorithms in Figure. First, DAPO
achieves slightly higher pass@1 scores across all three datasets; however, its dynamic sampling strategy
requires approximately ∼6×more samples per batch during training compared to other algorithms.
Moreover, its performance drops significantly atk= 256. Second, RLOO and Reinforce++ perform
consistently well across the entirekrange (from 1 to 256), while maintaining efficient training costs,
achieving a good balance between effectiveness and efficiency. Third, ReMax shows lower performance
at both pass@1 and pass@256. We hypothesize that this is due to its use of the greedy response reward
as the advantage baseline, which in the RLVR setting is binary (0 or 1) and highly variable. This likely
results in unstable gradient updates during training.
Table 3: Detailed values for each point at pass@1 and pass@256 across different RL algorithms in Figure.
Model Omni-MATH-Train Omni-MATH-Test MATH500
pass@1 pass@256 pass@1 pass@256 pass@1 pass@256
Qwen2.5-7B 9.9 67.2 10.2 69.1 34.5 96.2
GRPO 26.1 66.3 25.1 68.3 74.4 97.2
PPO 27.2 65.8 26.8 69.2 75.2 97.2
ReMax 24.4 65.5 23.8 67.5 73.5 96.6
RLOO 28.6 66.4 28.1 75.0
Reinforce++28.2 28.0 69.7 75.4 96.8
DAPO 31.4 26.5 67.0 75.6
Table 4: Detailed values at pass@1 and pass@256 across different RL training steps in Figure
Model Omni-MATH-Train Omni-MATH-Test MATH500
pass@1 pass@256 pass@1 pass@256 pass@1 pass@256
Qwen2.5-7B 9.9 10.2 34.5 96.2
GRPO-step150 26.1 66.3 25.1 68.3 74.4
GRPO-step300 33.6 65.3 27.1 66.6 75.4 96.0
GRPO-step450 42.5 28.3 76.3
22

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
D.6. Effects of KL and Rollout Number1248163264128256
Number of Samples k
0.4
0.5
0.6
0.7
Omni-MATH-Train
1248163264128256
Number of Samples k
0.3
0.4
0.5
0.6
0.7
Omni-MATH-Test (In Domain)
1248163264128256
Number of Samples k
0.7
0.8
0.9
MATH500 (Out of Domain)
GRPO (n=8, 450 steps) n=32, 220 steps +kl=0.001, 450 steps Qwen2.5-7B
Figure 16: n. nfrom 8 to 32,
we keep the prompt batch size unchanged, which results in increased computation per training step.
Due to resource constraints, we train for only 220 steps under this setting, leading to lower pass@1 as
the model has not yet converged. Nevertheless, the model withn= 32
highlighting the positive effect of larger rollout numbers in improving pass@k
D.7. Solvable Problem Coverage Analysis
Table
problem at least once, (2) only the base model solves it, (3) only the RLVR model solves it, and (4)
neither model solves it in any of theksamples. It highlights that there are many cases where the base
model solves a problem but RLVR fails (type 2), and very few where RLVR succeeds while the base
does not (type 3). Even in the rare type 3 cases (e.g., 1% or about 5 problems in MATH500), the base
model is able to solve all of them when sampling 1024 times. These results support our conclusion that
RLVR rarely solves problems the base model cannot and generally results in reduced coverage.
Table 5: Indices of solvable problems in AIME24 (starting from 0). An approximate subset relationship
can be observed: most problems solved by the RL model are also solvable by the base model.
Models Problem Indices
Qwen2.5-7B-Base 0, 1, 4, 6, 7, 8, 9, 11, 12, 14, 15, 16, 17, 18, 19, 22, 23, 24, 25, 26, 27, 28, 29
SimpleRL-Qwen2.5-7B 0, 1, 6, 7, 8, 9, 12, 14, 15, 16, 18, 22, 23, 24, 25, 26, 27, 28, 29
Table 6: Indices of solvable problems in LiveCodeBench (ranging from 400 to 450, starting from 0).
Model Solvable Problem Indices
Qwen2.5-7B-Instruct-1M
400, 402, 403, 407, 409, 412, 413, 417, 418, 419, 422, 423,
427, 432, 433, 436, 438, 439, 440, 444, 445, 448, 449
Coder-R1
400, 402, 403, 407, 412, 413, 417, 418, 419, 422, 423,
427, 430, 433, 438, 439, 440, 444, 445, 449
23

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
D.8. Temperature and Entropy Analysis1416642561024
Number of Samples k
0
0.2
0.4
0.6
0.8
Coverage (pass@k)
Base
RL
1416642561024
0
0.2
0.4
0.6
0.8
1416642561024
0
0.2
0.4
0.6
0.8
1248163264128
0.2
0.4
0.6
0.8
1
1248163264128
0.2
0.4
0.6
0.8
1
1248163264128
0.2
0.4
0.6
0.8
1
AIME24
MATH500
T = 0.6 T = 1.0 T = 1.2
Figure 17: We found that the base model’s performance drops when the temperature exceeds 1.0, as it tends to generate
more random and less coherent tokens. In contrast, the RL model’s performance remains relatively stable across different
temperature settings. Therefore, we useT= 0.6
best reasoning performance.1416642561024
0
0.2
0.4
0.6
0.8
AIME24 (Ebase=0.47)
Base (T=0.6)
RL (T=0.6)
RL (T=1.5)
1416642561024
0.2
0.4
0.6
0.8
1
AMC23 (Ebase=0.51)
Base (T=0.6)
RL (T=0.6)
RL (T=0.9)
1248163264128
0.8
0.85
0.9
0.95
1.0
GSM8K (Ebase=0.75)
Base (T=0.6)
RL (T=0.6)
RL (T=1.0)
1248163264128
Number of Samples k
0.6
0.7
0.8
0.9
1
Coverage (pass@k)
MATH500 (Ebase=0.72)
Base (T=0.6)
RL (T=0.6)
RL (T=0.85)
1248163264128
0.2
0.4
0.6
0.8
1
Minerva (Ebase=0.63)
Base (T=0.6)
RL (T=0.6)
RL (T=0.85)
1248163264128
0.2
0.4
0.6
0.8
1
Olympiad (Ebase=0.5)
Base (T=0.6)
RL (T=0.6)
RL (T=1.0)
Figure 18:
(Qwen2.5-7B) on each dataset using temperatureT= 0.6 Ebasein the title of each figure.
To enable a fair comparison, we increase the temperature of the RLVR model (SimpleRLZoo) until its output entropy
approximately matchesEbase. For example, on AMC23, we setT= 0.9 ERL= 0.47. We also include RLVR
results at.6
24

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
D.9. Training Dynamics0 60 120 180 240 300 360 420 480
Training Steps
0.10
0.15
0.20
0.25
0.30
0.35
0.40
Reward
Actor Reward
GRPO
PPO
Reinforce++
GRPO (n=32)
GRPO (KL=0.001) 0 60 120 180 240 300 360 420 480
Training Steps
400
500
600
700
800
900
1000
1100
1200
Response Length
Actor Response Length
GRPO
PPO
Reinforce++
GRPO (n=32)
GRPO (KL=0.001) 0 60 120 180 240 300 360 420 480
Training Steps
0.0
0.2
0.4
0.6
0.8
1.0
Entropy
Actor Generation Entropy
GRPO
PPO
Reinforce++
GRPO (n=32)
GRPO (KL=0.001)
Figure 19: The curves of training reward, response length, and generation entropy during training,
corresponding to experiments in Section.
25

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
D.10. CoT Case Analysis
Question

(AIME24 Question No.16)
Among the 900 residents of Aimeville, there are 195 who own a diamond ring, 367 who own a set of golf clubs, and 562 who own a garden
spade. In addition, each of the 900 residents owns a bag of candy hearts. There are 437 residents who own exactly two of these things, and
234 residents who own exactly three of these things. Find the number of residents of Aimeville who own all four of these things.
Qwen-2.5-7B Answer Response Length: 1058

To solve this problem, we can use the principle of inclusion-exclusion. We will denote the number of residents who own diamond rings, golf
clubs, garden spades, and candy hearts as D, G, S, and C, respectively. We are given the following information:
- D = 195
- G = 367
- S = 562
- C = 900 (since every resident owns a bag of candy hearts)
We are also given that 437 residents own exactly two of these things and 234 residents own exactly three of these things.
Let's denote the number of residents who own all four of these things as X. We need to find X.
First, we need to find the number of residents who own at least one of these things. This can be calculated using the principle of inclusion-
exclusion:
Total=&#3627408439;+��+��+&#3627408438;−(number of residents who own exactly two things)
−2×(number of residents who own exactly three things)−3×(number of residents who own all four things)

Since there are 900 residents in total, we can write the equation as:
900=195+367+562+900−437−2×234−3×&#3627408459;
Now we need to solve for X:
900=195+367+562+900−437−468−3×&#3627408459;
900=1717−905−3×&#3627408459;
900=812−3×&#3627408459;
3×&#3627408459;=812−900
3×&#3627408459;=−88
&#3627408459;=−
88
3

Since the number of residents cannot be negative, there is an error in our calculations. Let's re-evaluate the problem.
We know that there are 437 residents who own exactly two things and 234 residents who own exactly three things. We can subtract these
from the total number of residents who own at least one thing to find the number of residents who own exactly one thing:
Number of residents who own exactly one thing=(&#3627408439;+��+��+&#3627408438;)−(number of residents who own exactly two things)
−2×(number of residents who own exactly three things)−3×(number of residents who own all four things)

Let's denote the number of residents who own exactly one thing as Y:
&#3627408460;=(195+367+562+900)−437−2×234−3×&#3627408459;
&#3627408460;=1924−905−3×&#3627408459;
&#3627408460;=1019−3×&#3627408459;

Since every resident owns at least one thing, we can write the equation:
900=&#3627408460;+437+2×234+3×&#3627408459;
Now we need to solve for X:
900=(1019−3×&#3627408459;)+437+468+3×&#3627408459;
900=1019+437+468
900=1924

This equation is not true, which means there is still an error in our calculations. Let's try a different approach.
We can use the fact that the total number of residents who own at least one thing is 900. We can write the equation:
900=(number of residents who own exactly one thing)+(number of residents who own exactly two things)
+(number of residents who own exactly three things)+(number of residents who own all four things)

Let's denote the number of residents who own exactly one thing as Y:
900=&#3627408460;+437+2×234+&#3627408459;
900=&#3627408460;+805+&#3627408459;

Now we need to find Y. We can use the fact that the total number of residents who own at least one thing is equal to the sum of the residents
who own each item minus the residents who own at least two items:
900=(195+367+562+900)−(number of residents who own exactly two things)
−2×(number of residents who own exactly three things)−3×(number of residents who own all four things)

900=1924−437−2×234−3×&#3627408459;
900=1924−805−3×&#3627408459;
900=1119−3×&#3627408459;

Now we can solve for X:
3×&#3627408459;=1119−900
3×&#3627408459;=219
&#3627408459;=73


So, the number of residents of Aimeville who own all four of these things is 73.

Figure 20: Qwen2.5-Base-7B Correct Response - Case 1.
26

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
Question

(AIME24 Question No.24)
Find the number of ways to place a digit in each cell of a 2x3 grid so that the sum of the two numbers
formed by reading left to right is 999, and the sum of the three numbers formed by reading top to bottom is
99. The grid below is an example of such an arrangement because 8+991=999 and 9+9+81=99.
Qwen-2.5-7B Answer Response Length: 967

To solve this problem, we need to determine the number of ways to place digits in a 2x3 grid such that the sum
of the two numbers formed by reading left to right is 999, and the sum of the three numbers formed by reading
top to bottom is 99. Let's break this down step by step.

1. Define the grid:
Let the grid be represented as:
&#3627408462;&#3627408463;&#3627408464;
&#3627408465;&#3627408466;&#3627408467;

The numbers formed by reading left to right are 100&#3627408462;+10&#3627408463;+&#3627408464; and 100&#3627408465;+10&#3627408466;+&#3627408467;. The numbers formed
by reading top to bottom are 10&#3627408462;+&#3627408465;, 10&#3627408463;+&#3627408466;, and 10&#3627408464;+&#3627408467;.

2. Set up the equations:
- The sum of the two numbers formed by reading left to right is 999:
100&#3627408462;+10&#3627408463;+&#3627408464;+100&#3627408465;+10&#3627408466;+&#3627408467;=999
- The sum of the three numbers formed by reading top to bottom is 99:
10&#3627408462;+&#3627408465;+10&#3627408463;+&#3627408466;+10&#3627408464;+&#3627408467;=99

3. Simplify the equations:
- From the first equation:
100&#3627408462;+10&#3627408463;+&#3627408464;+100&#3627408465;+10&#3627408466;+&#3627408467;=999
This implies:
&#3627408462;+&#3627408465;=9,&#3627408463;+&#3627408466;=9,&#3627408464;+&#3627408467;=9
- From the second equation:
10&#3627408462;+&#3627408465;+10&#3627408463;+&#3627408466;+10&#3627408464;+&#3627408467;=99
This is consistent with the previous simplification since:
10(&#3627408462;+&#3627408463;+&#3627408464;)+(&#3627408465;+&#3627408466;+&#3627408467;)=99
Given &#3627408462;+&#3627408465;=9, &#3627408463;+&#3627408466;=9, and &#3627408464;+&#3627408467;=9, it follows that:
10(&#3627408462;+&#3627408463;+&#3627408464;)+(9−&#3627408462;)+(9−&#3627408463;)+(9−&#3627408464;)=99
Simplifying, we get:
10(&#3627408462;+&#3627408463;+&#3627408464;)+27−(&#3627408462;+&#3627408463;+&#3627408464;)=99
9(&#3627408462;+&#3627408463;+&#3627408464;)+27=99
9(&#3627408462;+&#3627408463;+&#3627408464;)=72
&#3627408462;+&#3627408463;+&#3627408464;=8
Since &#3627408462;+&#3627408465;=9, &#3627408463;+&#3627408466;=9, and &#3627408464;+&#3627408467;=9, and &#3627408462;+&#3627408463;+&#3627408464;=8, it follows that:
&#3627408465;=9−&#3627408462;,&#3627408466;=9−&#3627408463;,&#3627408467;=9−&#3627408464;

4. Determine the possible values:
- &#3627408462;,&#3627408463;,&#3627408464; must be digits (0-9) such that &#3627408462;+&#3627408463;+&#3627408464;=8.
- For each valid combination of &#3627408462;,&#3627408463;,&#3627408464;, &#3627408465;,&#3627408466;,&#3627408467; are uniquely determined as &#3627408465;=9−&#3627408462;, &#3627408466;=9−&#3627408463;, &#3627408467;=9−&#3627408464;.

5. Count the combinations:
- We need to count the number of non-negative integer solutions to &#3627408462;+&#3627408463;+&#3627408464;=8 with 0≤&#3627408462;,&#3627408463;,&#3627408464;≤9.
- This is a classic "stars and bars" problem with an additional constraint. The number of solutions is given by
the binomial coefficient:
(
8+3−1
3−1
)=(
10
2
)=45

Therefore, the number of ways to place the digits in the grid is 45.


Figure 21: Qwen2.5-Base-7B Correct Response - Case 2.
27

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
E. Prompt Templates
We provide the prompt templates used for training and evaluation in our experiments. The prompt
for SimpleRL training and evaluation is shown in Figure, while the prompt for Oat-Zero is shown
in Figure. For Code-R1 training, prompt in Figure
the original codebase and adopt the default templates from the benchmarks, including LiveCodeBench
prompt (Figure), HumanEval+, and MBPP+ prompt (Figure). The prompt used for EasyR1
training and evaluation is shown in Figure. For VeRL-trained RL models, as discussed in Section
and Section, the training and evaluation prompts are provided in Figure. For evaluating Mistral
and Magistral models on AIME24/25, prompts are provided in Figure. To ensure a fair comparison,
the base models use the same prompts as their corresponding RL-trained counterparts during evaluation.<|im_start|>system
You are a helpful assistant.<|im_end|>
<|im_start|>user
{question}
Please reason step by step, and put your final answer within\\boxed{{}}.<|im_end|>
<|im_start|>assistant
SimpleRLPrompt
Figure 22: Prompt for SimpleRL Training and Evaluation. The base model uses the same prompt as
the RL model during evaluation.<|im_start|>system
Please reason step by step, and put your final answer within \\boxed{}.<|im_end|>
<|im_start|>user
{question}<|im_end|>
<|im_start|>assistant
OatPrompt
Figure 23: Prompt for Oat-Zero training and evaluation.<|im_start|>system
You are a helpful programming assistant. The user will ask you a question and you
as the assistant solve it. The assistant first thinks how to solve the task through
reasoning and then provides the user with the final answer. The reasoning process
and answer are enclosed within <think>...</think> and <answer>...</answer> tags,
respectively.<|im_end|>
<|im_start|>user
{question}<|im_end|>
<|im_start|>assistant
Code-R1 Prompt Figure 24: Prompt for Code-R1 training.
28

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?<|im_start|>system
You are a helpful assistant.<|im_end|>
<|im_start|>user
You will be given a question (problem specification) and will generate a correct
Python program that matches the specification and passes all tests. You will NOT
return anything except for the program.
Question: {question.question_content}
{ if question.starter_code }
### Format:
{PromptConstants.FORMATTING_MESSAGE_WITH_STARTER_CODE }
```python
{question.starter_code}
```
<|im_end|>
{ else }
### Format:
{PromptConstants.FORMATTING_MESSAGE_WITHOUT_STARTER_CODE }
```python
# YOUR CODE HERE
```
<|im_end|>
<|im_start|>assistant
```python
LiveCodeBench (Code Generation) Prompt
Figure 25: Since Code-R1 does not specify an evaluation prompt, we adopt the original LiveCodeBench
evaluation prompt. To encourage both the base and RL-trained models to generate code, we append
“‘pythonto the end of the prompt. Using this setup, we reproduce a pass@1
close to the reported 29.7.<|im_start|>system
You are a helpful assistant.<|im_end|>
<|im_start|>user
Please provide a self-contained Python script that solves the following problem in a
markdown code block:
```
{python_task_prompt}
```
<|im_end|>
<|im_start|>assistant
Below is a Python script with a self-contained function that solves the problem and
passes corresponding tests:
```python
HumanEval+ & MBPP+Prompt
Figure 26: Prompt for Code-R1 Evaluation on HumanEval+ and MBPP+.
29

Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?<|im_start|>system
You are Qwen, created by Alibaba Cloud. You are a helpful assistant. You FIRST
think about the reasoning process as an internal monologue and then provide the
final answer. The reasoning process MUST BE enclosed within <think> </think>
tags. The final answer MUST BE put in \boxed{}.<|im_end|>
<|im_start|>user
<|vision_start|>{image_token}<|vision_end|>
{question}<|im_end|>
<|im_start|>assistant
EasyR1 Prompt
Figure 27: Prompt for EasyR1 training and evaluation.A conversation between User and Assistant. The user asks a question, and the
Assistant solves it. The assistant first thinks about the reasoning process in the mind
and then provides the user with the answer. The reasoning process and answer are
enclosed within <think> </think> and <answer> </answer> tags, respectively, i.e.,
<think> reasoning process here </think> <answer> answer here </answer>. User: {}
Assistant:
VeRLTraining and Evaluation Prompt Figure 28: Prompt for VeRL training on Omni-math-train and evaluation on Omni-math-eval and
MATH500.A conversation between User and Assistant. The user asks a question, and the
Assistant solves it. The assistant first thinks about the reasoning process in the
mind and then provides the user with the answer. The reasoning process and answer
are enclosed within <think> </think> and <answer> </answer> tags, respectively,
i.e., <think> reasoning process here </think> <answer> answer here </answer>.
User: {question}.
Assistant:
Mistral& MagistralPrompt
Figure 29: Prompt for Mistral and Magistral model evaluation.
F. Broader Impacts
The potential negative social impacts of our method align with those typically associated with general
LLM reasoning technologies. We emphasize the importance of adhering to the principles of fair and safe
deployment in LLM systems.
30