Fast-Slow Thinking GRPO for Large Vision-Language Model Reasoning

Wenyi Xiao 1
Zhejiang University
wenyixiao@zju.edu.cn
Leilei Gan\(^{\dagger}\)
Zhejiang University
leileigan@zju.edu.cn
2


Abstract

When applying reinforcement learning—typically through GRPO—to large vision-language model reasoning struggles to effectively scale reasoning length or generates verbose outputs across all tasks with only marginal gains in accuracy. To address this issue, we present FAST-GRPO, a variant of GRPO that dynamically adapts reasoning depth based on question characteristics. Through empirical analysis, we establish the feasibility of fast-slow thinking in LVLMs by investigating how response length and data distribution affect performance. Inspired by these observations, we introduce two complementary metrics to estimate the difficulty of the questions, guiding the model to determine when fast or slow thinking is more appropriate. Next, we incorporate adaptive length-based rewards and difficulty-aware KL divergence into the GRPO algorithm. Experiments across seven reasoning benchmarks demonstrate that FAST achieves state-of-the-art accuracy with over 10% relative improvement compared to the base model, while reducing token usage by 32.7-67.3% compared to previous slow-thinking approaches, effectively balancing reasoning length and accuracy.

1 Introduction↩︎

Slow-thinking reasoning has demonstrated remarkable capabilities in solving complex tasks in Large Language Models (LLMs) [1][3] by applying large-scale reinforcement learning (RL), exemplified by OpenAI’s o1 [4], DeepSeek-R1 [5], and Qwen’s QwQ [6].

Figure 1: FAST achieves higher average accuracy with shorter average response lengths across seven benchmarks. All methods are built upon Qwen2.5-VL.

Unlike fast-thinking models [7], [8], slow-thinking models undertake more deliberate and thorough reasoning before reaching an answer, which facilitates the exploration of diverse solution paths for a given problem.

Researchers [9][12] have begun exploring similar slow thinking approaches for large vision-language models (LVLMs) to enhance visual reasoning, which can be categorized into SFT-RL two-stage methods [9], [13][16] and RL-only methods [17][20]. SFT-RL methods collect large-scale distilled data from slow-thinking models before applying reinforcement learning, while RL-only methods directly employ reinforcement learning on curated high-quality data.

Despite these efforts, several challenges persist in slow-thinking for LVLM reasoning. First, while RL-only methods enable slow-thinking LVLMs to improve reasoning accuracy, they struggle to effectively scale reasoning length [15], [17], [19], with observed changes ranging only from –20% to +10% compared to base models. This limited adaptability in reasoning length may constrain their effectiveness on complex tasks. Second, in contrast, we observe that slow-thinking LVLMs with SFT-RL methods [9], [13], [14], [21] exhibit a pronounced overthinking phenomenon—producing overly verbose responses across tasks while yielding only marginal improvements in accuracy. This observation suggests that excessive verbosity may arise from the SFT stage, which performs behavior cloning from distilled data. As evidenced in Table 1, R1-OneVision (one slow thinking model with SFT-RL) produces reasoning chains approximately 2× longer than its base model across all difficulty levels on the Geometry [22] test set.

r7cm

Table 1: Comparison of accuracy and response length on Geometry 3K [22] test set across difficulty levels for Qwen2.5-VL-7B, R1-OneVision, and FAST.
Test Qwen2.5-VL R1-OneVision FAST
2-3 (lr)4-5 (lr)6-7 Acc. Len. Acc. Len. Acc. Len.
Easy 72.7 318 69.5 623 78.2 189
Med 33.9 406 40.4 661 49.2 220
Hard 5.5 412 10.2 835 12.3 304
All 37.7 378 40.3 731 46.4 239

Notably, this overthinking proves detrimental for simpler questions, where extended reasoning results in accuracy degradation (69.5% vs. 72.7%), highlighting the need for adaptive fast-slow thinking.

We notice that current research on addressing the overthinking phenomenon primarily focuses on large language models (LLMs) and can be classified into two categories based on the stage of application. In the training stage, they design length reward shaping in RL training to explicitly encourage concise model responses. [23][26] In the inference stage, they enforce concise reasoning via prompts, e.g., use less than 50 tokens, to constrain response length [27], [28]. However, these methods to address overthinking in LLMs ignore challenges of visual inputs and question characteristics in visual reasoning [23][25], leaving their effectiveness in LVLMs largely unexplored. To our knowledge, no existing work effectively balances fast and slow thinking in LVLMs.

To address these issues, we propose FAST-GRPO, a tailored variant of GRPO [5], [7] that balances fast and slow reasoning by incorporating adaptive length-based rewards and dynamic regularization conditioned on the characteristics of multimodal inputs. Our approach begins with an investigation of the relationship between reasoning length and accuracy in LVLMs, empirically demonstrating how length rewards and data distributions impact reasoning performance. Based on these findings, our methodology first introduces two complementary metrics to estimate the difficulty of the questions, guiding the model to determine when fast or slow thinking is more appropriate. Next, we incorporate adaptive length-based rewards and difficulty-aware KL divergence into the GRPO algorithm. The former dynamically incentivizes concise or detailed reasoning based on question characteristics, while the latter modulates exploration constraints based on the estimated difficulty of each question.

We conduct extensive experiments on a range of reasoning benchmarks for LVLMs, and the experimental results have demonstrated the effect of the proposed method. As shown in Figure  1, compared with slow thinking or fast thinking methods, our model achieves state-of-the-art reasoning accuracy with an average accuracy improvement of over 10% compared to the base model, while significantly reducing reasoning length against slow thinking models from 32.7% to 67.3%.

2 Background: Group Relative Policy Optimization↩︎

Group Relative Policy Optimization (GRPO;[5], [7]) extends PPO [29] by replacing the value model with group relative rewards estimation, optimizing the objective in Equation 1 . \[\begin{align} \mathcal{J}_{GRPO}(\pi_\theta) &= \mathbb{E}_{q \sim P(Q), \{o_i\}_{i=1}^G \sim \pi_{\theta_{old}}(\cdot|q)} \Bigg[\frac{1}{G}\sum_{i=1}^G \frac{1}{|o_i|} \sum_{t=1}^{|o_i|} \left\{ \min \Bigg[ \frac{\pi_\theta(o_{i,t}|q,o_{i,<t})}{\pi_{\theta_{old}}(o_{i,t}|q,o_{i,<t})} \hat{A}_{i,t}, \right. \\ &\!\!\!\!\left. \text{clip} \left( \frac{\pi_\theta(o_{i,t}|q,o_{i,<t})}{\pi_{\theta_{old}}(o_{i,t}|q,o_{i,<t})}, 1-\epsilon, 1+\epsilon \right) \hat{A}_{i,t} \Bigg] \right\} - \beta D_{KL}\bigl(\pi_\theta \,\|\, \pi_{\text{ref}}\bigr) \Bigg] \end{align} \label{eq:GRPO-obj}\tag{1}\] where \(\varepsilon\) and \(\beta\) are the clipping hyperparameter and the coefficient controlling the KL regularization [5]. \(\hat{A}_{i,t}\) is the advantage, estimated through group relative rewards \(\hat{A}_{i,t} = \frac{r_i - \operatorname{mean}(\{r_1, r_2, \dots, r_G\})}{\operatorname{std}(\{r_1, r_2, \dots, r_G\})}\) with two rule-based rewards: (1) accuracy reward (\(r_a\)) gives a reward when the response is equivalent to the answer, and (2) format reward (\(r_f\)) ensures responses adhere to the specified format.

3 Pilot Experiments↩︎

As discussed in §1, when applying reinforcement learning—typically through GRPO—to LVLM reasoning struggles to effectively scale reasoning length (RL-only methods; [15], [17], [19]) or generates verbose outputs across all tasks with only marginal gains in accuracy (SFT-RL methods; [9], [13], [14], [21]).

Figure 2: Effect of length rewards on reasoning length and accuracy.

To better understand the factors affecting response length and overall performance in GRPO [5] for LVLM reasoning, we conduct a series of experiments on the Geometry 3K dataset [22]. In particular, we analyze the impact of length-based reward strategies (§3.1) and the influence of data distribution characteristics (§3.2).

3.1 Length Rewards Analysis↩︎

Prior research has established that while GRPO effectively scales response length in text-only LLMs [5], [18], [30], this effect does not transfer to LVLMs [17], [20]. To verify this phenomenon and explore potential solutions, we performed GRPO-Zero on Qwen2.5-VL [31] with rule-based accuracy reward, and tested explicit length rewards that either encourage longer correct responses (\(r_\text{lengthy reward} = L_{correct}/L_{max}\)) or shorter correct ones (\(r_\text{short reward} = 1- L_{correct}/L_{max}\)) as extended rewards, where \(L_{max}\) is the maximum token length, \(L_{correct}\) denotes the length of the correct response. As shown in Figure 2, with increasing training steps, GRPO with lengthy reward steadily increases to 700 tokens, GRPO with short reward decreases to 180 tokens, while Naive GRPO remains stable around 330 tokens. These length rewards successfully manipulated response length, producing variations from 180 to 700 tokens, but with only modest changes in accuracy (±3%). This decoupling between length and accuracy suggests that LVLMs can maintain reasoning performance across different response lengths, challenging the assumption that longer reasoning is always better.

Based on the findings above, we can draw the following conclusions:

Observation : LVLMs can produce significantly different reasoning lengths with modest changes in accuracy via length rewards, suggesting potential for balancing reasoning depth and performance.

3.2 Data Distribution Analysis↩︎

Figure 3: Effect of data distribution, especially difficulty on reasoning length and accuracy.

Considering that overthinking models tend to generate verbose reasoning responses regardless of question difficulty, we next investigated how data distribution—particularly the presence of samples with varying difficulty levels—might naturally influence reasoning length and performance.

To this end, we stratified the Geometry3K training dataset into three difficulty tiers using the pass@8 metric (i.e., the probability of correctly solving a question within eight attempts): Easy (\(0.75 \leq pass@8\)), Medium (\(0.25 < pass@8 < 0.75\)), and Hard (\(pass@8 \leq 0.25\)). This categorization resulted in approximately 35% Easy, 25% Medium, and 40% Hard samples.

Figure 3 illustrates how training on these different difficulty distributions affected model behavior. Models trained exclusively on Hard samples generated significantly longer responses but showed only marginal accuracy improvements. In contrast, training on Easy samples produced shorter responses while improving accuracy. Models trained on Medium samples showed modest length increase and the highest accuracy. Based on the findings above, we can draw the following conclusions:

Observation : Question difficulty acts as an implicit regulator of reasoning length, suggesting that data distribution can be strategically leveraged to achieve adaptive fast-slow thinking.

4 Fast-Slow Thinking GRPO↩︎

Building upon the aforementioned observations, we begin by introducing two complementary metrics to quantify the difficulty of multimodal questions, which facilitates dynamic data selection during reinforcement learning(§4.1). Subsequently, we introduce FAST-GRPO, a variant of GRPO specifically designed to balance fast and slow reasoning by leveraging adaptive length-based rewards conditioned on question difficulty(§4.2).

4.1 Multimodal Question Difficulty Estimation↩︎

Given our findings that data distribution influences both reasoning length and performance, accurately gauging data difficulty becomes essential for making dynamic adjustments to data distribution. To address this need, we propose two complementary metrics to measure the difficulty of a given question for the policy model: one directly evaluates the intrinsic difficulty of the multimodal question itself, while the other measures its difficulty relative to the policy model’s current capabilities.

Extrinsic Difficulty. We first quantify question difficulty relative to the policy model through the empirical success rate \(S_{\mathrm{extrinsic}} = 1 - \mathrm{pass@k}\) , where \(\text{pass@k} = c/k\) represents the proportion of correct solutions among \(k\) rollouts. This metric is computed online, reflecting the model’s evolving capabilities.

Intrinsic Difficulty. While extrinsic difficulty reflects a model’s ability to solve problems, it may not fully capture the inherent visual complexity of the questions. We therefore introduce image complexity as an indicator that specifically evaluates the intrinsic difficulty within questions. Specifically, we use the Gray-Level Co-occurrence Matrix (GLCM) score, which analyzes how frequently pairs of pixels with specific intensity values occur at defined spatial relationships [32], [33]. However, GLCM captures only low-level image complexity based on pixel-level interactions and fails to account for higher-level semantic information. We therefore additionally employ the ViT classification entropy based on the output of the feature layer [34], [35] for image semantics complexity, providing a high-level representation of conceptual difficulty. The image complexity is computed as follows: \[H_{image} = - \frac{1}{P} \sum_{p=1}^{P} H(g_p) - H(v)\]

where \(g_{p}\) represents the GLCM for image patch \(p\), \(H(g_p)\) is the entropy of the co-occurrence probabilities across multiple radii and orientations; \(v\) denotes the feature output from the final layer of the ViT classifier, and \(H(v) = -\sum_{j}^{N} p_{j} \log p_{j}\) is the entropy of the predicted probability distribution over \(N\) classes, with \(p_{j}\) being the probability for class \(j\).

We integrate these metrics to get a comprehensive difficulty estimation (\(S_{\mathrm{difficulty}}\)) of a multimodal question. Questions with higher difficulty scores typically correspond to greater visual complexity and greater extrinsic difficulty for the policy model. \[S_{\mathrm{difficulty}} = S_{\mathrm{extrinsic}}\cdot{H_{image}}\]

Slow-to-Fast Sampling. We adopt curriculum strategies for the fast–slow thinking paradigm, using online-computed difficulty metrics. Two variants are considered: (i) Binary: distinct training phases. In early epochs, exclude easy samples (\(S_{\mathrm{extrinsic}} \leq 0.25\)) to strengthen reasoning on hard questions; in later epochs, exclude hard samples (\(S_{\mathrm{extrinsic}} \geq 0.75\)) to practice concise reasoning. (ii) Continuous: smoothly shift sampling probability from harder to easier questions over epochs, enabling a gradual transition from slow to fast thinking. Binary enforces a clear capability-efficiency separation, while Continuous offers a gentler progression.

4.2 FAST-GRPO↩︎

With the carefully designed difficulty estimation methods for multimodal question, we introduce FAST-GRPO, a tailored variant of GRPO that balances fast and slow reasoning by incorporating adaptive length-based rewards conditioned on question difficulty, as illustrated in Algorithm 4.

Figure 4: FAST-GRPO Training

Difficulty-Aware Length Reward Shaping. In addition to the accuracy reward \(r_a\) and format reward \(r_f\), we propose a difficulty-aware length reward \(r_t\) as follows, which guides the model to employ the appropriate reasoning approach based on question difficulty. \[{!}{ r_t = \begin{cases} 1-\frac{L}{L_{avg}} & \text{if } (S_{\text{d}} < \theta) \land (r_a = 1)\\ min(\frac{L}{L_{avg}} - 1, 1) & \text{if } (\theta \leq S_{\text{d}}) \land (r_a = 0) \\ 0 & \text{otherwise} \end{cases} }\] where \(S_{\text{d}}\) is the difficulty score, \(\theta\) is the 80th percentile difficulty threshold across the batch, and \(L_{avg}\) is the average length computed in the batch of responses. For less complex questions (\(S_{\text{d}} < \theta\)), the reward encourages fast thinking for correct trajectories, specifically rewards trajectories towards shorter than average length. Conversely, for complex questions, the reward encourages thorough reasoning for incorrect trajectories. Importantly, this reward is capped at 1, preventing excessive verbosity even for complex problems. Difficulty threshold \(\theta\) is a hyperparameter, for which we analyse sensitivity in  §5.4.

Following Deepseek-R1 setting [7], [17], we define the final reward function as a linear combination of these components: \(r_i = r_a + \lambda_f r_f + \lambda_t r_t\). This difficulty-aware length reward necessitates encouraging exploration for complex problems while maintaining efficient, accurate responses for simpler ones.

Difficulty-Aware KL Regularization. In addition to the aforementioned length reward that encourages adaptive response length for questions of varying difficulty, the KL divergence term constrains the policy model’s deviation from the reference model to achieve an exploitation-exploration balance, which impacts learning effectiveness across questions of different difficulty levels  [5], [29]. Our KL coefficient sensitivity analysis in  § 5.4 also reveals that no single static \(\beta\) value optimally serves questions across difficulty levels. Lower KL constraints benefit challenging questions by enabling broader exploration, while stronger regularization maintains performance on simpler tasks. To address this issue, we implement a dynamic coefficient \(\beta_{d}\) for difficulty-aware regularization. \[\beta_d = \beta_{\min} + (\beta_{\max} - \beta_{\min}) \cdot (1 - S_{\text{extrinsic}})\]

We give a simple theoretical analysis to demonstrate how difficulty-aware \(\beta_d\) enhances learning on varying questions by decomposing the gradient coefficient from the gradient Equation 2 : \[\begin{align} \nabla_{\theta}\mathcal{J}_{GRPO}(\theta) = \mathbb{E}_{[q \sim P(Q), \{o_i\}_{i=1}^G \sim \pi_{\theta_{old}}(O|q)]} \Bigg[ \frac{1}{G}\sum_{i=1}^G\frac{1}{|o_i|} \sum_{t=1}^{|o_i|} \left[GC_{GRPO}(q, o)\right] \nabla_{\theta}\log \pi_\theta(o_{i,t} | q, o_{i,<t}) \Bigg] \end{align} \label{eq:32gradient32of32GRPO}\tag{2}\] \[GC_{GRPO}(q, o) = \underbrace{\hat{A}_{i}}_\text{Advantage Signal} + \underbrace{\beta_d \left(\frac{\pi_{\text{ref}}(o_{i}|q)}{\pi_{\theta}(o_{i}|q)} -1\right)}_\text{Adaptive KL Regularization} \label{eq:gc-fast}\tag{3}\]

As shown in Equation 3 , the gradient coefficient consists of the advantage signal driving policy improvement and the adaptive KL regularization term. For high-difficulty questions, \(\beta_d\) approaches \(\beta_{\min}\), weakening the KL regularization and allowing the policy update to be dominated by the advantage signal. For low-difficulty questions, \(\beta_d\) approaches \(\beta_{\max}\), restricting policy deviation to ensure stability. Besides, the length normalization term \(\frac{1}{|o_i|}\) explicitly affects gradient updates, providing theory insight into Observation  [obs:data_distribution]: for incorrect responses, it encourages longer outputs by reducing per-token penalties, while for correct responses, it encourages brevity through stronger per-token updates. This creates an implicit bias toward increasing response length for difficult questions where models generate more incorrect rollouts, while naturally promoting shorter responses for simpler questions that yield more correct solutions.

5 Experiments↩︎

In this section, we evaluate the efficacy of our method for LVLM reasoning.

5.1 Experimental Setup↩︎

r0.5

4pt

Table 2: Comparison of different training methods and training samples.
Method Training Stage
2-5 SFT Sample RL Sample
Virgo [36] 5K
Mulberry [37] 260K
LMM-R1 [15] 105K
MM-R1 [19] 6K
MM-Eureka [20] 56K
Curr-ReFT [16] 1.5K 9K
OpenVLThinker [9] 35K 15K
Vision-R1 [13] 200K 10K
R1-OneVision [14] 155K 10K
FAST (Ours) 18K

Training Dataset. Starting with 500K questions from LLaVA-CoT [21], Mulberry [37], and MathV-360K [38], we first apply filters for answer verifiability. We deduplicate questions, retain only rule-based verifiable answers [39], and standardize to closed-form questions (e.g., multiple-choice, numeric answers). Second, we apply Slow-to-Fast sampling to remove questions with extreme extrinsic difficulty scores (\(S_{\mathrm{extrinsic}}=0\) or \(1\)), yielding 18K training questions. We display its distribution in Figure  7 and the specific source in the appendix.

Evaluation Benchmarks. we evaluate on 7 widely used multimodal benchmarks: (1) MathVision [40], (2) MathVerse [41], (3) MathVista [42], (4) MM-Math [43], (5) WeMath [44], (6) DynaMath [45], and (7) MM-Vet [46]. The first six cover various mathematical reasoning tasks, while MM-Vet examines general multimodal abilities. We report both accuracy and response length for all benchmarks. We also conduct additional cross-domain evaluations, including science reasoning (MM-K12 [20]), open-domain VQA (Bingo [47], MMHAL [48]), low-level visual perception (MMVP [49]), and comprehensive calibrated evaluation (MMEval-Pro [50]). Details are provided in Appendix 19.

Baselines. For slow thinking reasoning, we compare with three categories of approaches: (1) SFT on distilled data (LLaVA-CoT, Mulberry, Virgo); (2) RL-only training (MM-Eureka, LMM-R1, MM-R1); and (3) Two-stage approaches combining SFT and RL (R1-OneVision, Curr-ReFT, OpenVLThinker, Vision-R1). A comparative analysis of training methodologies and samples across these baselines is presented in Table 2. For fast-slow thinking comparison, we evaluate against fast thinking methods using various reward shaping techniques: Kimi 1.5’s length penalty [23], cosine function rewards [24], and DAST [25]. Table 8 details these different reward formulations.

3.5pt

Table 3: Main results on reasoning benchmarks compared with slow-thinking methods. For each benchmark, we report both accuracy (acc.) and response length (len.). Tokens are counted with Qwen2.5-VL’s tokenizer.
Method MathVision MathVerse MathVista MM-Math WeMath DynaMath MM-Vet
2-3 (lr)4-5 (lr)6-7 (lr)8-9 (lr)10-11 (lr)12-13 (lr)14-15 Acc. Len. Acc. Len. Acc. Len. Acc. Len. Acc. Len. Acc. Len. Acc. Len.
Closed-Source Model
GPT-4o 30.4 49.9 63.8 31.8 69.0 63.7 80.8
Claude-3.5 Sonnet 37.9 46.3 67.7 64.8 68.7
Qwen-VL-Max 39.3 47.3 74.2 45.6 73.2
MM-Eureka 26.9 40.4 67.1 60.7
LLaVA-CoT 16.4 20.3 54.8 22.6 44.8 60.3
Base Qwen2-VL-7B
Qwen2-VL-7B 18.8 443.0 31.9 388.9 58.2 265.9 20.2 661.7 50.5 294.3 39.8 298.4 62.0 132.5
Mulberry 23.4 349.2 39.5 364.3 62.1 275.0 23.7 467.0 50.4 372.1 46.8 273.3 43.9 218.3
Virgo 24.0 36.7
Base Qwen2.5-VL-3B
Qwen2.5-VL-3B 21.2 450.6 34.6 362.3 62.3 212.9 33.1 627.9 50.4 323.7 48.2 270.9 61.3 138.8
Curr-ReFT 20.1 240.1 36.3 121.6 61.9 95.9 28.6 301.5 57.3 156.0 43.8 146.4 62.0 117.6
LMM-R1 25.2 447.8 41.8 423.9 63.2 245.0 36.5 634.5 62.9 382.5 53.1 341.6 65.9 166.3
FAST-3B (Ours) 26.8 323.5 43.0 286.3 66.2 158.7 39.4 425.0 63.1 244.9 54.4 213.7 64.0 112.7
Base Qwen2.5-VL-7B
Qwen2.5-VL-7B 25.6 443.0 46.9 388.9 68.2 189.1 34.1 666.7 61.0 294.3 58.0 273.3 67.1 132.5
MM-R1 30.2 324.6 49.8 283.9 71.0 185.6 41.9 528.5 67.9 235.7 57.5 254.2 70.6 137.9
Vision-R1 52.4 73.5 40.4
R1-OneVision 29.9 692.8 46.4 631.5 64.1 402.5 34.1 688.6 61.8 591.9 53.5 560.6 71.6 440.7
OpenVLThinker 29.6 457.2 47.9 398.4 70.2 305.7 33.1 549.7 64.5 326.7 57.4 382.1 68.5 312.7
FAST-7B (Ours) 30.6 204.8 50.6 201.0 73.8 120.7 44.3 335.6 68.8 170.3 58.3 164.8 71.2 114.1

5pt

Table 4: Main results of accuracy and length compared with fast-thinking reward shaping methods.
Method MathV MathVista MathVer. WeMath MM-Vet Avg.
2-3 (lr)4-5 (lr)6-7 (lr)8-9 (lr)10-11 (lr)12-13 Acc. Len. Acc. Len. Acc. Len. Acc. Len. Acc. Len. Acc. Len.
Kimi 25.9 78.9 71.1 58.1 48.2 105.8 66.2 75.3 67.1 57.1 55.7 75.0
CosFn 27.9 396.4 72.1 247.2 49.6 383.9 68.1 311.9 71.1 148.9 57.8 297.7
DAST 27.0 281.1 72.9 93.5 48.5 194.5 67.4 148.9 67.6 66.3 56.7 156.9
FAST 30.6 204.8 73.8 120.7 50.6 201.0 68.8 170.3 71.2 114.1 59.0 162.2

5.2 Main Results↩︎

We report the main results concerning reasoning accuracy and reasoning length.

Reasoning Accuracy. Table 4 reports the main results of reasoning performance. First, FAST achieves state-of-the-art results on MathVista with 73.8 and MathVerse with 50.6, outperforming leading-edge closed-source LVLMs like GPT-4o. Second, on more challenging benchmarks, MathVision and MM-Math, FAST achieves competitive results, validating FAST’s ability to solve complex questions. Third, FAST improves Qwen2.5-VL-7B, our base model, with an average accuracy improvement of over 10%. Fourth, FAST improves Qwen2.5-VL-3B with an average accuracy improvement of over 14%, demonstrating that our method can be applied to different-sized models.  Further scalability results on a 32B-parameter model are provided in Appendix 18. Lastly, FAST maintains its general multimodal ability, evidenced by improved performance on MM-Vet, and further demonstrates strong generalization beyond math-centric benchmarks in science reasoning and open-domain VQA. In these evaluations, FAST improves its base model by 7–9% in physics, chemistry, and biology, and on open-domain VQA matches or surpasses strong baselines, showing effectiveness across diverse reasoning domains. Detailed results are provided in Appendix 19.

Reasoning Length. Tables 3 and 4 report the main results of reasoning length. First, FAST achieves a significant reduction of average reasoning length compared to slow thinking methods, from 32.7% against MM-R1 to 67.3% versus R1-OneVision, while preserving comparable or better reasoning accuracy. Second, compared to other fast thinking methods in LLMs, FAST achieves a modest reasoning length reduction and better reasoning accuracy. Third, FAST achieves slower thinking on more challenging questions, producing 60% longer responses on Hard than Easy of Geometry 3K as shown in Table 1 and averaging 79% more tokens on MM-Math. Lastly, in cross-domain evaluations ( §19), FAST yields substantially shorter responses: on MM-K12, average length drops by \(\sim\)​106 tokens (\(33.8\%\)) vs.its base model and over \(30\%\) vs.strong slow-thinking baselines. In open-domain VQA (Bingo, MMHal), outputs are consistently \(15\)\(25\%\) shorter, showing effective control of reasoning length beyond math tasks.

5.3 Ablations↩︎

r0.5

1.5pt

Table 5: Ablation Results on MathVista, MathVision, and MathVerse. More details of naive GRPO refer to appendix § 14.
Model MathVista MathV. MathVer. Len.
Qwen-2.5-VL-7B 68.2 25.6 46.9 340.3
FAST 73.8 30.6 50.6 175.5
w/o Data Sampling 69.9 27.2 48.4 257.3
w/o Thinking Reward 73.6 31.5 45.9 302.2
w/o Difficulty Aware 72.0 29.5 49.2 171.6
Naive GRPO 67.2 25.3 47.6 205.4
+ early stop 70.4 28.1 48.9 243.6

We conduct ablation studies to validate the effectiveness of each design of our method: Data Sampling, thinking reward shaping, and difficulty-aware optimization. The results are represented in Table  5. We can draw the following conclusions. First, without Data Sampling, reasoning accuracy seriously degrades on all benchmarks, highlighting the critical role of proper data distribution. Second, our thinking reward significantly reduces relative 42% response length with minor reasoning accuracy degradation, from 31.5 to 30.6 on MathVision. Third, the difficulty-aware regularization demonstrates robust improvement across all benchmarks, with a 1.8-point absolute increase on MathVista.

5.4 Analyses↩︎

r7cm

1.5pt

Table 6: Results on the effect of Slow to Fast Sampling.
Method MathV. MathVista MathVer. Len.
No Selection 25.3 67.2 47.6 205.4
Dynamic Sampling 27.0 73.2 50.3 317.9
Fast to Slow 26.3 72.9 50.2 266.1
Slow to Fast 30.6 73.8 50.6 175.5

5.4.0.1 Effect of Slow to Fast Sampling.

We further investigate the effect of Slow to Fast sampling by comparing our Slow to Fast sampling with alternative approaches: Fast to Slow, i.e., excluding hard samples early, easy samples later, and Dynamic Sampling [51], i.e., always filtering out Easy and Hard samples). As shown in Table 6, Fast to Slow yields comparable accuracy but shows degradation on challenging MathVision, while Dynamic Sampling leads to 80% longer responses without better accuracy improvements. We also compared our binary Slow-to-Fast sampling against a continuous variant to examine the effect of gradual curriculum shifts. This additional comparison is reported in Appendix 17.

image
Correlation between image complexity metrics and human judgments.

Effect of SFT versus GRPO. As shown in Table  [Tab:32effect32of32SFT32versus32GRPO], to further verify the efficacy of our FAST compared to SFT methods, we compare our method with SFT using: (1) 260K structured CoT data from GPT-4o [37] and (2) 200K long CoT from Deepseek-R1 [13]. SFT on Deepseek-R1 data produces overthinking responses with degraded reasoning, while SFT on GPT-4o data mimics fixed structures without substantial gains. In contrast, FAST with just 18K samples demonstrates superior performance across all benchmarks.

Validation of Image Complexity. In our image complexity design, we utilize the GLCM entropy score [32], [33] to measure texture complexity and ViT classification entropy [34], [35] for semantic complexity. Zhang et al. [32] demonstrated that GLCM entropy achieves strong human alignment with a Spearman Rank-order Correlation Coefficient (SRCC) of 0.75 and a Pearson Linear Correlation Coefficient (PLCC) of 0.77. To validate the effectiveness of our combined metric \(H_{img}\) on our specific dataset, we followed the methodology in [32], having three participants rate 200 sampled training images on a 5-point scale based on visual detail complexity. As shown in Table [tab:image95complexity95validation], while our combined metric \(H_{img}\) demonstrates moderate correlation with human judgments (SRCC=0.49, PLCC=0.54), it maintains well alignment with human perception.

Difficulty Threshold Analysis. We use the 80th percentile of batch difficulty as our threshold \(\theta\). Figure 5 shows our grid search results: the 100th percentile yields concise responses (140.8 tokens) but reduces accuracy, while a 0 threshold produces excessive verbosity (486.2 tokens) with performance degradation. The 50th percentile achieves comparable accuracy to our 80th percentile but with 37% longer responses, confirming our choice effectively balances accuracy and conciseness.

Figure 5: Left: Results on the effect of difficulty threshold. The average accuracy is computed across MathVision, MathVerse, and MathVista. Middle: Test set results with different difficulty level training split comparisons on Geometry 3K. Right: Test set results with different \beta value comparison in pilot experiments on Geometry 3K. OOD results comparison on MM-Vet benchmark.

Extrinsic Difficulty Analysis. Figure  5 reveals how training on different extrinsic difficulty splits affects model performance. First, training on Medium difficulty samples yields the best overall performance (45.4%), providing optimal balance for learning. Second, we observe clear difficulty-specific transfer effects: Easy training improves Easy test performance (76.3%), Medium training benefits Medium tests (44.4%), and Hard training significantly boosts Hard test performance (15.1% vs. base 5.5%). However, Hard sample training degrades Easy performance (68.2% vs. base 72.7%), while Easy training shows limited transfer to Hard problems. These findings support our Slow-to-Fast sampling strategy, demonstrating that no single difficulty level is optimal for all test cases.

KL Coefficient Analysis. Recent works [51], [52] suggest that removing KL constraints can enhance long-form reasoning in language models. We explored this effect in visual reasoning through a grid search on the KL coefficient \(\beta\) (Figure  5). Our analysis reveals that lower \(\beta\) values significantly improve performance on Hard questions (16.9% at \(\beta=0\) vs. base 5.5%) by enabling greater exploration, but risk catastrophic forgetting on previously mastered tasks. Conversely, higher \(\beta\) values maintain strong performance on Easy questions and improve out-of-distribution generalization (69.2% at \(\beta=5e-2\) on MM-Vet), but restrict exploration on complex reasoning tasks. These findings demonstrate no static \(\beta\) value optimally serves questions across all difficulty levels—Hard questions benefit from looser constraints while Easy ones and generalization require stronger regularization.

5.4.0.2 In-depth Analysis of Multiplying Estimated Difficulties.

The multiplicative form \(S_{\mathrm{difficulty}} = S_{\mathrm{extrinsic}}\cdot H_{\mathrm{image}}\) jointly captures empirical hardness and intrinsic visual complexity. One theoretical concern is that this product could give low scores for cases that are hard for the model but visually simple, potentially leading to fast-thinking behaviour when slow reasoning is needed. In practice, such mismatches are rare (less than 5% of our training set), and our reward design in  §4.2 assigns zero length reward in these cases, avoiding contradictory signals. We also tested a weighted-sum alternative \(S_{\mathrm{difficulty}}^{(\mathrm{sum})} = \alpha S_{\mathrm{extrinsic}} + (1-\alpha) H_{\mathrm{image}},\) where \(\alpha=0.5,\) and found almost identical performance to the multiplicative form across MathVista, MathVision, and MathVerse (differences within 0.5% accuracy). These results confirm that FAST’s difficulty estimation is robust to this potential corner case and to the choice of combination strategy. Details are provided in  §16.

5.5 Case Studies and Failure Mode Analysis↩︎

We complement our quantitative results with qualitative illustrations of FAST’s behaviour. Appendix §20 provides examples where FAST adapts its reasoning, from concise answers on simple problems to expanded chains on complex ones, as well as typical failure cases. Here, we focus on a systematic analysis of these failures.

To understand when and why FAST-GRPO succeeds or fails, we analyse all incorrect responses from FAST-7B, R1-OneVision-7B, and the base model (Qwen2.5-VL-7B) on MathVista. We observe three recurring failure patterns (Figure 6): (i) Visual Perception Failures — where the model incorrectly extracts or interprets visual cues (e.g., scales, chart values, spatial relations); (ii) Reasoning Error Propagation — where a mid-chain mistake contaminates subsequent logical steps; and (iii) Knowledge Conflict & Gap — where language priors override contradictory visual evidence, or the model hallucinates in the absence of domain knowledge.

Figure 6: Error breakdown by category.

5.5.0.1 Key Insights.

First, adaptive fast–slow thinking substantially reduces reasoning-related failures: FAST-7B cuts Reasoning Error Propagation and Knowledge Conflict cases by \(\sim\)​27% and \(\sim\)​19% relative to its base model. Shorter, targeted reasoning chains leave fewer opportunities for mid-proof errors and help suppress hallucinations caused by overextended thought.

Second, perception, not reasoning, is the dominant bottleneck: over half of FAST-7B’s errors stem from visual misinterpretation. Once a spatial relation is mis-localised or a key numeric value misread, even perfectly structured reasoning will converge to an incorrect answer. Future gains will likely come from strengthening the input stage, e.g., fine-grained OCR, calibrated scale reading, robust chart and graph value extraction, and accurate spatial grounding, so adaptive reasoning can operate on correct evidence.

6 Related Work↩︎

We review approaches for LVLM reasoning and methods addressing overthinking in LLMs.

Slow-thinking methods for LVLMs. SFT-RL two-stage methods leverage high-quality reasoning trajectories while inadvertently behavior cloning overthinking. Examples include Mulberry [37], [53] using MCTS from GPT-4o, LLaVA-CoT [21] with structured reasoning stages, and Virgo [36] finetuning on text-only reasoning chains. Vision-R1 [13], R1-OneVision [14], and OpenVLThinker [9] first collect distilled data from advanced models before applying SFT and RL. RL-only methods directly employ RL to improve reasoning accuracy but struggle with scaling response length [19], [20], [54]. Visual-RFT [54] uses GRPO for various vision tasks, while MM-R1 [19], LMM-R1 [15], and MM-Eureka [20] apply RL on base models with curated visual reasoning questions.

Fast-Slow thinking methods for LLMs. Methods addressing overthinking in LLMs include inference-stage approaches include TALE [28] enforcing token budgets in prompt and CCoT [27] providing concise examples in context. Training-stage approaches include O1-Pruner [26] using a tailored RL objective to reduce verbosity, CoT-Value [55] fine-tuning on varied-length reasoning chains to learn dynamic thinking, and Kimi [23] proposing length penalty rewards in RL and Long2Short DPO [56] to shorten length. DAST [25] and CosineReward [24] encourage shorter correct responses and longer in- correct responses via curated length rewards. While effective for text-only tasks, these approaches remain largely unexplored for LVLM reasoning.

7 Conclusion↩︎

We presented FAST, a framework enabling LVLMs to dynamically adapt reasoning depth based on question characteristics, addressing the overthinking phenomenon. Through empirical analysis, we developed FAST-GRPO with three components: model-based metrics for question characterization, adaptive thinking rewards, and difficulty-aware KL regularization. Extensive experiments demonstrated that FAST achieves state-of-the-art accuracy with over 10% improvement compared to the base model while reducing token usage compared to previous slow-thinking approaches, effectively balancing reasoning length and accuracy.

8 Acknowledgement↩︎

This work was supported by the Earth System Big Data Platform of the School of Earth Sciences, Zhejiang University, and Alibaba-Zhejiang University Joint Research Institute of Frontier Technologies.

This Appendix for "Fast-Slow Thinking GRPO for Large Vision-Language Model Reasoning" is organized as follows:

  • Experimental Setup and Reproducibility. In §10 we detail the implementation settings;  §11 describes the training and evaluation datasets;  §12 compares different length reward shaping methods;  §13 provides the human evaluation prompt for image complexity.

  • Additional Experimental Analyses.  §14 analyses naive GRPO behaviors;  §15 reports statistical significance of main results;  §20 presents case studies and failure mode examples;  §19 gives cross-domain evaluation results (science reasoning, open-domain VQA, low-level visual perception, calibrated evaluation);  §18 shows scalability experiments on a 32B-parameter LVLM.

  • Discussion and Limitations.  §9 discusses potential limitations of FAST-GRPO and directions for future work.

9 Limitations↩︎

While our FAST framework demonstrates significant improvements in balancing reasoning length and accuracy, we acknowledge several limitations in our current work. Due to computational resource constraints, we were only able to evaluate our approach on models up to 32B parameters (Qwen2.5-VL-32B). The effectiveness of fast-slow thinking mechanisms may scale differently with larger models (e.g., models with 70B+ parameters), which could potentially exhibit different reasoning patterns and overthinking behaviors.

10 Implementation Details↩︎

r0.62

Table 7: Training Hyperparameters
Hyperparameter Value
Model Qwen2.5-VL
Epochs 10
Learning Rate 1e-6
Train Batch Size 512
Temperature 1.0
Rollout per Prompt 8
Prompt Max Length 4096
Generation Max Length 4096
Max KL Coefficient 0.03
Min KL Coefficient 0.001
Precision BF16
Max Pixels 1000000
\(\lambda_f\) 0.5
\(\lambda_t\) 0.5
Difficulty \(\begin{cases} \text{Easy} & \text{if }0.75 \leq \text{pass@k}\\ \text{Hard} & \text{if pass@k} \leq 0.25 \\ \text{Medium} & \text{otherwise} \\ \end{cases}\)
Difficulty Threshold 80th percentile

We implement FAST using Qwen2.5-VL-3B and 7B as our base models. Below we detail our training setup and hyperparameters.

General Training Hyperparameters. For FAST training, we use our 18K dataset with a learning rate of 1e-6, a batch size of 512. We set the maximum sequence length to 4096 for both prompts and generation, and apply BF16 precision throughout training. The training process runs for 10 epochs, requiring approximately 600 H100 GPU hours. We use the prompt: You FIRST think about the reasoning process as an internal monologue and then provide the final answer. The reasoning process MUST BE enclosed within \(<think>\) \(</think>\) tags. The final answer MUST BE put in \(<answer>\) \(</answer>\) tags.

Method-specific Training Hyperparameters. For our reinforcement learning approach, we employ a temperature of 1.0, 8 rollouts per question, and a KL coefficient ranging from 0.001 (min) to 0.03 (max). The reward weighting factors are set to 0.5. The difficulty threshold is set at the 80th percentile. For GLCM computation, following prior setting [32], \(g_{p}\) is derived from local patch \(p\) in the original image with 64 gray levels, defined by radius \(\delta=[1,2,3,4]\) and orientation \(\theta=[0^{\circ}, 45^{\circ}, 90^{\circ}, 135^{\circ}]\). In practice, we divide the gray image into local patches of size 64.

Computation Environment. All training experiments were conducted using H20 GPUs. Model inference in evaluations is performed using the vLLM framework [57], and our training implementation extends the VeRL codebase [58].

The complete set of hyperparameters is provided in Table 7. We commit to releasing all the code, data, and model checkpoints for experimental results reproducibility.

11 Datasets↩︎

Figure 7: Distribution of Training Dataset Sources by Category.

Our training dataset comprises samples from four main categories: (1) Mathematical problems, including data from MathV360K, Geometry3K, and other mathematical reasoning datasets; (2) Visual QA tasks, sourced from ShareGPT4V, Vizwiz, and additional visual question answering benchmarks; (3) Science problems from AI2D, ScienceQA, and other scientific reasoning datasets; and (4) Figure Understanding tasks from DocVQA, ChartQA, and other document and chart comprehension datasets. The distribution is balanced across these categories, with Mathematical problems constituting the largest portion, followed by Figure Understanding, Science, and Visual QA tasks.

12 Length Rewards↩︎

We provide a comparison of different length rewards in Talbe  8.

Table 8: Comparison of different length reward shaping methods.
Method Length Reward
Kimi Length Penalty [23] \(\begin{cases} 0.5 - \frac{\mathrm{len}(i) - \mathrm{min\_len}}{\mathrm{max\_len} - \mathrm{min\_len}} & \text{if correct} \\ \min(0, 0.5 - \frac{\mathrm{len}(i) - \mathrm{min\_len}}{\mathrm{max\_len} - \mathrm{min\_len}}) & \text{otherwise} \end{cases}\)
CosFn [24] \(\eta_{min} + \frac{1}{2}(\eta_{max} - \eta_{min})(1 + \cos(\frac{t\pi}{T}))\)
where \(t\) is generation length, \(T\) is maximum length
\(\eta_{min}/\eta_{max}\) are min/max rewards
For correct answers: \(\eta_{min}=r_0^c, \eta_{max}=r_L^c\)
For wrong answers: \(\eta_{min}=r_0^w, \eta_{max}=r_L^w\)
DAST [25] \(\begin{cases} \max(-0.5 \lambda + 0.5, 0.1) & \text{if correct} \\ \min(0.9 \lambda - 0.1, -0.1) & \text{if incorrect} \end{cases}\)
where \(\lambda = \frac{L_i-L_{budget}}{L_{budget}}\)
and \(L_{budget} = p\cdot L_{\overline{r}} + (1-p)\cdot L_{max}\), \(p = \frac{c}{N}\)
FAST \(r_t = \begin{cases} 1-\frac{L}{L_{avg}} & \text{if } S_{\text{difficulty}} < \theta \text{ and } r_a = 1\\ \min(\frac{L}{L_{avg}} - 1, 1) & \text{if } \theta \leq S_{\text{difficulty}} \text{ and } r_a = 0 \\ 0 & \text{otherwise} \end{cases}\)

13 Human Evaluation Prompt↩︎

Figure 8: image.

14 Naive GRPO Results↩︎

Figure 9: Training accuracy of Naive GRPO.

As shown in Figure 9, the training accuracy for naive GRPO continues to increase throughout the training process, similar to other methods like Dynamic Sampling and FAST. However, when we examine the validation accuracy in Figure [fig:validation-accuracy], we observe a concerning trend. The validation accuracy of naive GRPO initially increases, reaching a peak of 0.40 around step 150, but then begins to decline despite continued improvements in training accuracy.

This discrepancy between training and validation performance indicates a clear overfitting phenomenon in naive GRPO. Upon closer examination of the model outputs in later training stages (e.g., case in Figure  10 ), we observe an increasing inconsistency between the chain-of-thought reasoning and the final answers produced by the model. Specifically, the model develops a tendency to generate reasoning chains that do not logically support its final conclusions, suggesting that it is memorizing training examples rather than learning generalizable reasoning patterns.

This overfitting behavior explains why the validation accuracy deteriorates even as training accuracy continues to improve. The model optimizes for the training distribution at the expense of generalization capability. Therefore, implementing early stopping based on validation performance is crucial when using naive GRPO to select the optimal checkpoint (around step 150 in our experiments) before overfitting severely impacts the model’s reasoning abilities.

Figure 10: Example of reasoning inconsistency in naive GRPO on training dataset: The model produces the correct final answer (110°) despite having flawed reasoning in its CoT (which incorrectly concludes 70°). This suggests memorization rather than proper understanding of geometric principles.

15 Statistical Significance Analysis↩︎

5pt

Table 9: Main results on reasoning benchmarks. We report accuracy (Acc.) and response length (Len.) with standard deviation across 5 runs with temperature set to 0.1. * indicates statistical significance (p<0.05).
Benchmark Qwen2.5-VL-3B FAST-3B Qwen2.5-VL-7B FAST-7B
2-9 Acc. Len. Acc. Len. Acc. Len. Acc. Len.
MathVision 21.2 450.6 26.8±0.3\(^{*}\) 323.5±14.2\(^{*}\) 25.6 443.0 30.6±0.4\(^{*}\) 204.8±12.3\(^{*}\)
MathVerse 34.6 362.3 43.0±0.4\(^{*}\) 286.3±12.8\(^{*}\) 46.9 388.9 50.6±0.5\(^{*}\) 201.0±10.5\(^{*}\)
MathVista 62.3 212.9 66.2±0.5\(^{*}\) 158.7±9.3\(^{*}\) 68.2 189.1 73.8±0.6\(^{*}\) 120.7±8.2\(^{*}\)
MM-Math 33.1 627.9 39.4±0.6\(^{*}\) 425.0±16.7\(^{*}\) 34.1 666.7 44.3±0.7\(^{*}\) 335.6±15.3\(^{*}\)
WeMath 50.4 323.7 63.1±0.4\(^{*}\) 244.9±11.5\(^{*}\) 61.0 294.3 68.8±0.5\(^{*}\) 170.3±9.8\(^{*}\)
DynaMath 48.2 270.9 54.4±0.3\(^{*}\) 213.7±10.6\(^{*}\) 58.0 273.3 58.3±0.4 164.8±11.2\(^{*}\)
MM-Vet 61.3 138.8 64.0±0.5\(^{*}\) 112.7±6.9\(^{*}\) 67.1 132.5 71.2±0.6\(^{*}\) 114.1±7.5\(^{*}\)
Figure 11: Performance comparison between FAST-7B and Qwen2.5-VL-7B across multiple benchmarks.

To rigorously evaluate the effectiveness of our approach, we conducted statistical significance analysis across all benchmarks. Table 9 presents comprehensive results comparing our FAST models with their respective Qwen2.5-VL baselines, including standard deviations from multiple runs.

Figure 11 visualizes the performance differences between FAST-7B and Qwen2.5-VL-7B. The top panel illustrates accuracy improvements in percentage points, while the bottom panel shows response length reduction percentages. Error bars represent standard deviation across 5 runs with temperature set to 0.1, and asterisks (*) indicate statistically significant differences (p<0.05).

Our analysis reveals that FAST-7B achieves statistically significant accuracy improvements on 6 out of 7 benchmarks, with only DynaMath showing a non-significant improvement (0.3 percentage points). The most substantial accuracy gains are observed on mathematical reasoning tasks (MathVision: +5.0%, MathVerse: +3.7%, MM-Math: +10.2%), demonstrating our method’s particular effectiveness on complex reasoning problems.

Regarding response length, FAST-7B consistently produces significantly more concise responses across all benchmarks, with length reductions ranging from 13.9% to 53.8%. This confirms that our approach successfully achieves both improved accuracy and enhanced efficiency in generating responses. The statistical significance of these improvements provides strong evidence for the effectiveness of our FAST framework in enhancing both the reasoning capabilities and efficiency.

16 Analysis of Multiplicative Difficulty Formulation and Weighted-Sum Alternative↩︎

The multiplicative combination \[S_{\mathrm{difficulty}} = S_{\mathrm{extrinsic}} \cdot H_{\mathrm{image}}\] was designed to jointly capture a model’s empirical success rate and the intrinsic visual complexity of a question. One concern raised in review is that when \(S_{\mathrm{extrinsic}}\) is high (model finds the problem hard) but \(H_{\mathrm{image}}\) is very low (visually simple), the product may be close to zero, potentially signalling "fast thinking" in a case that is actually challenging.

Reward design avoids conflict↩︎

As shown in Algorithm (4), the difficulty-aware length reward \(r_t\) applies non-zero shaping only in two aligned cases:

  • Correct and Not Complex (\(S_{\mathrm{difficulty}} < \theta\)): encourage shorter responses.

  • Incorrect and Complex (\(S_{\mathrm{difficulty}} \geq \theta\)): encourage longer responses.

The misaligned case in question (Incorrect but Not Complex, or Correct but Complex) yields \(r_t = 0\), so no penalty or incorrect encouragement is applied.

Empirical rarity of corner cases↩︎

We ranked 1,000 random training samples and computed correlations between \(S_{\mathrm{extrinsic}}\), \(H_{\mathrm{image}}\), and \(S_{\mathrm{difficulty}}\). High \(S_{\mathrm{extrinsic}}\) combined with low \(H_{\mathrm{image}}\) was rare (\(<5\%\) of samples).

Weighted-sum alternative↩︎

We compared the multiplicative form with a weighted sum: \[S_{\mathrm{difficulty}}^{(\mathrm{sum})} = \alpha S_{\mathrm{extrinsic}} + (1-\alpha) H_{\mathrm{image}}, \quad \alpha=0.5.\] Table 10 shows near-identical results.

Table 10: Multiplicative vs weighted-sum difficulty formulation.
Formulation MathVision MathVista MathVerse Avg.Len.
Multiplicative 30.6 73.8 50.6 175.5
Weighted Sum 29.1 73.9 50.2 183.8

These results confirm (i) corner cases are rare in our actual training distribution, (ii) reward shaping avoids contradictory signals for such cases, and (iii) performance is robust to the choice of multiplicative vs sum combination.

17 Continuous Slow-to-Fast Sampling↩︎

In the main text (Section [sec:slow95to95fast95sampling95effect]), we compared our Slow-to-Fast sampling strategy against alternative approaches such as Fast-to-Slow and Dynamic Sampling [51]. Here, we further contrast a continuous variant of Slow-to-Fast scheduling: .

17.0.0.1 Binary Slow-to-Fast.

In this setting, the training curriculum makes a hard switch at the halfway point of total epochs: the first half samples only hard and medium questions, and the second half incorporates easy questions, following the procedure in Algorithm 4.

17.0.0.2 Continuous Slow-to-Fast.

Here, the probability of drawing an easy sample, \(p_{\mathit{easy}}\), increases linearly with the training epoch \(t\) from \(0\) at the start to a maximum \(P_{\max}\) at the final epoch: \[p_{\mathit{easy}}(t) = P_{\max} \cdot \frac{t}{T},\] where \(T\) is the total number of epochs. We set \(P_{\max} = 0.4\) following initial tuning, ensuring a gradual transition from hard/medium-focus to more balanced sampling.

17.0.0.3 Results.

Table 11 compares the two schedules under identical training settings on MathVista, MathVision, and MathVerse.

Table 11: Binary vs.Continuous Slow-to-Fast scheduling. Accuracy (%) / Avg.length (tokens).
Method MathVista MathVision MathVerse Avg.Len.
Binary 73.8 30.6 50.6 175.5
Continuous 74.4 30.9 51.0 221.2

17.0.0.4 Findings.

Continuous scheduling provides accuracy gains (e.g., +0.6pp on MathVista) and increases average output length by 26%, reducing efficiency. We hypothesize that the chosen \(P_{\max}\) was insufficient to sample a large enough proportion of easy questions in later epochs, limiting potential efficiency gains. Additional tuning or adaptive \(P_{\max}\) may yield more favourable trade-offs.

18 Scalability to Larger Models↩︎

To evaluate the scalability of FAST-GRPO beyond mid-sized LVLMs, we train and test the framework on the 32B-parameter model Qwen-2.5-VL-32B using the same 18K-question training set as in our main experiments. Due to compute constraints, training was stopped after 3 epochs (\(\sim\)​1,200 GPU hours), which likely results in a sub-optimal checkpoint. We compare FAST-32B to strong slow-thinking baselines including Vision-R1-32B and MM-Eureka-32B on six benchmarks, including MM-K12 [20], a 2,000-question scientific reasoning benchmark evenly covering math, physics, chemistry, and biology. Due to compute constraints, we stopped training after three epochs (1200 GPU hours), resulting in a likely sub-optimal checkpoint.

5pt

Table 12: Performance of FAST-GRPO on 32B models compared to baselines. Accuracy (%) / Avg length (tokens).
Model MathVision MathVista MathVerse WeMath MM-K12 MM-Vet Avg Acc / Len
Qwen-2.5-VL-32B 38.4 / 651 71.7 / 331 49.9 / 550 69.1 / 515 66.8 / 840 71.1 / 312 61.1 / 533.2
Vision-R1-32B 39.1 / 976 76.4 / 410 60.9 / 818 74.2 / 637 64.8 / 1039 72.2 / 384 64.6 / 710.6
MM-Eureka-32B 34.4 / 639 74.8 / 352 56.5 / 560 73.4 / 524 72.2 / 857 73.4 / 344 64.1 / 546.0
FAST-32B 37.2 / 531 75.4 / 268 57.6 / 430 74.4 / 420 68.4 / 629 72.6 / 254 64.3 / 422.1

18.0.0.1 Findings.

Despite shorter training, FAST-32B matches or slightly exceeds the accuracy of stronger slow-thinking baselines while using notably fewer tokens:

  • Versus Vision-R1-32B, average output length is reduced by \(\sim 40\%\) (422.1 vs. tokens) with comparable accuracy (64.3% vs.%).

  • Versus MM-Eureka-32B, length is reduced by \(\sim 22\%\) (422.1 vs. tokens) while slightly improving average accuracy (64.3% vs.%).

These results indicate that FAST-GRPO scales effectively to larger LVLMs, maintaining its accuracy-efficiency trade-off. We leave exploration on ultra-large (\(\geq 70\)B) LVLMs for future work.

19 Cross-Domain Evaluation↩︎

To validate FAST-GRPO beyond math-intensive benchmarks, we conducted additional experiments on science reasoning, open-domain VQA, hallucination analysis, and low-level visual perception. These evaluations were added in response to reviewer requests for broader task coverage.

19.1 MM-K12 Scientific Reasoning↩︎

The MM-K12 benchmark [20] consists of 2,000 multimodal reasoning questions evenly covering four domains: mathematics, physics, chemistry, and biology.

Table 13: Accuracy (%) and average output length (tokens) on MM-K12 across subjects.
Model Math Physics Chemistry Biology Avg Acc / Len
Qwen-2.5-VL-7B 58.4 45.4 56.4 54.0 53.6 / 477.6
FAST-7B 69.0 53.8 63.4 62.8 62.2 / 371.2
MM-Eureka-7B 71.2 56.2 65.2 65.2 64.5 / 537.8
OpenVLThinker-7B 63.0 53.8 60.6 65.0 60.6 / 561.0
R1-OneVision-7B 44.8 33.8 39.8 40.8 39.8 / 817.5
FAST-3B 56.0 50.6 56.2 57.6 55.1 / 318.1

19.1.0.1 Findings.

Compared to its base model, FAST-7B improves accuracy by \(+8.4\)pp in physics, \(+7.0\)pp in chemistry, and \(+8.8\)pp in biology, while reducing output length by \(\sim 33.8\%\). Even against strong slow-thinking models such as MM-Eureka-7B, accuracy remains comparable with \(\sim30\%\) fewer tokens.

19.2 Additional General Benchmarks↩︎

Table 14: Performance across diverse benchmarks.
Model Bingo \(\uparrow\) MMHALU \(\uparrow\) MMVP \(\uparrow\) MMEval-Pro \(\uparrow\) MM-K12 \(\uparrow\)
Qwen2.5-VL-7B 3.70 3.50 47.3 76.0 53.6
FAST-7B 3.72 3.40 47.0 75.0 62.2
Vision-R1-7B 3.62 3.10 44.0 72.2
MM-Eureka-7B 3.69 3.20 46.7 74.8 64.5
OpenVLThinker-7B 3.45 3.00 46.5 71.5 60.6

We further evaluate on four benchmarks covering open-domain VQA and visual robustness:

  • Bingo Score [47]: Open-domain VQA benchmark for hallucination analysis.

  • MMHal [48]: Hallucination and informativeness evaluation in open-domain VQA.

  • MMVP [49]: Low-level visual perception probing.

  • MMEval-Pro [50]: Calibrated multimodal benchmark spanning math, science, and general VQA.

19.2.0.1 Findings.

FAST matches or slightly outperforms strong slow-thinking baselines in open-domain VQA and hallucination-robustness benchmarks, while producing shorter outputs. This supports the observation that adaptive reasoning length mitigates hallucination risk in multimodal reasoning [59].

20 Case Study↩︎

Figure 12: Case studies of different models’ reasoning. Mistakes and overthinking are highlighted in red and green.

Figure 12 illustrates how FAST balances reasoning length and accuracy. For simple coordinate identification, R1-OneVision exhibits overthinking with 349 tokens output (highlighted in green), while FAST delivers a concise 59-token solution. For complex geometry, the base model makes a critical error in angle calculations, while R1-OneVision produces a correct but verbose 676-token solution. FAST demonstrates adaptive slow thinking with a more efficient and correct 375-token solution. validating our approach’s ability to adjust reasoning depth based on question complexity.

In addition to these efficiency-focused examples, we present three representative error cases illustrating the main failure categories discussed in Section 5.5: one each for Visual Perception Failure, Reasoning Error Propagation, and Knowledge Conflict & Gap. These cases are shown in Figures 1315 and provide visual, task-specific instances of how such errors manifest across different problem types.

Figure 13: Visual Perception Failure example: Model confuses the beaker’s maximum capacity label with the highest visible measurement marking, leading to an incorrect answer. This highlights the bottleneck in visual extraction accuracy.
Figure 14: Reasoning Error Propagation example: The model applies the correct trigonometric method but misaligns the computed value with the problem’s actual target, causing subsequent steps to be built on a wrong assumption.
Figure 15: Knowledge Conflict & Gap example: The model ignores the specific sequence number information in the visual diagram and instead applies an incorrect general TCP rule, leading to a wrong ACK number.

NeurIPS Paper Checklist↩︎

  1. Claims

  2. Question: Do the main claims made in the abstract and introduction accurately reflect the paper’s contributions and scope?

  3. Answer:

  4. Justification: The abstract and introduction accurately reflect the paper’s contributions, including the FAST framework for balancing fast-slow thinking in LVLMs, the empirical analysis of reasoning length and accuracy, and the three key components of the approach.

  5. Guidelines:

    • The answer NA means that the abstract and introduction do not include the claims made in the paper.

    • The abstract and/or introduction should clearly state the claims made, including the contributions made in the paper and important assumptions and limitations. A No or NA answer to this question will not be perceived well by the reviewers.

    • The claims made should match theoretical and experimental results, and reflect how much the results can be expected to generalize to other settings.

    • It is fine to include aspirational goals as motivation as long as it is clear that these goals are not attained by the paper.

  6. Limitations

  7. Question: Does the paper discuss the limitations of the work performed by the authors?

  8. Answer:

  9. Justification: The paper includes a dedicated Limitations section 9 that acknowledges computational resource constraints limited evaluation to models up to 7B parameters, and notes that effectiveness may scale differently with larger models (70B+) which could exhibit different reasoning patterns.

  10. Guidelines:

    • The answer NA means that the paper has no limitation while the answer No means that the paper has limitations, but those are not discussed in the paper.

    • The authors are encouraged to create a separate "Limitations" section in their paper.

    • The paper should point out any strong assumptions and how robust the results are to violations of these assumptions (e.g., independence assumptions, noiseless settings, model well-specification, asymptotic approximations only holding locally). The authors should reflect on how these assumptions might be violated in practice and what the implications would be.

    • The authors should reflect on the scope of the claims made, e.g., if the approach was only tested on a few datasets or with a few runs. In general, empirical results often depend on implicit assumptions, which should be articulated.

    • The authors should reflect on the factors that influence the performance of the approach. For example, a facial recognition algorithm may perform poorly when image resolution is low or images are taken in low lighting. Or a speech-to-text system might not be used reliably to provide closed captions for online lectures because it fails to handle technical jargon.

    • The authors should discuss the computational efficiency of the proposed algorithms and how they scale with dataset size.

    • If applicable, the authors should discuss possible limitations of their approach to address problems of privacy and fairness.

    • While the authors might fear that complete honesty about limitations might be used by reviewers as grounds for rejection, a worse outcome might be that reviewers discover limitations that aren’t acknowledged in the paper. The authors should use their best judgment and recognize that individual actions in favor of transparency play an important role in developing norms that preserve the integrity of the community. Reviewers will be specifically instructed to not penalize honesty concerning limitations.

  11. Theory assumptions and proofs

  12. Question: For each theoretical result, does the paper provide the full set of assumptions and a complete (and correct) proof?

  13. Answer:

  14. Justification: The paper is primarily empirical and does not include formal theoretical results requiring proofs.

  15. Guidelines:

    • The answer NA means that the paper does not include theoretical results.

    • All the theorems, formulas, and proofs in the paper should be numbered and cross-referenced.

    • All assumptions should be clearly stated or referenced in the statement of any theorems.

    • The proofs can either appear in the main paper or the supplemental material, but if they appear in the supplemental material, the authors are encouraged to provide a short proof sketch to provide intuition.

    • Inversely, any informal proof provided in the core of the paper should be complemented by formal proofs provided in appendix or supplemental material.

    • Theorems and Lemmas that the proof relies upon should be properly referenced.

  16. Experimental result reproducibility

  17. Question: Does the paper fully disclose all the information needed to reproduce the main experimental results of the paper to the extent that it affects the main claims and/or conclusions of the paper (regardless of whether the code and data are provided or not)?

  18. Answer:

  19. Justification: All the the implementation details can be found in § 5.1 and  10.

  20. Guidelines:

    • The answer NA means that the paper does not include experiments.

    • If the paper includes experiments, a No answer to this question will not be perceived well by the reviewers: Making the paper reproducible is important, regardless of whether the code and data are provided or not.

    • If the contribution is a dataset and/or model, the authors should describe the steps taken to make their results reproducible or verifiable.

    • Depending on the contribution, reproducibility can be accomplished in various ways. For example, if the contribution is a novel architecture, describing the architecture fully might suffice, or if the contribution is a specific model and empirical evaluation, it may be necessary to either make it possible for others to replicate the model with the same dataset, or provide access to the model. In general. releasing code and data is often one good way to accomplish this, but reproducibility can also be provided via detailed instructions for how to replicate the results, access to a hosted model (e.g., in the case of a large language model), releasing of a model checkpoint, or other means that are appropriate to the research performed.

    • While NeurIPS does not require releasing code, the conference does require all submissions to provide some reasonable avenue for reproducibility, which may depend on the nature of the contribution. For example

      1. If the contribution is primarily a new algorithm, the paper should make it clear how to reproduce that algorithm.

      2. If the contribution is primarily a new model architecture, the paper should describe the architecture clearly and fully.

      3. If the contribution is a new model (e.g., a large language model), then there should either be a way to access this model for reproducing the results or a way to reproduce the model (e.g., with an open-source dataset or instructions for how to construct the dataset).

      4. We recognize that reproducibility may be tricky in some cases, in which case authors are welcome to describe the particular way they provide for reproducibility. In the case of closed-source models, it may be that access to the model is limited in some way (e.g., to registered users), but it should be possible for other researchers to have some path to reproducing or verifying the results.

  21. Open access to data and code

  22. Question: Does the paper provide open access to the data and code, with sufficient instructions to faithfully reproduce the main experimental results, as described in supplemental material?

  23. Answer:

  24. Justification: The code and model checkpoint are stated to be open source. Data generation for the training dataset is described, and evaluation benchmarks are public.

  25. Guidelines:

    • The answer NA means that paper does not include experiments requiring code.

    • Please see the NeurIPS code and data submission guidelines (https://nips.cc/public/guides/CodeSubmissionPolicy) for more details.

    • While we encourage the release of code and data, we understand that this might not be possible, so “No” is an acceptable answer. Papers cannot be rejected simply for not including code, unless this is central to the contribution (e.g., for a new open-source benchmark).

    • The instructions should contain the exact command and environment needed to run to reproduce the results. See the NeurIPS code and data submission guidelines (https://nips.cc/public/guides/CodeSubmissionPolicy) for more details.

    • The authors should provide instructions on data access and preparation, including how to access the raw data, preprocessed data, intermediate data, and generated data, etc.

    • The authors should provide scripts to reproduce all experimental results for the new proposed method and baselines. If only a subset of experiments are reproducible, they should state which ones are omitted from the script and why.

    • At submission time, to preserve anonymity, the authors should release anonymized versions (if applicable).

    • Providing as much information as possible in supplemental material (appended to the paper) is recommended, but including URLs to data and code is permitted.

  26. Experimental setting/details

  27. Question: Does the paper specify all the training and test details (e.g., data splits, hyperparameters, how they were chosen, type of optimizer, etc.) necessary to understand the results?

  28. Answer:

  29. Justification: We list experimental setting in § 10 and § 5.1.

  30. Guidelines:

    • The answer NA means that the paper does not include experiments.

    • The experimental setting should be presented in the core of the paper to a level of detail that is necessary to appreciate the results and make sense of them.

    • The full details can be provided either with the code, in appendix, or as supplemental material.

  31. Experiment statistical significance

  32. Question: Does the paper report error bars suitably and correctly defined or other appropriate information about the statistical significance of the experiments?

  33. Answer:

  34. Justification: See Figure  11 and Table  9.

  35. Guidelines:

    • The answer NA means that the paper does not include experiments.

    • The authors should answer "Yes" if the results are accompanied by error bars, confidence intervals, or statistical significance tests, at least for the experiments that support the main claims of the paper.

    • The factors of variability that the error bars are capturing should be clearly stated (for example, train/test split, initialization, random drawing of some parameter, or overall run with given experimental conditions).

    • The method for calculating the error bars should be explained (closed form formula, call to a library function, bootstrap, etc.)

    • The assumptions made should be given (e.g., Normally distributed errors).

    • It should be clear whether the error bar is the standard deviation or the standard error of the mean.

    • It is OK to report 1-sigma error bars, but one should state it. The authors should preferably report a 2-sigma error bar than state that they have a 96% CI, if the hypothesis of Normality of errors is not verified.

    • For asymmetric distributions, the authors should be careful not to show in tables or figures symmetric error bars that would yield results that are out of range (e.g. negative error rates).

    • If error bars are reported in tables or plots, The authors should explain in the text how they were calculated and reference the corresponding figures or tables in the text.

  36. Experiments compute resources

  37. Question: For each experiment, does the paper provide sufficient information on the computer resources (type of compute workers, memory, time of execution) needed to reproduce the experiments?

  38. Answer:

  39. Justification: See §10. We specify that training requires approximately 600 H100 GPU hours, with a global batch size of 512 and 8 rollouts per question over 10 epochs.

  40. Guidelines:

    • The answer NA means that the paper does not include experiments.

    • The paper should indicate the type of compute workers CPU or GPU, internal cluster, or cloud provider, including relevant memory and storage.

    • The paper should provide the amount of compute required for each of the individual experimental runs as well as estimate the total compute.

    • The paper should disclose whether the full research project required more compute than the experiments reported in the paper (e.g., preliminary or failed experiments that didn’t make it into the paper).

  41. Code of ethics

  42. Question: Does the research conducted in the paper conform, in every respect, with the NeurIPS Code of Ethics https://neurips.cc/public/EthicsGuidelines?

  43. Answer:

  44. Justification: We promise that this paper conforms, in every respect, with the NeurIPS Code of Ethics.

  45. Guidelines:

    • The answer NA means that the authors have not reviewed the NeurIPS Code of Ethics.

    • If the authors answer No, they should explain the special circumstances that require a deviation from the Code of Ethics.

    • The authors should make sure to preserve anonymity (e.g., if there is a special consideration due to laws or regulations in their jurisdiction).

  46. Broader impacts

  47. Question: Does the paper discuss both potential positive societal impacts and negative societal impacts of the work performed?

  48. Answer:

  49. Justification: The paper does not explicitly discuss potential positive and negative societal impacts of the work. The research focuses on improving reasoning efficiency in vision-language models, which has general benefits for AI applications but could benefit from a discussion of broader implications.

  50. Guidelines:

    • The answer NA means that there is no societal impact of the work performed.

    • If the authors answer NA or No, they should explain why their work has no societal impact or why the paper does not address societal impact.

    • Examples of negative societal impacts include potential malicious or unintended uses (e.g., disinformation, generating fake profiles, surveillance), fairness considerations (e.g., deployment of technologies that could make decisions that unfairly impact specific groups), privacy considerations, and security considerations.

    • The conference expects that many papers will be foundational research and not tied to particular applications, let alone deployments. However, if there is a direct path to any negative applications, the authors should point it out. For example, it is legitimate to point out that an improvement in the quality of generative models could be used to generate deepfakes for disinformation. On the other hand, it is not needed to point out that a generic algorithm for optimizing neural networks could enable people to train models that generate Deepfakes faster.

    • The authors should consider possible harms that could arise when the technology is being used as intended and functioning correctly, harms that could arise when the technology is being used as intended but gives incorrect results, and harms following from (intentional or unintentional) misuse of the technology.

    • If there are negative societal impacts, the authors could also discuss possible mitigation strategies (e.g., gated release of models, providing defenses in addition to attacks, mechanisms for monitoring misuse, mechanisms to monitor how a system learns from feedback over time, improving the efficiency and accessibility of ML).

  51. Safeguards

  52. Question: Does the paper describe safeguards that have been put in place for responsible release of data or models that have a high risk for misuse (e.g., pretrained language models, image generators, or scraped datasets)?

  53. Answer:

  54. Justification: The paper focuses on improving reasoning capabilities rather than releasing models with high risk for misuse. The proposed method enhances existing models rather than creating new potentially harmful capabilities.

  55. Guidelines:

    • The answer NA means that the paper poses no such risks.

    • Released models that have a high risk for misuse or dual-use should be released with necessary safeguards to allow for controlled use of the model, for example by requiring that users adhere to usage guidelines or restrictions to access the model or implementing safety filters.

    • Datasets that have been scraped from the Internet could pose safety risks. The authors should describe how they avoided releasing unsafe images.

    • We recognize that providing effective safeguards is challenging, and many papers do not require this, but we encourage authors to take this into account and make a best faith effort.

  56. Licenses for existing assets

  57. Question: Are the creators or original owners of assets (e.g., code, data, models), used in the paper, properly credited and are the license and terms of use explicitly mentioned and properly respected?

  58. Answer:

  59. Justification: The paper properly cites and credits the original sources of datasets (LLaVA-CoT, Mulberry, MathV-360K) and models (Qwen2.5-VL) used in the experiments, as well as the evaluation benchmarks.

  60. Guidelines:

    • The answer NA means that the paper does not use existing assets.

    • The authors should cite the original paper that produced the code package or dataset.

    • The authors should state which version of the asset is used and, if possible, include a URL.

    • The name of the license (e.g., CC-BY 4.0) should be included for each asset.

    • For scraped data from a particular source (e.g., website), the copyright and terms of service of that source should be provided.

    • If assets are released, the license, copyright information, and terms of use in the package should be provided. For popular datasets, paperswithcode.com/datasets has curated licenses for some datasets. Their licensing guide can help determine the license of a dataset.

    • For existing datasets that are re-packaged, both the original license and the license of the derived asset (if it has changed) should be provided.

    • If this information is not available online, the authors are encouraged to reach out to the asset’s creators.

  61. New assets

  62. Question: Are new assets introduced in the paper well documented and is the documentation provided alongside the assets?

  63. Answer:

  64. Justification: The paper describes the training data generation process and mentions releasing model checkpoints as well as code, with sufficient documentation of the methodology to understand the assets.

  65. Guidelines:

    • The answer NA means that the paper does not release new assets.

    • Researchers should communicate the details of the dataset/code/model as part of their submissions via structured templates. This includes details about training, license, limitations, etc.

    • The paper should discuss whether and how consent was obtained from people whose asset is used.

    • At submission time, remember to anonymize your assets (if applicable). You can either create an anonymized URL or include an anonymized zip file.

  66. Crowdsourcing and research with human subjects

  67. Question: For crowdsourcing experiments and research with human subjects, does the paper include the full text of instructions given to participants and screenshots, if applicable, as well as details about compensation (if any)?

  68. Answer:

  69. Justification: Human evaluation instruction is provided in §13.

  70. Guidelines:

    • The answer NA means that the paper does not involve crowdsourcing nor research with human subjects.

    • Including this information in the supplemental material is fine, but if the main contribution of the paper involves human subjects, then as much detail as possible should be included in the main paper.

    • According to the NeurIPS Code of Ethics, workers involved in data collection, curation, or other labor should be paid at least the minimum wage in the country of the data collector.

  71. Institutional review board (IRB) approvals or equivalent for research with human subjects

  72. Question: Does the paper describe potential risks incurred by study participants, whether such risks were disclosed to the subjects, and whether Institutional Review Board (IRB) approvals (or an equivalent approval/review based on the requirements of your country or institution) were obtained?

  73. Answer:

  74. Justification: The human evaluation conducted appears to be a minimal risk assessment of image complexity rather than research requiring IRB approval, as it doesn’t involve personal data or potential harm to participants.

  75. Guidelines:

    • The answer NA means that the paper does not involve crowdsourcing nor research with human subjects.

    • Depending on the country in which research is conducted, IRB approval (or equivalent) may be required for any human subjects research. If you obtained IRB approval, you should clearly state this in the paper.

    • We recognize that the procedures for this may vary significantly between institutions and locations, and we expect authors to adhere to the NeurIPS Code of Ethics and the guidelines for their institution.

    • For initial submissions, do not include any information that would break anonymity (if applicable), such as the institution conducting the review.

  76. Declaration of LLM usage

  77. Question: Does the paper describe the usage of LLMs if it is an important, original, or non-standard component of the core methods in this research? Note that if the LLM is used only for writing, editing, or formatting purposes and does not impact the core methodology, scientific rigorousness, or originality of the research, declaration is not required.

  78. Answer:

  79. Justification: The core method development in this research does not involve LLMs as any important, original, or non-standard components.

  80. Guidelines:

    • The answer NA means that the core method development in this research does not involve LLMs as any important, original, or non-standard components.

    • Please refer to our LLM policy (https://neurips.cc/Conferences/2025/LLM) for what should or should not be described.

References↩︎

[1]
Z.-Z. Li, D. Zhang, M.-L. Zhang, J. Zhang, Z. Liu, Y. Yao, H. Xu, J. Zheng, P.-J. Wang, X. Chen, Y. Zhang, F. Yin, J. Dong, Z. Guo, L. Song, and C.-L. Liu, “From system 1 to system 2: A survey of reasoning large language models,” 2025.
[2]
P. Hu, J. Qi, X. Li, H. Li, X. Wang, B. Quan, R. Wang, and Y. Zhou, “Tree-of-mixed-thought: Combining fast and slow thinking for multi-hop visual reasoning,” 2023.
[3]
W. Xiao, Z. Wang, L. Gan, S. Zhao, W. He, L. A. Tuan, L. Chen, H. Jiang, Z. Zhao, and F. Wu, “A comprehensive survey of datasets, theories, variants, and applications in direct preference optimization,” CoRR, vol. abs/2410.15595, 2024.
[4]
OpenAI, “Learning to reason with large language models,” September 2024.
[5]
D.-A. Team, “Deepseek-r1: Incentivizing reasoning capability in llms via reinforcement learning,” 2025.
[6]
Qwen, “Qwq: Reflect deeply on the boundaries of the unknown,” November 2024.
[7]
Z. Shao, P. Wang, Q. Zhu, R. Xu, J. Song, X. Bi, H. Zhang, M. Zhang, Y. K. Li, Y. Wu, and D. Guo, “Deepseekmath: Pushing the limits of mathematical reasoning in open language models,” 2024.
[8]
A. Yang, B. Yang, B. Hui, B. Zheng, B. Yu, C. Zhou, C. Li, C. Li, D. Liu, F. Huang, et al., “Qwen2 technical report,” arXiv preprint arXiv:2407.10671, 2024.
[9]
Y. Deng, H. Bansal, F. Yin, N. Peng, W. Wang, and K.-W. Chang, “Openvlthinker: An early exploration to complex vision-language reasoning via iterative self-improvement,” 2025.
[10]
G. Sun, M. Jin, Z. Wang, C.-L. Wang, S. Ma, Q. Wang, T. Geng, Y. N. Wu, Y. Zhang, and D. Liu, “Visual agents as fast and slow thinkers,” in The Thirteenth International Conference on Learning Representations, 2025.
[11]
Y. Wang, W. Chen, X. Han, X. Lin, H. Zhao, Y. Liu, B. Zhai, J. Yuan, Q. You, and H. Yang, “Exploring the reasoning abilities of multimodal large language models (mllms): A comprehensive survey on emerging trends in multimodal reasoning,” 2024.
[12]
W. Xiao, Z. Huang, L. Gan, W. He, H. Li, Z. Yu, F. Shu, H. Jiang, and L. Zhu, “Detecting and mitigating hallucination in large vision language models via fine-grained AI feedback,” in AAAI-25, Sponsored by the Association for the Advancement of Artificial Intelligence, February 25 - March 4, 2025, Philadelphia, PA, USA(T. Walsh, J. Shah, and Z. Kolter, eds.), pp. 25543–25551, AAAI Press, 2025.
[13]
W. Huang, B. Jia, Z. Zhai, S. Cao, Z. Ye, F. Zhao, Z. Xu, Y. Hu, and S. Lin, “Vision-r1: Incentivizing reasoning capability in multimodal large language models,” 2025.
[14]
Y. Yang, X. He, H. Pan, X. Jiang, Y. Deng, X. Yang, H. Lu, D. Yin, F. Rao, M. Zhu, B. Zhang, and W. Chen, “R1-onevision: Advancing generalized multimodal reasoning through cross-modal formalization,” arXiv preprint arXiv:2503.10615, 2025.
[15]
Y. Peng, G. Zhang, M. Zhang, Z. You, J. Liu, Q. Zhu, K. Yang, X. Xu, X. Geng, and X. Yang, “Lmm-r1: Empowering 3b lmms with strong reasoning abilities through two-stage rule-based rl,” arXiv preprint arXiv:2503.07536, 2025.
[16]
H. Deng, D. Zou, R. Ma, H. Luo, Y. Cao, and Y. Kang, “Boosting the generalization and reasoning of vision language models with curriculum reinforcement learning,” 2025.
[17]
L. Chen, L. Li, H. Zhao, Y. Song, and Vinci, “R1-v: Reinforcing super generalization ability in vision-language models with less than $3.” https://github.com/Deep-Agent/R1-V, 2025. Accessed: 2025-02-02.
[18]
H. Face, “Open r1: A fully open reproduction of deepseek-r1,” January 2025.
[19]
S. Leng, J. Wang, J. Li, H. Zhang, Z. Hu, B. Zhang, H. Zhang, Y. Jiang, X. Li, D. Zhao, F. Wang, Y. Rong, A. Sun, and S. Lu, “Mmr1: Advancing the frontiers of multimodal reasoning.” https://github.com/LengSicong/MMR1, 2025.
[20]
F. Meng, L. Du, Z. Liu, Z. Zhou, Q. Lu, D. Fu, B. Shi, W. Wang, J. He, K. Zhang, et al., “Mm-eureka: Exploring visual aha moment with rule-based large-scale reinforcement learning,” arXiv preprint arXiv:2503.07365, 2025.
[21]
G. Xu, P. Jin, H. Li, Y. Song, L. Sun, and L. Yuan, “Llava-cot: Let vision language models reason step-by-step,” 2024.
[22]
P. Lu, R. Gong, S. Jiang, L. Qiu, S. Huang, X. Liang, and S.-C. Zhu, Inter-GPS: Interpretable geometry problem solving with formal language and symbolic reasoning,” in Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)(C. Zong, F. Xia, W. Li, and R. Navigli, eds.), (Online), pp. 6774–6786, Association for Computational Linguistics, Aug. 2021.
[23]
K. Team, A. Du, B. Gao, B. Xing, C. Jiang, C. Chen, C. Li, C. Xiao, C. Du, C. Liao, et al., “Kimi k1.5: Scaling reinforcement learning with llms,” arXiv preprint arXiv:2501.12599, 2025.
[24]
E. Yeo, Y. Tong, M. Niu, G. Neubig, and X. Yue, “Demystifying long chain-of-thought reasoning in llms,” 2025.
[25]
Y. Shen, J. Zhang, J. Huang, S. Shi, W. Zhang, J. Yan, N. Wang, K. Wang, and S. Lian, “Dast: Difficulty-adaptive slow-thinking for large reasoning models,” 2025.
[26]
H. Luo, L. Shen, H. He, Y. Wang, S. Liu, W. Li, N. Tan, X. Cao, and D. Tao, “O1-pruner: Length-harmonizing fine-tuning for o1-like reasoning pruning,” 2025.
[27]
M. Renze and E. Guven, “The benefits of a concise chain of thought on problem-solving in large language models,” in 2024 2nd International Conference on Foundation and Large Language Models (FLLM), p. 476–483, IEEE, Nov. 2024.
[28]
T. Han, Z. Wang, C. Fang, S. Zhao, S. Ma, and Z. Chen, “Token-budget-aware llm reasoning,” 2025.
[29]
J. Schulman, F. Wolski, P. Dhariwal, A. Radford, and O. Klimov, “Proximal policy optimization algorithms,” 2017.
[30]
J. Hu, Y. Zhang, Q. Han, D. Jiang, X. Zhang, and H.-Y. Shum, “Open-reasoner-zero: An open source approach to scaling up reinforcement learning on the base model,” 2025.
[31]
S. Bai, K. Chen, X. Liu, J. Wang, W. Ge, S. Song, K. Dang, P. Wang, S. Wang, J. Tang, H. Zhong, Y. Zhu, M. Yang, Z. Li, J. Wan, P. Wang, W. Ding, Z. Fu, Y. Xu, J. Ye, X. Zhang, T. Xie, Z. Cheng, H. Zhang, Z. Yang, H. Xu, and J. Lin, “Qwen2.5-vl technical report,” arXiv preprint arXiv:2502.13923, 2025.
[32]
J. Zhang, Q. Huang, J. Liu, X. Guo, and D. Huang, “Diffusion-4k: Ultra-high-resolution image synthesis with latent diffusion models,” 2025.
[33]
R. M. Haralick, K. Shanmugam, and I. Dinstein, “Textural features for image classification,” IEEE Transactions on Systems, Man, and Cybernetics, vol. SMC-3, no. 6, pp. 610–621, 1973.
[34]
B. Wu, C. Xu, X. Dai, A. Wan, P. Zhang, Z. Yan, M. Tomizuka, J. Gonzalez, K. Keutzer, and P. Vajda, “Visual transformers: Token-based image representation and processing for computer vision,” 2020.
[35]
J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei, “Imagenet: A large-scale hierarchical image database,” in 2009 IEEE conference on computer vision and pattern recognition, pp. 248–255, Ieee, 2009.
[36]
Y. Du, Z. Liu, Y. Li, W. X. Zhao, Y. Huo, B. Wang, W. Chen, Z. Liu, Z. Wang, and J.-R. Wen, “Virgo: A preliminary exploration on reproducing o1-like mllm,” 2025.
[37]
H. Yao, J. Huang, W. Wu, J. Zhang, Y. Wang, S. Liu, Y. Wang, Y. Song, H. Feng, L. Shen, and D. Tao, “Mulberry: Empowering mllm with o1-like reasoning and reflection via collective monte carlo tree search,” 2024.
[38]
W. Shi, Z. Hu, Y. Bin, J. Liu, Y. Yang, S.-K. Ng, L. Bing, and R. K.-W. Lee, “Math-LLaVA: Bootstrapping mathematical reasoning for multimodal large language models,” in Findings of the Association for Computational Linguistics: EMNLP 2024(Y. Al-Onaizan, M. Bansal, and Y.-N. Chen, eds.), (Miami, Florida, USA), pp. 4663–4680, Association for Computational Linguistics, Nov. 2024.
[39]
hiyouga, “Mathruler.” https://github.com/hiyouga/MathRuler, 2025.
[40]
K. Wang, J. Pan, W. Shi, Z. Lu, H. Ren, A. Zhou, M. Zhan, and H. Li, “Measuring multimodal mathematical reasoning with math-vision dataset,” in The Thirty-eight Conference on Neural Information Processing Systems Datasets and Benchmarks Track, 2024.
[41]
R. Zhang, D. Jiang, Y. Zhang, H. Lin, Z. Guo, P. Qiu, A. Zhou, P. Lu, K.-W. Chang, Y. Qiao, P. Gao, and H. Li, “Mathverse: Does your multi-modal llm truly see the diagrams in visual math problems?,” in Computer Vision – ECCV 2024(A. Leonardis, E. Ricci, S. Roth, O. Russakovsky, T. Sattler, and G. Varol, eds.), (Cham), pp. 169–186, Springer Nature Switzerland, 2025.
[42]
P. Lu, H. Bansal, T. Xia, J. Liu, C. Li, H. Hajishirzi, H. Cheng, K.-W. Chang, M. Galley, and J. Gao, “Mathvista: Evaluating mathematical reasoning of foundation models in visual contexts,” in International Conference on Learning Representations (ICLR), 2024.
[43]
K. Sun, Y. Bai, J. Qi, L. Hou, and J. Li, MM-MATH: Advancing multimodal math evaluation with process evaluation and fine-grained classification,” in Findings of the Association for Computational Linguistics: EMNLP 2024(Y. Al-Onaizan, M. Bansal, and Y.-N. Chen, eds.), (Miami, Florida, USA), pp. 1358–1375, Association for Computational Linguistics, Nov. 2024.
[44]
R. Qiao, Q. Tan, G. Dong, M. Wu, C. Sun, X. Song, Z. Gongque, S. Lei, Z. Wei, M. Zhang, R. Qiao, Y. Zhang, X. Zong, Y. Xu, M. Diao, Z. Bao, C. Li, and H. Zhang, “We-math: Does your large multimodal model achieve human-like mathematical reasoning?,” CoRR, vol. abs/2407.01284, 2024.
[45]
C. Zou, X. Guo, R. Yang, J. Zhang, B. Hu, and H. Zhang, “Dynamath: A dynamic visual benchmark for evaluating mathematical reasoning robustness of vision language models,” 2024.
[46]
W. Yu, Z. Yang, L. Li, J. Wang, K. Lin, Z. Liu, X. Wang, and L. Wang, “Mm-vet: Evaluating large multimodal models for integrated capabilities,” in International conference on machine learning, PMLR, 2024.
[47]
C. Cui, Y. Zhou, X. Yang, S. Wu, L. Zhang, J. Zou, and H. Yao, “Holistic analysis of hallucination in gpt-4v (ision): Bias and interference challenges,” arXiv preprint arXiv:2311.03287, 2023.
[48]
Z. Sun, S. Shen, S. Cao, H. Liu, C. Li, Y. Shen, C. Gan, L. Gui, Y.-X. Wang, Y. Yang, et al., “Aligning large multimodal models with factually augmented rlhf,” in Findings of the Association for Computational Linguistics ACL 2024, pp. 13088–13110, 2024.
[49]
S. Tong, Z. Liu, Y. Zhai, Y. Ma, Y. LeCun, and S. Xie, “Eyes wide shut? exploring the visual shortcomings of multimodal llms,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9568–9578, 2024.
[50]
J. Huang, L. Chen, T. Guo, F. Zeng, Y. Zhao, B. Wu, Y. Yuan, H. Zhao, Z. Guo, Y. Zhang, et al., “Mmevalpro: Calibrating multimodal benchmarks towards trustworthy and efficient evaluation,” arXiv preprint arXiv:2407.00468, 2024.
[51]
G. Cui, L. Yuan, Z. Wang, H. Wang, W. Li, B. He, Y. Fan, T. Yu, Q. Xu, W. Chen, J. Yuan, H. Chen, K. Zhang, X. Lv, S. Wang, Y. Yao, X. Han, H. Peng, Y. Cheng, Z. Liu, M. Sun, B. Zhou, and N. Ding, “Process reinforcement through implicit rewards,” 2025.
[52]
Q. Yu, Z. Zhang, R. Zhu, Y. Yuan, X. Zuo, Y. Yue, T. Fan, G. Liu, L. Liu, X. Liu, H. Lin, Z. Lin, B. Ma, G. Sheng, Y. Tong, C. Zhang, M. Zhang, W. Zhang, H. Zhu, J. Zhu, J. Chen, J. Chen, C. Wang, H. Yu, W. Dai, Y. Song, X. Wei, H. Zhou, J. Liu, W.-Y. Ma, Y.-Q. Zhang, L. Yan, M. Qiao, Y. Wu, and M. Wang, “Dapo: An open-source llm reinforcement learning system at scale,” 2025.
[53]
J. Zhang, J. Huang, H. Yao, S. Liu, X. Zhang, S. Lu, and D. Tao, “R1-vl: Learning to reason with multimodal large language models via step-wise group relative policy optimization,” 2025.
[54]
Z. Liu, Z. Sun, Y. Zang, X. Dong, Y. Cao, H. Duan, D. Lin, and J. Wang, “Visual-rft: Visual reinforcement fine-tuning,” arXiv preprint arXiv:2503.01785, 2025.
[55]
X. Ma, G. Wan, R. Yu, G. Fang, and X. Wang, “Cot-valve: Length-compressible chain-of-thought tuning,” 2025.
[56]
W. Xiao, Z. Wang, L. Gan, S. Zhao, W. He, L. A. Tuan, L. Chen, H. Jiang, Z. Zhao, and F. Wu, “A comprehensive survey of direct preference optimization: Datasets, theories, variants, and applications,” 2024.
[57]
W. Kwon, Z. Li, S. Zhuang, Y. Sheng, L. Zheng, C. H. Yu, J. E. Gonzalez, H. Zhang, and I. Stoica, “Efficient memory management for large language model serving with pagedattention,” in Proceedings of the ACM SIGOPS 29th Symposium on Operating Systems Principles, 2023.
[58]
G. Sheng, C. Zhang, Z. Ye, X. Wu, W. Zhang, R. Zhang, Y. Peng, H. Lin, and C. Wu, “Hybridflow: A flexible and efficient rlhf framework,” arXiv preprint arXiv: 2409.19256, 2024.
[59]
C. Liu, Z. Xu, Q. Wei, J. Wu, J. Zou, X. E. Wang, Y. Zhou, and S. Liu, “More thinking, less seeing? assessing amplified hallucination in multimodal reasoning models,” arXiv preprint arXiv:2505.21523, 2025.

  1.  https://github.com/Mr-Loevan/FAST↩︎

  2. \(^{\dagger}\)Correspondence to Leilei Gan.↩︎