pith. machine review for the scientific record. sign in

arxiv: 2604.22266 · v1 · submitted 2026-04-24 · 💻 cs.CL

Recognition: unknown

Large Language Models Decide Early and Explain Later

Authors on Pith no claims yet

Pith reviewed 2026-05-08 12:02 UTC · model grok-4.3

classification 💻 cs.CL
keywords large language modelschain-of-thought reasoningearly stoppinganswer stabilizationinference efficiencypost-decision explanationforced answer completion
0
0 comments X

The pith

Large language models often fix their final answer early during chain-of-thought reasoning, making most later tokens post-decision explanation.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper examines when large language models reach their final answer while generating chain-of-thought explanations. By checking what the model would answer after each partial reasoning step, it finds that answers change in only about a third of cases. The remaining reasoning steps often just elaborate on an already fixed decision. This suggests that much of the generated text is unnecessary for correctness but adds to computation time. Simple stopping rules can cut hundreds of tokens with little accuracy loss.

Core claim

The authors use forced answer completion to probe the model's predicted answer at various points in the reasoning process. For the Qwen3-4B model across multiple datasets, predicted answers change in only 32% of queries. After the last change, an average of 760 additional reasoning tokens are generated, representing a substantial portion of the total output. Early stopping heuristics, such as probe-based stopping, reduce token usage by 500 per query with a 2% accuracy drop.

What carries the argument

Forced answer completion, which elicits the model's intermediate predictions by appending a prompt for the final answer at partial reasoning prefixes. It tracks when the answer stabilizes.

If this is right

  • A substantial portion of chain-of-thought generation consists of post-decision explanation rather than active decision making.
  • Simple heuristics for early stopping can reduce token usage by hundreds per query.
  • Accuracy remains nearly the same when generation stops after the answer has stabilized.
  • Inference latency and cost can be lowered by halting redundant reasoning steps.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • Models may benefit from training that encourages shorter reasoning paths from the start.
  • The early-fixation pattern could appear in other tasks or larger models beyond those tested.
  • Early stopping could combine with existing efficiency methods to further cut inference costs.
  • The results raise the possibility that chain-of-thought serves partly as formatting rather than discovery.

Load-bearing premise

Forcing the model to output an answer at an intermediate reasoning prefix accurately reveals the point at which its final decision is fixed, without the forcing itself altering the decision process.

What would settle it

Comparing answer changes and final accuracy when using forced probes versus uninterrupted generation on identical queries to check whether the forcing step itself shifts the model's decision.

Figures

Figures reproduced from arXiv: 2604.22266 by Alexander Mehler, Ayan Datta, Bhuvanesh Verma, Mounika Marreddy, Radhika Mamidi, Zhixue Zhao.

Figure 1
Figure 1. Figure 1: Top: illustration of an obtained answer trajectory for a multiple-choice task. A brief transient flip (answer C → answer B → answer C) reflects local instability, whereas the later sustained transition to D constitutes a genuine answer switch. We study this in Section 4. Bottom: Early-stopping to save token usages with (detailed figures are in Section 5) traces are typically treated as a monolithic process… view at source ↗
Figure 2
Figure 2. Figure 2: Distribution of answer switch counts across denoised answer trajectories for different task types using Qwen3-4B. 4.1.2. ANSWER TRAJECTORIES WITH NO ANSWER SWITCHES Next, we analyze the fraction of examples in which the predicted answer does not change at any reasoning step, i.e., the answer trajectory satisfies A0 = A1 = · · · = An or t ⋆ = −1. In such cases, the predicted answer under forced completion r… view at source ↗
Figure 3
Figure 3. Figure 3: Illustration of Answer Switch Denoising for MCQ task. After denoising, the decision path becomes clearer across the entire reasoning process. lower overall accuracy compared to the datasets considered earlier. Across all three benchmarks, we observe qualitatively con￾sistent behavior with our main results. In particular, the predicted answer often stabilizes before the completion of the full reasoning trac… view at source ↗
Figure 4
Figure 4. Figure 4: Early stopping performance for random, task-specific and generic probe-based gates on Qwen3-4B reasoning. Despite its promising results, our study is limited to the models, tasks, and reasoning lengths evaluated in this work. We do not study substantially larger models or much longer reasoning traces, where answer dynamics may differ. In addition, our experiments focus on language based reason￾ing tasks, a… view at source ↗
Figure 5
Figure 5. Figure 5: Early stopping performance for random, task-specific and generic probe-based gates on Qwen3-8B reasoning view at source ↗
Figure 6
Figure 6. Figure 6: Distribution of the mean of answer switches for raw and denoised outputs in Qwen3-4B, evaluated on the MCQ Questions. 0 0.5 1 1.5 2 mmlu pro (raw) mmlu pro (den) Transient flips view at source ↗
Figure 7
Figure 7. Figure 7: Distribution of the mean of transient flips for raw and denoised outputs in Qwen3-4B, evaluated on the MCQ Questions. 600 800 mmlu pro (raw) mmlu pro (den) Tokens after final switch view at source ↗
Figure 8
Figure 8. Figure 8: Distribution of the mean of tokens after final answer switch for raw and denoised outputs in Qwen3-4B, evaluated on the MCQ Questions. D.1.2. NUMERIC 2 3 4 5 6 7 8 numeric (raw) numeric (den) Answer switches view at source ↗
Figure 9
Figure 9. Figure 9: Distribution of the mean of answer switches for raw and denoised outputs in Qwen3-4B, evaluated on the Numeric Answers. 15 view at source ↗
Figure 10
Figure 10. Figure 10: Distribution of the mean of transient flips for raw and denoised outputs in Qwen3-4B, evaluated on the Numeric Answers. 800 1100 numeric (raw) numeric (den) Tokens after final switch view at source ↗
Figure 11
Figure 11. Figure 11: Distribution of the mean of tokens after final answer switch for raw and denoised outputs in Qwen3-4B, evaluated on the Numeric Answers. D.1.3. SEARCH 0.4 0.5 0.6 0.7 0.8 0.9 search (raw) search (den) Answer switches view at source ↗
Figure 16
Figure 16. Figure 16: Distribution of the mean of transient flips for raw and denoised outputs in Qwen3-4B, evaluated on the Tool Selection. 100 500 1000 1500 tool choice (raw) tool choice (den) Tokens after final switch view at source ↗
Figure 17
Figure 17. Figure 17: Distribution of the mean of tokens after final answer switch for raw and denoised outputs in Qwen3-4B, evaluated on the Tool Selection. D.2. Qwen3-8B Figures 18-29 shows distribution of metrics on Qwen3-8B. 16 view at source ↗
Figure 22
Figure 22. Figure 22: Distribution of the mean of transient flips for raw and denoised outputs in Qwen3-8b, evaluated on the Numeric Answers. 1500 2000 numeric (raw) numeric (den) Tokens after final switch view at source ↗
Figure 23
Figure 23. Figure 23: Distribution of the mean of tokens after final answer switch for raw and denoised outputs in Qwen3-8b, evaluated on the Numeric Answers. D.2.3. SEARCH 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 search (raw) search (den) Answer switches view at source ↗
Figure 26
Figure 26. Figure 26: Distribution of the mean of tokens after final answer switch for raw and denoised outputs in Qwen3-8B, evaluated on the Search Query. 17 view at source ↗
Figure 31
Figure 31. Figure 31: Distribution of the mean of transient flips for raw and denoised outputs in Qwen3-14B, evaluated on the MCQ Questions. 400 500 600 700 800 900 1000 mmlu pro (raw) mmlu pro (den) Tokens after final switch view at source ↗
Figure 32
Figure 32. Figure 32: Distribution of the mean of tokens after final answer switch for raw and denoised outputs in Qwen3-14B, evaluated on the MCQ Questions. D.3.2. NUMERIC 2 4 6 8 10 12 numeric (raw) numeric (den) Answer switches view at source ↗
Figure 29
Figure 29. Figure 29: Distribution of the mean of tokens after final answer switch for raw and denoised outputs in Qwen3-8B, evaluated on the Tool Selection. D.3. Qwen/Qwen3-14B Figures 30-41 shows the distribution of metrics on different datasets on Qwen3-14B. D.3.1. MCQ 0.5 1 1.5 2 2.5 3 3.5 mmlu pro (raw) mmlu pro (den) Answer switches view at source ↗
Figure 34
Figure 34. Figure 34: Distribution of the mean of transient flips for raw and denoised outputs in Qwen3-14B, evaluated on the Numeric answers. 18 view at source ↗
Figure 35
Figure 35. Figure 35: Distribution of the mean of tokens after final answer switch for raw and denoised outputs in Qwen3-14B, evaluated on the Numeric answers. D.3.3. SEARCH 0.2 0.3 0.4 0.5 0.6 0.7 search (raw) search (den) Answer switches view at source ↗
Figure 40
Figure 40. Figure 40: Distribution of the mean of transient flips for raw and denoised outputs in Qwen3-14B, evaluated on the Tool Selection. 50 100 200 400 800 tool choice (raw) tool choice (den) Tokens after final switch view at source ↗
Figure 41
Figure 41. Figure 41: Distribution of the mean of tokens after final answer switch for raw and denoised outputs in Qwen3-14B, evaluated on the Tool Selection. D.4. Qwen3-30B-A3B Figures 42-53 show bootstrap experiments performed on different datasets. D.4.1. MCQ 0.5 1 1.5 2 2.5 mmlu pro (raw) mmlu pro (den) Answer switches view at source ↗
Figure 38
Figure 38. Figure 38: Distribution of the mean of tokens after final answer switch for raw and denoised outputs in Qwen3-14B, evaluated on the Search query. D.3.4. TOOL SELECTION 0 5 · 10−2 0.1 0.15 0.2 0.25 0.3 tool choice (raw) tool choice (den) Answer switches view at source ↗
Figure 43
Figure 43. Figure 43: Distribution of the mean of transient flips for raw and denoised outputs in Qwen3-30B-A3B, evaluated on the MCQ Questions. 600 700 800 900 1000 mmlu pro (raw) mmlu pro (den) Tokens after final switch view at source ↗
Figure 44
Figure 44. Figure 44: Distribution of the mean of tokens after final answer switch for raw and denoised outputs in Qwen3-30B-A3B, evalu￾ated on the MCQ Questions. D.4.2. NUMERIC 0 2 4 6 8 10 12 numeric (raw) numeric (den) Answer switches view at source ↗
Figure 45
Figure 45. Figure 45: Distribution of the mean of answer switches for raw and denoised outputs in Qwen3-30B-A3B, evaluated on the Numeric Answers. 0 2 4 6 8 numeric (raw) numeric (den) Transient flips view at source ↗
Figure 46
Figure 46. Figure 46: Distribution of the mean of transient flips for raw and denoised outputs in Qwen3-30B-A3B, evaluated on the Numeric Answers. 500 1500 2500 3500 numeric (raw) numeric (den) Tokens after final switch view at source ↗
Figure 51
Figure 51. Figure 51: Distribution of the mean of answer switches for raw and denoised outputs in Qwen3-30B-A3B, evaluated on the Tool Selection. 0 1 2 3 4 5 ·10−2 tool choice (raw) tool choice (den) Transient flips view at source ↗
Figure 52
Figure 52. Figure 52: Distribution of the mean of transient flips for raw and denoised outputs in Qwen3-30B-A3B, evaluated on the Tool Selection. 101 102 tool choice (raw) Tokens after final switch view at source ↗
Figure 53
Figure 53. Figure 53: Distribution of the mean of tokens after final answer switch for raw and denoised outputs in Qwen3-30B-A3B, evalu￾ated on the Tool Selection. E. Training Details Positive and negative training examples. Probe training examples are constructed by sampling intermediate reason￾ing steps from each trace rather than using all steps. Positive examples correspond to reasoning states at which compu￾tation could b… view at source ↗
read the original abstract

Large Language Models often achieve strong performance by generating long intermediate chain-of-thought reasoning. However, it remains unclear when a model's final answer is actually determined during generation. If the answer is already fixed at an intermediate stage, subsequent reasoning tokens may constitute post-decision explanation, increasing inference cost and latency without improving correctness. We study the evolution of predicted answers over reasoning steps using forced answer completion, which elicits the model's intermediate predictions at partial reasoning prefixes. Focusing on Qwen3-4B and averaging results across all datasets considered, we find that predicted answers change in only 32% of queries. Moreover, once the final answer switch occurs, the model generates an average of 760 additional reasoning tokens per query, accounting for a substantial fraction of the total reasoning budget. Motivated by these findings, we investigate early stopping strategies that halt generation once the answer has stabilized. We show that simple heuristics, including probe-based stopping, can reduce reasoning token usage by 500 tokens per query while incurring only a 2% drop in accuracy. Together, our results indicate that a large portion of chain-of-thought generation is redundant and can be reduced with minimal impact on performance.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 1 minor

Summary. The paper claims that LLMs fix their final answers early during chain-of-thought generation. Using forced answer completion (inserting an answer-eliciting prompt after partial CoT prefixes) on Qwen3-4B averaged across datasets, predicted answers change in only 32% of queries; after the last switch the model still emits an average of 760 reasoning tokens. Simple early-stopping heuristics (including probe-based) are shown to cut ~500 tokens per query at a 2% accuracy cost, implying that much post-stabilization CoT is redundant post-decision explanation.

Significance. If the core measurements are valid, the work supplies a practical route to lower inference cost and latency for CoT models while preserving accuracy. The empirical quantification of answer stabilization timing and the token-budget savings are concrete and actionable; the early-stopping results constitute a falsifiable, immediately deployable contribution.

major comments (2)
  1. [Methods (forced answer completion procedure)] The central measurement (32% answer-change rate and the 760-token post-switch budget) rests on forced answer completion. No ablation tests whether inserting the probe prompt (e.g., “So the answer is”) itself shifts token probabilities or internal state relative to uninterrupted continuation; without such controls (prompt wording, temperature, or alternative read-outs such as logit inspection), the reported statistics may be artifacts of the intervention rather than evidence of natural early fixation.
  2. [Results / Experiments] Results section: the headline numbers (32% change rate, 760 additional tokens, 500-token savings at 2% accuracy drop) are presented without dataset sizes, number of queries, per-dataset breakdowns, variance across runs, or statistical significance tests. These omissions make it impossible to assess whether the averaged figures are robust or driven by a few datasets.
minor comments (1)
  1. [Notation and figures] Notation for “final answer switch” and “probe-based stopping” should be defined once with a short equation or pseudocode before being used in figures and tables.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for their constructive and detailed review. The comments highlight important aspects of our methodology and presentation that we will address in the revision. Below we respond point-by-point to the major comments.

read point-by-point responses
  1. Referee: [Methods (forced answer completion procedure)] The central measurement (32% answer-change rate and the 760-token post-switch budget) rests on forced answer completion. No ablation tests whether inserting the probe prompt (e.g., “So the answer is”) itself shifts token probabilities or internal state relative to uninterrupted continuation; without such controls (prompt wording, temperature, or alternative read-outs such as logit inspection), the reported statistics may be artifacts of the intervention rather than evidence of natural early fixation.

    Authors: We acknowledge that the forced-answer-completion procedure is an intervention and that we did not include explicit ablations comparing probe insertion to fully uninterrupted generation or alternative read-outs such as logit inspection. Our choice of a short, standard prompt (“So the answer is”) was intended to minimize disruption while still eliciting the model’s current prediction; the low observed change rate (32%) and consistency across datasets provide indirect support that the measurement reflects genuine stabilization rather than prompt-induced artifacts. Nevertheless, to strengthen the claim, the revised manuscript will add a dedicated paragraph in the Methods section discussing potential intervention effects and will report new ablation results using (i) varied prompt phrasings, (ii) different temperatures, and (iii) logit-based answer extraction on a subset of queries. These additions will be placed before the main results to allow readers to assess robustness. revision: partial

  2. Referee: [Results / Experiments] Results section: the headline numbers (32% change rate, 760 additional tokens, 500-token savings at 2% accuracy drop) are presented without dataset sizes, number of queries, per-dataset breakdowns, variance across runs, or statistical significance tests. These omissions make it impossible to assess whether the averaged figures are robust or driven by a few datasets.

    Authors: We apologize for the omission of these details in the submitted version. The experiments were run on the full set of datasets referenced in the paper (totaling several thousand queries). In the revised manuscript we will expand the Results section to include: (1) a table listing each dataset, its size, and the number of queries evaluated; (2) per-dataset breakdowns of the 32% change rate, post-switch token count, and early-stopping savings; (3) standard deviations across multiple runs; and (4) statistical significance tests (paired t-tests or Wilcoxon tests) for the accuracy differences between full CoT and early-stopping conditions. These additions will make the robustness of the headline averages transparent. revision: yes

Circularity Check

0 steps flagged

No circularity; claims rest on direct empirical measurements

full rationale

The paper reports observational statistics obtained by applying forced answer completion probes to partial CoT prefixes and counting answer changes plus subsequent tokens. These quantities (32% change rate, 760-token average) are measured outputs, not quantities fitted to data and then re-presented as predictions, nor self-defined via the result itself. No equations, uniqueness theorems, or ansatzes are invoked that reduce the central claims to prior inputs by construction. The early-stopping heuristics are motivated by the observations but remain independent proposals whose performance is separately evaluated. No load-bearing self-citations appear in the derivation chain.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

The central empirical claim rests on the assumption that forced answer completion is a valid probe of internal decision timing and that the chosen datasets and model are representative.

axioms (1)
  • domain assumption Forcing an answer at an intermediate prefix does not change the model's committed final answer relative to uninterrupted generation.
    This is required for the forced-completion measurements to reflect natural decision timing.

pith-pipeline@v0.9.0 · 5520 in / 1134 out tokens · 64588 ms · 2026-05-08T12:02:23.840351+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 1 Pith paper

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. LLMs Should Not Yet Be Credited with Decision Explanation

    cs.AI 2026-05 unverdicted novelty 4.0

    LLMs support decision prediction and rationale generation but lack evidence for genuine decision explanation, requiring stricter standards to avoid over-crediting.

Reference graph

Works this paper leans on

20 extracted references · 17 canonical work pages · cited by 1 Pith paper · 6 internal anchors

  1. [1]

    Bogdan, Uzay Macar, Neel Nanda, and Arthur Conmy

    URL https://matharena.ai/. Bogdan, P. C., Macar, U., Nanda, N., and Conmy, A. Thought anchors: Which llm reasoning steps matter? arXiv preprint arXiv:2506.19143,

  2. [2]

    Humanity's Last Exam

    doi: 10.1038/s41586-025-09962-4. URL https://arxiv.org/abs/2501.14249. Karpas, E., Abend, O., Belinkov, Y ., Lenz, B., Lieber, O., Ratner, N., Shoham, Y ., Bata, H., Levine, Y ., Leyton- Brown, K., et al. Mrkl systems: A modular, neuro- symbolic architecture that combines large language mod- els, external knowledge sources and discrete reasoning. arXiv pr...

  3. [3]

    Can language models learn from explanations in context? InFindings of the Association for Computa- tional Linguistics: EMNLP 2022, pp

    Lampinen, A., Dasgupta, I., Chan, S., Mathewson, K., Tessler, M., Creswell, A., McClelland, J., Wang, J., and Hill, F. Can language models learn from explanations in context? InFindings of the Association for Computa- tional Linguistics: EMNLP 2022, pp. 537–563,

  4. [4]

    Measuring Faithfulness in Chain-of-Thought Reasoning

    Lanham, T., Chen, A., Radhakrishnan, A., Steiner, B., Deni- son, C., Hernandez, D., Li, D., Durmus, E., Hubinger, E., Kernion, J., Luko ˇsi¯ut˙e, K., Nguyen, K., Cheng, N., Joseph, N., Schiefer, N., Rausch, O., Larson, R., McCan- dlish, S., Kundu, S., Kadavath, S., Yang, S., Henighan, T., Maxwell, T., Telleen-Lawton, T., Hume, T., Hatfield- Dodds, Z., Kap...

  5. [5]

    Internal consistency and self-feedback in large language models: A survey.arXiv preprint arXiv:2407.14507,

    Liang, X., Song, S., Zheng, Z., Wang, H., Yu, Q., Li, X., Li, R.-H., Wang, Y ., Wang, Z., Xiong, F., et al. Internal consistency and self-feedback in large language models: A survey.arXiv preprint arXiv:2407.14507,

  6. [6]

    Let's Verify Step by Step

    Lightman, H., Kosaraju, V ., Burda, Y ., Edwards, H., Baker, B., Lee, T., Leike, J., Schulman, J., Sutskever, I., and 9 Submission and Formatting Instructions for ICML 2026 Cobbe, K. Let’s verify step by step.arXiv preprint arXiv:2305.20050,

  7. [7]

    and Venkat, A

    Mao, Z. and Venkat, A. Recurrent confidence chain: Temporal-aware uncertainty quantification in large lan- guage models.arXiv preprint arXiv:2601.13368,

  8. [8]

    K., Singhu, S., and Ruchkin, I

    Mao, Z., Venkat, A., Bisliouk, A., Kothiyal, A., Subra- manian, S. K., Singhu, S., and Ruchkin, I. Confidence over time: Confidence calibration with temporal logic for large language model reasoning.arXiv preprint arXiv:2601.13387,

  9. [9]

    Reasoning about uncertainty: Do reason- ing models know when they don’t know?arXiv preprint arXiv:2506.18183,

    Mei, Z., Zhang, C., Yin, T., Lidard, J., Shorinwa, O., and Majumdar, A. Reasoning about uncertainty: Do reason- ing models know when they don’t know?arXiv preprint arXiv:2506.18183,

  10. [10]

    Plaat, A., Wong, A., Verberne, S., Broekens, J., Van Stein, N., and B¨ack, T

    URLhttps://arxiv.org/abs/2406.11811. Plaat, A., Wong, A., Verberne, S., Broekens, J., Van Stein, N., and B¨ack, T. Multi-step reasoning with large language models, a survey.ACM Computing Surveys, 58(6):1–35,

  11. [11]

    GPQA: A Graduate-Level Google-Proof Q&A Benchmark

    URL https://arxiv.org/abs/2311.12022. Shao, Z., Wang, P., Zhu, Q., Xu, R., Song, J., Bi, X., Zhang, H., Zhang, M., Li, Y ., Wu, Y ., et al. Deepseekmath: Push- ing the limits of mathematical reasoning in open language models.arXiv preprint arXiv:2402.03300,

  12. [12]

    What evidence do language models find convincing?arXiv preprint arXiv:2402.11782,

    Wan, A., Wallace, E., and Klein, D. What evidence do language models find convincing?arXiv preprint arXiv:2402.11782,

  13. [13]

    Self-Consistency Improves Chain of Thought Reasoning in Language Models

    Wang, X., Wei, J., Schuurmans, D., Le, Q., Chi, E., Narang, S., Chowdhery, A., and Zhou, D. Self-consistency im- proves chain of thought reasoning in language models. arXiv preprint arXiv:2203.11171,

  14. [14]

    MMLU-Pro: A More Robust and Challenging Multi-Task Language Understanding Benchmark

    URL https://arxiv.org/abs/2406.01574. Wei, J., Wang, X., Schuurmans, D., Bosma, M., Xia, F., Chi, E., Le, Q. V ., Zhou, D., et al. Chain-of-thought prompting elicits reasoning in large language models.Advances in neural information processing systems, 35:24824–24837,

  15. [15]

    arXiv preprint arXiv:1908.04626 , year=

    Wiegreffe, S. and Pinter, Y . Attention is not not explanation. arXiv preprint arXiv:1908.04626,

  16. [16]

    Qwen3 Technical Report

    URL https: //arxiv.org/abs/2505.09388. Yao, S., Zhao, J., Yu, D., Du, N., Shafran, I., Narasimhan, K. R., and Cao, Y . React: Synergizing reasoning and acting in language models. InThe eleventh international conference on learning representations,

  17. [17]

    Reasoning models better express their confidence.arXiv preprint arXiv:2505.14489,

    Yoon, D., Kim, S., Yang, S., Kim, S., Kim, S., Kim, Y ., Choi, E., Kim, Y ., and Seo, M. Reasoning models better express their confidence.arXiv preprint arXiv:2505.14489,

  18. [18]

    and Zhang, R

    Zhang, B. and Zhang, R. Cot-uq: Improving response-wise uncertainty quantification in llms with chain-of-thought. arXiv preprint arXiv:2502.17214,

  19. [19]

    answer":

    URL https://openreview.net/forum? id=WZH7099tgfM. 10 Submission and Formatting Instructions for ICML 2026 A. Datasets We use the sample sizes described in Table 2 to conduct our analyses and early stopping experiments for all the datasets. We shuffle and randomly sample from the dataset to obtain samples. DatasetS MCQ 1000 Numeric-answer 500 Search-query ...

  20. [20]

    C.2.1. RANDOM PROBING Table 16.MCQ Accuracy drop Tokens saved Token % 24.390 1764 99.5 24.429 1762 99.4 24.527 1759 99.2 24.527 1755 99.0 24.473 1751 98.8 24.571 1745 98.4 24.732 1735 97.9 24.707 1721 97.0 24.141 1690 95.3 22.459 1608 90.7 19.107 1468 82.8 9.112 854 48.2 0.488 0 0.0 Table 17.Numeric-answer Accuracy drop Tokens saved Token % 59.722 2920 99...