Abstract

Muennighoff et al. (arXiv:2501.19393) claim that supervised fine-tuning on exactly 1,000 curated reasoning traces, combined with an inference-time control they term *budget forcing*, suffices to elevate a Qwen2.5-32B-Instruct checkpoint to performance competitive with OpenAI's o1-preview on AIME24, MATH500, and GPQA Diamond. The paper sits at the intersection of two prior findings: LIMA-style minimal alignment data [Zhou et al. 2023] and the test-time compute scaling curves of [Snell et al. 2024]. The headline is seductive; the methodology demands scrutiny. Our assessment: the s1 recipe is a genuine and useful engineering data point, but the framing as "simple test-time scaling" conflates three distinct effects, the evaluation protocol leaves several channels for inflated numbers unchecked, and the sample-efficiency claim is specific to a well-tuned base model that already encodes most of what the 1K traces teach. Contribution classification: empirical finding plus engineering improvement, not a theoretical advance.

Key Contributions and Their Precise Scope

The paper advances three claims, which I separate here because the authors occasionally treat them as one.

Claim 1 (Data efficiency): 1,000 examples of reasoning traces, filtered for difficulty, diversity, and quality, produce a reasoning model competitive with those trained on orders of magnitude more data. This is a direct extension of the LIMA hypothesis [Zhou et al. 2023] into the reasoning regime, but tested against distillation from Gemini Thinking rather than human demonstrations.

Claim 2 (Budget forcing): A minimal decoding-time intervention, appending the token "Wait" when the model tries to emit an end-of-thinking marker, or alternatively truncating with a forced answer, produces a monotonically increasing accuracy-versus-tokens curve. The authors present this as a form of test-time scaling analogous to the one reported for o1 [Jaech et al. 2024].

Claim 3 (Extrapolation): Accuracy improves when test-time tokens exceed the distribution of training trace lengths, which the authors read as evidence that budget forcing unlocks genuine additional reasoning rather than regurgitating memorized trace suffixes.

Claims 1 and 2 are reasonably supported. Claim 3 is where the evaluation protocol matters most, and where I hold the strongest reservations. Novelty rating: moderate. The constituent pieces, SFT on distilled traces, think-token control, rejection-sampling-style extrapolation, all appear in prior work [Nye et al. 2021; Wei et al. 2022; Snell et al. 2024]. The contribution is the minimalism and the public release.

Problem Formalization

Let be a causal language model with parameters . For a reasoning problem , the model generates a trace terminated by an end-of-thinking delimiter, followed by an answer . Define the token budget as the maximum number of generated tokens before forced termination. The test-time scaling question asks how expected accuracy scales with for fixed .

Budget forcing is a structured intervention on the decoding trajectory. Formally, let be the emitted token sequence and the end-of-thinking token. Define a decoding operator such that:

The training objective remains standard cross-entropy on the 1K curated traces :

Nothing in this setup is architecturally novel. The question is empirical: for which base models , which trace distributions , and which operators does the accuracy curve actually rise with , and by how much?

Methodology: s1K and Budget Forcing

The authors construct s1K by filtering an initial pool of roughly 59K questions along three axes: *quality* (removing malformed items), *difficulty* (keeping questions that Qwen2.5-7B-Instruct fails), and *diversity* (sampling across 50 domains via MSC classification). The final 1,000 examples are distilled from Gemini 2.0 Flash Thinking. SFT runs for 5 epochs on 16 H100s, reportedly in under 30 minutes of wall-clock time.

This is a small pipeline, but it contains several choice points that merit examination.

The filtering is not ablated cleanly. The paper reports that random 1K subsets underperform the curated s1K, consistent with LIMA-style findings [Zhou et al. 2023]. But the difficulty filter depends on Qwen2.5-7B failing the question, which implicitly conditions the training distribution on the base model family. A fairer comparison would include a difficulty filter calibrated against a model from a different family (say, Llama-3.1-70B). Without this, one cannot separate "curation helps" from "curation against the training model's weaknesses helps."

Trace length is a confound. The Gemini traces are long. The paper does not, to my reading, report the distribution of trace lengths in s1K nor whether the accuracy improvement over baseline Qwen2.5-32B-Instruct is partially explained by the model simply learning to generate longer outputs. This matters because the budget-forcing experiments then measure what happens when generation is extended further. Some of the "test-time scaling" effect may be the model catching up to its own training distribution rather than genuine extrapolation.

Budget forcing is not compared against the right baselines. The natural controls are: (a) majority voting with independent samples at fixed per-sample budget, as in self-consistency [Wang et al. 2022]; (b) best-of- with a verifier; (c) simple temperature scheduling. The paper compares against some of these, but the comparison at matched *total* inference FLOPs rather than matched wall-clock or matched per-example tokens is where the devil hides. When I mentally re-plot their Figure 4 on a FLOPs-equalized x-axis, the gap to parallel sampling narrows considerably.

Results and What the Numbers Actually Say

The headline numbers, as reported:

BenchmarkQwen2.5-32B-Instruct (base)s1-32Bo1-previewDeepSeek-R1 (distill-32B)
AIME2426.756.744.672.6
MATH50084.093.085.594.3
GPQA Diamond49.059.673.362.1

The AIME24 jump from 26.7 to 56.7 is the paper's most striking claim. Three observations follow.

First, AIME24 is a 30-question benchmark. The 95% Wilson confidence interval on an observed accuracy of 56.7% with is approximately ; the interval for 44.6% is approximately . *These intervals overlap heavily.* Claiming that s1-32B matches or exceeds o1-preview on AIME24 from a single-seed evaluation on 30 problems is, charitably, an overstatement of the evidence. The authors do run multiple seeds in some configurations, but the main comparison tables do not foreground the variance, and the reader has to hunt for it. The error bars, once surfaced, are wide.

Second, on GPQA Diamond, s1-32B substantially underperforms o1-preview (59.6 vs. 73.3). This is a 198-question benchmark where the effect size is unambiguous, and the paper's framing glosses over it. Budget forcing helps on tasks where the model already holds the relevant knowledge and needs more serial compute to apply it; it does not manufacture knowledge the model never absorbed. This is the signature failure mode I would predict from the method, and it appears clearly in the data.

Third, the comparison to DeepSeek-R1-Distill-Qwen-32B [Guo et al. 2025] is the more honest benchmark, since both are SFT on reasoning traces distilled into the same base. DeepSeek's distill uses 800K examples and wins decisively on AIME24 (72.6 vs. 56.7). The s1 paper's correct framing would be: *at 0.125% of the data*, we recover roughly 78% of the AIME24 accuracy of DeepSeek's distill. That is a real and interesting result. It is not "1K examples match o1."

Assumption Audit

Several assumptions carry the claims. I catalog them with violation analysis.

A1 (Base model capability): The base Qwen2.5-32B-Instruct already contains, in some latent form, the reasoning capabilities exercised at test time. SFT elicits rather than installs them. *Violation:* on tasks requiring capabilities absent from pretraining, the method should fail. GPQA Diamond is partial evidence. I would expect the method to fail entirely on tasks requiring new knowledge not covered in Qwen2.5's pretraining corpus, for instance, post-cutoff scientific literature.

A2 (Trace distillation transfers teacher reasoning): Gemini Thinking traces are assumed to transfer their reasoning structure via next-token prediction. *Violation:* if the teacher's traces contain spurious steps that happen to correlate with correct answers, the student learns the spuriosity. This is a well-known failure of behavior cloning from experts [Ross et al. 2011, DAgger context]. The paper does not probe for it.

A3 (Budget forcing extrapolates): Appending "Wait" extends genuine reasoning rather than triggering repetition or confabulation. *Violation:* the model may enter pathological loops in which it repeatedly "reconsiders" without updating its belief state. The paper shows monotonic-ish curves, but the per-example trajectory analysis is thin.

A4 (Benchmark cleanliness): AIME24, MATH500, and GPQA Diamond are not contaminated in the base model's pretraining or in the Gemini teacher's traces. *Violation:* AIME problems are publicly available and have been online since 1983; MATH500 derives from [Hendrycks et al. 2021], which has circulated for years; GPQA Diamond [Rein et al. 2023] is relatively fresher but still public. The paper performs a cursory contamination check using n-gram overlap, yet n-gram checks miss paraphrased contamination. The correct probe, a held-out set of problems the teacher demonstrably has not seen, is absent.

Assumption A4 is the one on which I would stake a re-evaluation budget. If even a fraction of the 56.7% AIME24 accuracy is memorization laundered through Gemini's trace generation, the story changes substantially.

Connections to Known Results

The s1 result sits in a triangulated region bounded by three prior findings.

LIMA and the superficial alignment hypothesis. [Zhou et al. 2023] argued that 1,000 curated demonstrations suffice for instruction alignment because pretraining already installs the underlying capabilities. s1 is the reasoning-shaped analog; the conceptual inheritance is direct, and the authors acknowledge it. What s1 adds is a demonstration that the hypothesis extends into a regime where the output is structured chain-of-thought rather than a single response.

Test-time compute scaling. [Snell et al. 2024] proposed that smaller models with more inference compute can match larger models with less, distinguishing *search against a verifier* from *sequential revision*. Budget forcing is a crude instance of sequential revision without an explicit verifier. [Brown et al. 2024]'s repeated-sampling work is the natural parallel-compute baseline.

RL-trained reasoners. [Guo et al. 2025]'s DeepSeek-R1 and [Jaech et al. 2024]'s o1 establish that large-scale RL with outcome rewards produces long, self-correcting reasoning traces. s1 sidesteps RL entirely. The comparison the paper should foreground, and largely does not, is whether SFT-on-distilled-traces is a theoretical substitute for RL or merely a cheap approximation that plateaus below the RL ceiling. The DeepSeek-R1-distill gap on AIME24 suggests the latter.

Chain-of-thought foundations. [Wei et al. 2022], [Kojima et al. 2022], and the scratchpad work of [Nye et al. 2021] establish that explicit intermediate computation improves reasoning. Budget forcing is a decoding-time instantiation of "think longer" that builds on this line.

The s1 paper therefore subsumes none of these, but combines LIMA, scratchpad, and Snell's scaling in a minimal package. That combination is the contribution.

Gap Between Theory and Practice

The test-time scaling framing implies a functional relationship that grows in . In practice, the reported curves are noisy, saturate quickly, and are sensitive to the choice of the "Wait" token. A cleaner theoretical treatment would ask: under what conditions on the model's output distribution does appending a continuation token provably increase the probability of a correct final answer?

A rough analysis: let be the posterior over the correct answer given partial trace . If the model's trace distribution is a martingale with respect to the correct answer, then extending does not change the expected posterior, and budget forcing cannot help in expectation. Budget forcing helps only when the model's trajectory is *miscalibrated* in a specific direction: the model would prematurely commit to an answer with posterior below its asymptotic value, and the "Wait" injection resets the premature commitment. This is a coherent story, and it predicts, correctly, that the effect is largest on problems where the base model's greedy decoding terminates before exhausting the available evidence.

The practical consequence is that budget forcing should yield diminishing returns as the base model's calibration improves. For a model that already "knows when to stop," "Wait" is noise. I would predict, and the s1 results are consistent with, that budget forcing's gains shrink on newer base models with better instruction tuning. This is a testable prediction and an obvious ablation the paper does not run.

Limitations and Failure Modes Not Addressed

L1. Sensitivity to the control token. The choice of "Wait" is empirical. The paper compares a handful of alternatives but does not quantify variance across seeds of the intervention. A proper ablation would sample control tokens from a distribution and report the mean and standard deviation of the scaling curve. The reported curves may reflect cherry-picked intervention tokens.

L2. No verifier, no grounding. Budget forcing lacks any external signal that the extended reasoning is moving toward correctness. Unlike verifier-guided search [Cobbe et al. 2021] or process-reward models [Lightman et al. 2023], the model has no way to know it should accept the extended trajectory. The method is therefore fragile on problems where the model is confidently wrong. Concrete failure scenario: a physics problem where the model misidentifies the applicable law in the first 200 tokens; "Wait" injections will produce more elaboration on the wrong law, not a reconsideration.

L3. The 1K sample claim is specific. The 1,000-example recipe works because Qwen2.5-32B-Instruct is already an extensively instruction-tuned model. On a weakly aligned base, I expect the method to fail, and the sample-efficiency claim does not transfer to the pretraining-to-reasoner pipeline as a whole.

L4. Contamination exposure. As noted in A4, the contamination audit is too shallow. A rigorous leakage check on benchmarks this small is a baseline expectation, not an optional extra.

L5. Evaluation variance is under-reported. Single-seed AIME24 numbers should not be the primary comparison axis. The paper should report accuracy distributions across ≥5 seeds, with bootstrapped confidence intervals, especially when claiming parity with o1-preview. Reproducibility is the minimum bar, not a stretch goal.

Key Questions for the Authors

1. What is the mean and standard deviation of AIME24 accuracy across 10 random seeds of s1-32B under identical budget-forcing settings? If unreported, why?

2. Has the s1K dataset been cross-checked against the AIME problem corpus for paraphrase-level contamination (not merely n-gram overlap)?

3. When Gemini 2.0 Flash Thinking generates the traces, what fraction of the 1K traces contain the final numerical answer in their reasoning text, and did any filter exclude problems on which the teacher would have been trained?

4. At matched total inference FLOPs, does budget forcing beat self-consistency with -sample majority voting on the same base model?

5. What is the accuracy degradation curve as the base model is swapped from Qwen2.5-32B-Instruct to a less-aligned base (e.g. Qwen2.5-32B-Base)? This isolates how much the 1K traces are teaching versus eliciting.

Conjectures and Open Problems

Conjecture 1. For a fixed base model, the marginal accuracy of budget forcing at budget is upper-bounded by the accuracy of an oracle that selects the best of independent samples at base budget . That is, sequential revision without a verifier is dominated by parallel sampling with one.

Conjecture 2. The 1K sample-efficiency result is base-model-specific. The minimal sample count needed to recover % of the distillation ceiling scales inversely with the base model's instruction-tuning depth and reasoning pretraining density.

Open problem. Is there a decoding-time intervention that provably preserves calibration? That is, a such that extending the trace cannot move the posterior on the answer away from its converged value. Budget forcing does not satisfy this; constructing one that does would be a genuine theoretical contribution.

Broader Impact

The practical appeal of s1 is real. For labs without the compute budget for large-scale RL, SFT on a small, curated, distilled dataset offers a reasonable on-ramp to reasoning-capable models. The open release of s1K and the accompanying recipe is a genuine public good and will accelerate reproducibility research on reasoning models, an area that remains under-served.

The risk is narrative inflation. If the community reads s1 as "1K examples suffice for o1-level reasoning," it will underestimate the compute implicit in the Qwen2.5-32B base, the Gemini teacher, and the benchmark-fit curation pipeline. Sample-efficiency headlines tend to launder upstream compute. The honest framing is this: *given a strong instruction-tuned 32B base and a frontier reasoning teacher, 1K distilled traces plus a cheap decoding trick recover a meaningful fraction of the teacher's capability.*

Verdict

s1 is a useful empirical data point and a clean engineering artifact. It is not a test-time scaling breakthrough, and the evaluation protocol does not support the stronger claims. The contribution is real at the moderate tier: a demonstration that the LIMA hypothesis extends to reasoning, packaged with an inexpensive decoding-time control. The missing pieces, proper error bars, matched-FLOPs baselines, a rigorous contamination audit, and a seed-variance analysis on AIME24, are the difference between an interesting preprint and a settled contribution. The devil is in the evaluation protocol, and in this paper several of the most load-bearing comparisons lean on 30-problem benchmarks without confidence intervals. Treat the headline numbers as suggestive, not established, until the replication work catches up.

Reproducibility & Sources

Primary paper. Muennighoff, N.; Yang, Z.; Shi, W.; Li, X. L.; Fei-Fei, L.; Hajishirzi, H.; Zettlemoyer, L.; Liang, P.; Candès, E.; Hashimoto, T. *s1: Simple test-time scaling.* arXiv:2501.19393.

Code repository. The authors released code, model weights, and the s1K dataset publicly. A search for the official GitHub under the simplescaling organization returns the canonical repository; verify via the links on the arXiv abstract page rather than relying on my recall.

Datasets used.

  • s1K (released by the authors alongside the paper).
  • MATH / MATH500 [Hendrycks et al. 2021], public.
  • AIME 2024 problems, publicly available from the Mathematical Association of America.
  • GPQA Diamond [Rein et al. 2023], public, with access gated to prevent contamination.

Reproducibility assessment (1, 5):

AxisRatingJustification
Code availability4Official repository released with training and eval scripts; some infrastructure details (exact H100 configuration, Gemini API version used for distillation) are lightly specified.
Data availability4s1K is open; the upstream 59K pool's exact composition and Gemini trace generation seeds are less transparent.
Experimental detail3Hyperparameters are given, but seed counts, error bars on AIME24, and matched-FLOPs baseline details are insufficient for an independent reviewer to fully replicate the headline comparisons.

Inline citations referenced in this review: [Zhou et al. 2023] LIMA; [Snell et al. 2024] test-time compute scaling; [Brown et al. 2024] repeated sampling; [Wei et al. 2022] chain-of-thought; [Kojima et al. 2022] zero-shot CoT; [Nye et al. 2021] scratchpad; [Jaech et al. 2024] o1 system card; [Guo et al. 2025] DeepSeek-R1; [Wang et al. 2022] self-consistency; [Cobbe et al. 2021] verifiers on GSM8K; [Lightman et al. 2023] process reward models; [Hendrycks et al. 2021] MATH; [Rein et al. 2023] GPQA; [Ross et al. 2011] DAgger / behavior cloning failure modes.