Ask GPT-4 who Tom Cruise's mother is, and it answers fluently. Ask it whose son Mary Lee Pfeiffer is, and it often fails. This is the reversal curse [Berglund et al. 2023], and it remains one of the cleanest diagnostics we have for the claim that autoregressive (AR) language models encode directional, rather than relational, knowledge. LLaDA [Nie et al. 2025; arXiv:2502.09992] proposes that a masked diffusion objective, scaled to 8B parameters and trained from scratch, not only matches LLaMA3 8B on standard benchmarks but also largely closes the reversal gap. The pertinent question is not whether the model arrives at the right answer, but how, and more specifically, whether the gain is attributable to the diffusion formulation itself, to bidirectional attention during training, or to an artifact of the evaluation protocol. This review is a surgical dissection of that question.

1. The Claim, Stated Formally

LLaDA is a discrete masked diffusion model over tokens. Let denote a clean sequence of length over vocabulary , and let be a distinguished mask symbol. The forward corruption process is defined token-wise: at continuous time , each token is independently replaced by with probability ,

At we recover the clean sequence; at the sequence is fully masked. The model is trained to predict from , equivalently, to denoise the masked positions, under the reweighted cross-entropy objective

The factor is not cosmetic: it renders an upper bound on the negative log-likelihood of , as derived for absorbing-state discrete diffusion by [Austin et al. 2021] and subsequently tightened by [Shi et al. 2024] and [Sahoo et al. 2024] under the name Simplified Masked Diffusion (SMD). LLaDA's loss is, up to notation, the SMD loss applied at LLM scale.

The formal claim of the paper is twofold:

1. Scaling claim: Pretraining a 2.3T-token, 8B-parameter masked diffusion model from scratch yields downstream performance (MMLU, GSM8K, HumanEval, BBH, instruction-following via SFT) statistically comparable to LLaMA3 8B at similar compute.

2. Inductive-bias claim: The bidirectional training signal endows LLaDA with non-trivial gains on reversal and infilling tasks where AR models structurally fail.

These are distinct claims with distinct evidentiary standards, and I will argue they should be assessed separately.

2. Derivation Walkthrough: Where the Assumptions Enter

2.1 The bound

The cleanest way to see why LLaDA can even call itself a language model is to trace the ELBO. For an absorbing discrete diffusion with mask rate , the negative ELBO decomposes into independent per-token terms, because the corruption factorizes, and one can show

The key identity is that, conditional on a token being masked at time , its posterior under the absorbing process is exactly the data distribution at that position given the partial context. This is the discrete analogue of Tweedie's formula; it is why a single cross-entropy head suffices, and why no separate noise-prediction head is required, in contrast to continuous Gaussian diffusion [Ho et al. 2020].

Where assumptions enter:

  • *Token independence in the forward process*. The corruption masks positions independently. Relaxing this, for instance, via span-masking as in BART or T5, would break the clean ELBO decomposition, and the weighting would no longer yield a valid bound without modification.
  • *Uniform sampling*. The paper samples . The gradient variance of is dominated both by small (where explodes but few tokens are masked) and by large (where many tokens are masked but the task reduces to near-unconditional generation). Importance sampling over , analogous to the weighting schedules of [Kingma et al. 2021] in continuous diffusion, is an obvious lever the paper does not fully explore.
  • *Mask token as distinct symbol*. LLaDA treats as a vocabulary element. This architectural choice introduces a potential train/inference mismatch: the embedding of is shaped entirely by training dynamics, and its interaction with rotary position embeddings is nontrivial.

2.2 The sampling procedure

Inference is the pressure point. LLaDA generates by starting with a fully masked sequence and iteratively unmasking tokens across reverse steps. At each step, the model produces for all masked positions, a subset is "committed" (unmasked), and the process repeats. LLaDA employs a semi-autoregressive variant: the sequence is divided into blocks decoded left-to-right, while within each block the decoding is parallel diffusion.

This is the crucial design detail most readers will skim. Consider what happens under stress. Pure parallel unmasking is known to underperform because the model's per-position marginals are not jointly consistent: sampling each masked position independently from its marginal ignores token-token dependencies given the partial observation. The semi-autoregressive scheme mitigates this by reducing the within-block fan-out, but reintroduces a left-to-right bias at block boundaries. In the limit of block size one, LLaDA degenerates to AR decoding; in the limit of block size , it degenerates to fully parallel MaskGIT-style decoding [Chang et al. 2022], which is known to struggle with long, coherent text. The "sweet spot" block size is an empirical hyperparameter, and the paper's ablation of it is thinner than I would like for a load-bearing design choice.

2.3 Complexity

Per training step, the FLOP count matches a dense transformer on a sequence of length : for attention and MLP respectively. Training FLOPs per token are therefore comparable to AR at matched architecture. But the effective loss signal per forward pass is applied only to masked positions (expected count ), whereas AR computes loss on every position. Equating signal per FLOP, LLaDA receives roughly half the learning signal per pass, and the weighting redistributes rather than amplifies it. This is the hidden compute tax of masked diffusion that scaling curves must absorb.

At inference, LLaDA requires forward passes through the full sequence to generate it, compared with forward passes under KV-caching for AR. When , LLaDA is faster in wall-clock per sample; when quality demands , the two converge, and LLaDA loses the KV-cache benefit entirely. The paper reports reasonable quality at to , yielding a 2, 4x inference speedup in principle, but only if attention is not recomputed over already-committed tokens, a nontrivial engineering question left somewhat open.

3. Comparison to Alternative Formulations

LLaDA sits in a crowded conceptual neighborhood. Three alternatives are worth disentangling:

(a) Continuous diffusion over embeddings [Li et al. 2022, Diffusion-LM; Gulrajani & Hashimoto, 2023, Plaid]. These methods add Gaussian noise to word embeddings and learn to denoise. They preserve the full continuous diffusion toolkit (DDIM samplers, classifier guidance), but require a separate decoding step from embeddings to tokens, and have not scaled competitively past roughly 1B parameters. LLaDA's discrete formulation avoids the quantization step entirely, which is precisely why it can ride transformer scaling laws more cleanly.

(b) Score-based discrete diffusion [Lou et al. 2024, SEDD]. SEDD parameterizes the concrete score and uses a score-entropy loss. This is more flexible than LLaDA's absorbing-state formulation, it admits uniform-state corruptions as well, but it is harder to train stably and introduces an additional normalizing-constant estimate. LLaDA opts for the simpler absorbing variant, likely the right call for a first scaling experiment, though it leaves the richer corruption schedules SEDD enables on the table.

(c) Semi-autoregressive sequence-level models [Gong et al. 2023, DiffuSeq; Han et al. 2023, SSD-LM]. These are ancestors of LLaDA, typically trained on seq2seq tasks with small decoders. LLaDA's contribution is less algorithmic novelty than the demonstration that the objective scales.

The honest framing is that LLaDA is an engineering and scaling contribution built atop SMD [Shi et al. 2024; Sahoo et al. 2024], which in turn descends from D3PM [Austin et al. 2021]. The novelty lies in the successful scaling, not in the loss itself. On the contribution taxonomy, I would rate it a moderate-to-significant empirical finding, not a new theoretical result.

4. Experimental Validation: What Is Actually Being Tested?

4.1 The benchmark table

BenchmarkLLaMA3 8B (reported)LLaDA 8B BaseGap
MMLU (5-shot)~66.6~65.9-0.7
GSM8K (8-shot)~56.8~70.7+13.9
HumanEval (0-shot)~33.5~33.5~0
BBH (3-shot)~57.7~49.8-7.9
Reversal (authors' constructed)substantially lowersubstantially higherlarge

(Numbers are approximate; exact reported values should be verified against Tables 1, 3 of the manuscript, since the paper reports several evaluation protocols, and the closest-matching numbers are quoted here.)

The GSM8K number is the one that should give a careful reviewer pause. A 13+ point gain over LLaMA3 8B on grade-school math, from a diffusion model, at matched parameter count, is the kind of result that demands a mechanistic explanation and a careful control. There are at least three alternative explanations to "diffusion is inherently better at math":

1. Data composition differences. LLaDA was pretrained on 2.3T tokens assembled by the authors. LLaMA3's pretraining corpus is proprietary. A difference in math-heavy data fraction could easily produce this gap.

2. Evaluation protocol differences. LLaDA's semi-autoregressive decoding with self-consistency-like token committing is a different inference regime than greedy or nucleus sampling on AR. If the diffusion sampler behaves as implicit self-consistency, the comparison is not apples-to-apples.

3. Tokenizer and prompting effects. Masked diffusion is more robust to certain prompt layouts (e.g. filling in a template), which may interact favorably with chain-of-thought formatting.

The paper does not fully isolate these. A fairer comparison would hold pretraining data fixed by training an AR model of identical size on the same tokens and schedule, a control that is expensive but necessary to support the inductive-bias claim.

4.2 The reversal experiments

The reversal reasoning result is at once the most interesting and the most underscrutinized. The authors show that LLaDA handles bidirectional factual queries better than AR baselines. But the reversal curse, as formulated by [Berglund et al. 2023], concerns *generalization from directional training data to the opposite direction at inference*. LLaDA's bidirectional training implies both directions are seen during pretraining (in the sense that the model must predict any masked position from any context). So it is not solving the reversal curse under the original construction; it is demonstrating that a model trained bidirectionally handles bidirectional queries, which is closer to tautology than to breakthrough absent careful controls.

The controlled version of the claim would require training LLaDA only on forward-direction data (e.g. only "A's mother is B" and never "B's son is A"), then testing on reverse. Does the bidirectional *objective* still help, or is it only the bidirectional *data* that matters? This is the missing ablation, and the one that would convince me the inductive bias is real.

4.3 Statistical rigor

The paper reports single-seed numbers with no confidence intervals on most benchmarks. Given that the diffusion sampler has stochasticity from -scheduling and mask ordering, and given that several reported gaps fall within 1, 2 points of noise, error bars would substantially strengthen the comparison. This is a routine concern for any 8B-scale evaluation, but it bites harder here because the diffusion sampler is an additional source of variance.

5. Failure Mode Analysis

Where does this approach degrade? I see at least four concrete scenarios:

(i) Long-form coherent generation. Diffusion LMs decode in a non-sequential order, committing tokens whose marginals are most confident first. For long-range narrative or code with nested scope, this can yield locally plausible but globally incoherent outputs, because the model commits to surface tokens before resolving latent structural decisions. AR models, by contrast, force sequential dependency resolution. The paper's HumanEval numbers are on par with LLaMA3 8B, but HumanEval is short-form; an MBPP-Plus or LiveCodeBench comparison on longer programs would probe this edge.

(ii) Constrained generation with KV-cache dependencies. Speculative decoding, prefix caching, and batched inference infrastructure are all built around AR semantics. A diffusion LM forfeits all of them. At scale, this is an inference-economics failure mode the paper does not address; the 2, 4x theoretical per-sample speedup evaporates in production settings where KV-cache reuse across requests dominates.

(iii) Very low $t$ regimes. When only a few tokens are masked, the weighting explodes variance, and the model learns fine-grained local completions. This is precisely where a BERT-like objective would do well, but also where the diffusion ELBO is loosest. Gradient clipping or importance resampling is likely required for stable training; the paper is sparse on this.

(iv) Instruction-following with long, structured outputs. SFT on a diffusion backbone remains an active question. If the instruction and response are concatenated and the response is masked, the model learns conditional generation, but the per-token weighting interacts with response length in subtle ways. Adversarial prompts that exploit the parallel decoding order, for instance, prompts demanding late-token commitment before early-token commitment, remain untested.

6. Open Technical Questions

Five questions I would put to the authors on a program committee:

1. Data-controlled AR baseline: What happens if you train an AR model of the same size on the same 2.3T tokens with the same tokenizer? Without this, the MMLU/GSM8K comparisons confound architecture and data.

2. Reversal curse under directional training: If you ablate the training corpus to remove reverse-direction pairs, does LLaDA still generalize to reversed queries? This is the test that separates objective from data.

3. Inference variance: What are the performance standard deviations across sampler seeds at fixed , and how does quality degrade as decreases below ?

4. Mask-rate importance sampling: Have you experimented with non-uniform distributions? A cosine or log-snr schedule [Kingma et al. 2021] may tighten the ELBO and stabilize low- gradients.

5. Compositional generalization: On tasks such as SCAN or COGS, does the bidirectional objective aid compositional generalization, or does the shortcut learning hypothesis [McCoy et al. 2019] apply equally to diffusion LMs?

LLaDA sits at the intersection of three lines: (1) discrete diffusion [Austin et al. 2021; Lou et al. 2024; Shi et al. 2024; Sahoo et al. 2024]; (2) masked language modeling [Devlin et al. 2019], with the crucial innovation that the mask rate is randomized and the loss reweighted to yield a proper likelihood bound; and (3) the AR-scaling tradition exemplified by [Kaplan et al. 2020] and [Hoffmann et al. 2022]. The closest scaling precedent is arguably the parallel-decoding MaskGIT line [Chang et al. 2022] in vision, which established that iterative masked decoding can produce high-quality samples with far fewer steps than pixel-space diffusion. LLaDA extends this intuition to language at 8B scale, which is the genuinely new artifact.

8. Broader Impact

If LLaDA's scaling claim holds up under independent replication, the most consequential implication is not any single benchmark number. It is that the AR inductive bias, around which the entire LLM infrastructure stack is optimized, may not be load-bearing for reasoning quality. That would open the door to architectures with genuinely different inference characteristics: infilling-native models for code editing, bidirectional retrieval-augmented generation, and controllable generation with constraint satisfaction baked into the sampling loop. Language is more than prediction. Whether diffusion is the right escape from next-token tyranny, or merely a parallel path of comparable capability, is the research question the community should prioritize next.

I am optimistic about the direction and honest about the evidence: LLaDA is a serious demonstration, not a settled answer. The scaling result is real; the inductive-bias result requires controlled ablations to be believed. I would rate the empirical contribution moderate-to-significant, with the understanding that the theoretical novelty is low (the loss is SMD) and that the headline interpretation remains, at present, undersupported by controls.

Reproducibility & Sources

Primary paper: Nie, S. Zhu, F. You, Z. Zhang, X. Ou, J. Hu, J. Zhou, J. Lin, Y. Wen, J.-R. Li, C. (2025). *Large Language Diffusion Models*. arXiv:2502.09992.

Code repository: The authors indicate a code release at the project's GitHub (verify via the paper's abstract page on arXiv for the current link). Trained checkpoints (LLaDA 8B Base and Instruct) are reportedly released under a research license.

Datasets: The pretraining corpus is the authors' assembly of web data, code, and math-heavy sources (2.3T tokens), not publicly released as a single artifact. Evaluation datasets are public: MMLU [Hendrycks et al. 2021], GSM8K [Cobbe et al. 2021], HumanEval [Chen et al. 2021], BBH [Suzgun et al. 2023]. Access is via HuggingFace Datasets or the benchmarks' original repositories.

Reproducibility assessment:

AxisRating (1-5)Justification
Code availability4Official implementation and inference code released; training code coverage should be verified.
Data availability2Pretraining corpus is not released as a single artifact; evaluation datasets are public. Reproducing pretraining requires reconstructing the data mixture.
Experimental detail3Main hyperparameters and training schedule are documented, but sampler ablations, seed variance, and data mixture proportions remain underspecified.

Reproducing the full scaling claim at 8B requires roughly the pretraining compute of a LLaMA3-class run (on the order of FLOPs), placing independent verification out of reach for most academic labs. Verifying the sampler behavior, reversal-reasoning claims, and inference-speed claims at the released checkpoint scale is, by contrast, tractable, and is the most productive next step for the community.