pith. machine review for the scientific record. sign in

arxiv: 2605.08565 · v1 · submitted 2026-05-08 · 💻 cs.LG

Recognition: 1 theorem link

· Lean Theorem

Finer is Better (with the Right Scaling)

Authors on Pith no claims yet

Pith reviewed 2026-05-12 01:06 UTC · model grok-4.3

classification 💻 cs.LG
keywords microscalingquantizationlarge language modelsFP4 formatsblock sizescaling factorsheavy-tailed distributionsperplexity
0
0 comments X

The pith

Finer microscaling blocks reduce quantization error in LLMs once scaling prevents underflow and corrects large values in heavy-tailed tensors.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper demonstrates that the recently observed paradox—where finer block sizes degrade LLM quality under standard abs-max scaling—is not intrinsic to smaller blocks. Instead the degradation arises when heavy-tailed value distributions interact with the coarse upper quantization bins of FP4 element formats. By stopping scaling factors from underflowing to zero and applying targeted corrections such as the 4-over-6 method for large elements, the authors restore the expected behavior: theoretical mean squared error strictly decreases as blocks become finer. A brute-force search supplies an optimal baseline that confirms this improvement, and the same recipe lets ordinary hardware formats like OCP E4M3 match the accuracy of custom wider-exponent formats like UE5M3 on multiple large models.

Core claim

The central claim is that microscaling degradation at small block sizes is fixable rather than fundamental. Preventing scaling-factor underflow and applying the 4-over-6 correction for large elements restores the theoretical MSE reduction that should accompany finer granularity, allowing standard E4M3 formats to reach the same downstream perplexity as UE5M3 while producing robust improvements across several LLMs.

What carries the argument

The combination of underflow prevention for scaling factors and the 4-over-6 methodology that adjusts quantization geometry for the largest elements in heavy-tailed tensors.

If this is right

  • Standard hardware-compliant FP4 formats achieve parity with custom wider-exponent formats on downstream perplexity.
  • Theoretical MSE decreases monotonically with smaller block sizes once the corrections are in place.
  • The block-size paradox is eliminated for the tested large language models.
  • Perplexity gains remain robust across multiple model scales and architectures.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • Hardware implementations could safely adopt finer block sizes if they also support the simple scaling corrections.
  • Analogous mismatches between value distributions and quantization bins may limit other ultra-low-precision schemes beyond FP4.
  • Applying the same recipe to emerging model families or new data types could yield further perplexity reductions.

Load-bearing premise

Heavy-tailed tensor distributions are the main driver of the observed degradation and the listed algorithmic fixes resolve it without creating new dominant errors.

What would settle it

A direct computation of quantization MSE on representative heavy-tailed tensors that still fails to show improvement as block size decreases after underflow prevention and 4-over-6 correction are applied would falsify the resolution of the paradox.

Figures

Figures reproduced from arXiv: 2605.08565 by Clemens Schaefer, Gil Tabak.

Figure 1
Figure 1. Figure 1: Comparing MSE against distribution scale behavior. [PITH_FULL_IMAGE:figures/full_fig_p003_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: (a,b,d,e) show the perplexity gap of the Granite-3.3-8B, Llama-3.1-8B, DeepSeek-7B, and Qwen2.5-14B models with various FP4 formats compared to a unquantized baseline. Note that only Granite and Llama demonstrate the paradoxical increased perplexity gap at lower block sizes, meanwhile DeepSeek and Qwen improve the perplexity with smaller block sizes. Techniques such as hierarchical scales (H), four over si… view at source ↗
Figure 3
Figure 3. Figure 3: Standard Abs-Max Scaling. (a) Region A: At extremely small scales, the standard algorithm frequently selects a scaling factor of zero, obliterating element information and causing a sharp MSE spike. (b) Regions B and C: At these ranges, we find smaller blocks have more entries at the bin of magnitude 6. At Region B the scales are relatively concentrated at only a few bins, while at Region C they are spread… view at source ↗
Figure 4
Figure 4. Figure 4: Prevent-Zero Adjustment. (a) Region A: Restricting the scaling factor to the smallest positive representable value avoids a collapse to zero, preserving information and eliminating the localized MSE spike. (b) Region B: Elements remain pushed into the coarse quantization gap between 4 and 6, visually confirming that preventing zero scales does not resolve the structural element-level anomaly. (c) Region C:… view at source ↗
Figure 5
Figure 5. Figure 5: 4-over-6 Methodology. (a) Region A: While providing better overall bin utilization, the scaling factors can still struggle with extreme underflow in this tiny magnitude regime. (b) Region B: By adaptively allowing a maximum target scale of 4, the algorithm successfully avoids forcing elements into the massive upper quantization gap of FP4, resolving the scaling inversion paradox. (c) Region C: Similar to R… view at source ↗
Figure 6
Figure 6. Figure 6: Brute-Force Search. (a) Region A: The scaling factors chosen align with the prevent-zero adjustment. Choosing a larger scaling value would result in underflowing too many entries (b) Region B: The scales show a somewhat wider distribution compared to 4-over-6, and significantly larger compared to abs-max. Although for smaller blocks a larger portion of entries fall into the magnitude 6 bin it is much less … view at source ↗
Figure 7
Figure 7. Figure 7: Clipping only MSE. This plot is analogous to Figure 1, showing only the clipping portion. Using 4-over-6 or brute [PITH_FULL_IMAGE:figures/full_fig_p007_7.png] view at source ↗
Figure 8
Figure 8. Figure 8: Granite 3.3 8B sample weights for layer 2 and layer 25 down projections. The distributions are similar at both parts of [PITH_FULL_IMAGE:figures/full_fig_p007_8.png] view at source ↗
Figure 9
Figure 9. Figure 9: Qwen 2.5 14B sample weights for layer 2 and layer 25 down projections. Earlier layers (a) show many columns with [PITH_FULL_IMAGE:figures/full_fig_p008_9.png] view at source ↗
Figure 10
Figure 10. Figure 10: Unquantized ablation with mapping of specific entry value ranges to zero (between lower and upper thresholds specified [PITH_FULL_IMAGE:figures/full_fig_p008_10.png] view at source ↗
read the original abstract

Microscaling is a critical technique for preserving the quality of Large Language Models (LLMs) quantized to ultra-low precision formats. Intuitively, finer block sizes should yield lower quantization error; however, a paradox recently identified in the literature demonstrates that standard abs-max scaling can actually degrade model quality as block sizes shrink. In this work, we investigate the underlying mechanics of this phenomenon. We demonstrate that this degradation is not an inherent limitation of finer granularity, but is primarily driven by heavy-tailed tensor distributions interacting poorly with the coarse upper quantization bins of the FP4 element format. Specifically, we show that i) preventing the scaling factor from underflowing to zero mitigates localized errors, ii) targeted algorithmic interventions like the 4-over-6 methodology effectively correct the quantization geometry for large elements, and iii) a brute-force search establishes an optimal baseline, confirming that the theoretical Mean Squared Error (MSE) strictly improves with finer block sizes. Ultimately, our findings reveal a valuable interchangeability: applying the correct algorithmic recipe allows standard, hardware-compliant formats (like OCP E4M3) to match the performance of custom, wider-exponent formats (like UE5M3). We validate these results across several large language models, fully resolving the block size paradox and achieving robust downstream perplexity improvements.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The paper investigates the 'block size paradox' in microscaling quantization for LLMs, where finer blocks unexpectedly degrade quality under standard abs-max scaling. It attributes this not to finer granularity itself but to heavy-tailed tensor distributions clashing with coarse upper bins in FP4 formats like E4M3. The authors propose fixes including preventing scaling-factor underflow, a '4-over-6' algorithmic intervention to correct geometry for large elements, and a brute-force search establishing an optimal baseline that confirms theoretical MSE strictly improves with finer blocks. They conclude that these allow standard hardware-compliant E4M3 to match custom wider-exponent UE5M3 performance, with validation across LLMs yielding perplexity gains and resolving the paradox.

Significance. If the central claims hold after verification, the work would be significant for low-precision LLM deployment: it provides a practical recipe to exploit finer blocks for lower error while using existing hardware formats, potentially improving efficiency without custom silicon. The brute-force optimal baseline and cross-format interchangeability result would be particularly useful if accompanied by reproducible code or parameter-free derivations.

major comments (2)
  1. [Abstract] Abstract: the assertion that 'the theoretical Mean Squared Error (MSE) strictly improves with finer block sizes' is presented as confirmed by brute-force search, but without an explicit derivation or comparison showing that the post-intervention error distribution matches the theoretical improvement across the full range of heavy-tailed tensor statistics in LLMs, the claim that interventions fully resolve the paradox without shifting error mass elsewhere remains unverified and load-bearing for the E4M3/UE5M3 interchangeability result.
  2. [Abstract] Abstract: the weakest assumption—that heavy-tailed distributions interacting with FP4 upper bins are the primary (and fixable) cause, with 4-over-6 plus underflow prevention resolving it without new errors—is not supported by shown analysis of how the fixes alter scaling dynamics or error geometry for large elements; this directly underpins the downstream perplexity improvements and must be demonstrated rather than asserted.
minor comments (2)
  1. [Abstract] The abstract references validation 'across several large language models' but provides no details on model sizes, datasets, or exact perplexity deltas; adding these would strengthen reproducibility.
  2. Notation for formats (E4M3, UE5M3, 4-over-6) is introduced without prior definition or reference to standards (e.g., OCP microscaling spec); a brief clarification in the introduction would aid readers.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for their thoughtful and constructive review of our manuscript on resolving the block size paradox in microscaling quantization. We address each major comment below in detail. To strengthen the verification of our central claims, we have revised the manuscript with additional derivations, error distribution comparisons, and mechanistic analyses as outlined in the responses.

read point-by-point responses
  1. Referee: [Abstract] Abstract: the assertion that 'the theoretical Mean Squared Error (MSE) strictly improves with finer block sizes' is presented as confirmed by brute-force search, but without an explicit derivation or comparison showing that the post-intervention error distribution matches the theoretical improvement across the full range of heavy-tailed tensor statistics in LLMs, the claim that interventions fully resolve the paradox without shifting error mass elsewhere remains unverified and load-bearing for the E4M3/UE5M3 interchangeability result.

    Authors: We agree that an explicit link between the theoretical MSE and post-intervention results strengthens the paper. In the revision, we have added a derivation of the theoretical MSE under adjusted scaling for heavy-tailed distributions, along with direct comparisons of error distributions across a range of tail heaviness representative of LLM tensors. These show that the interventions align the observed errors with the theoretical improvement without shifting error mass. The brute-force search, performed over multiple block sizes and tensor statistics drawn from LLMs, serves as empirical confirmation of the strict improvement, supporting the E4M3/UE5M3 interchangeability. revision: yes

  2. Referee: [Abstract] Abstract: the weakest assumption—that heavy-tailed distributions interacting with FP4 upper bins are the primary (and fixable) cause, with 4-over-6 plus underflow prevention resolving it without new errors—is not supported by shown analysis of how the fixes alter scaling dynamics or error geometry for large elements; this directly underpins the downstream perplexity improvements and must be demonstrated rather than asserted.

    Authors: We acknowledge the value of demonstrating the mechanistic effects rather than asserting them. The revised manuscript includes expanded analysis of scaling dynamics, illustrating how underflow prevention stabilizes scaling factors for fine blocks and how the 4-over-6 intervention corrects bin geometry specifically for large elements in heavy-tailed data. We add visualizations of error geometry pre- and post-intervention, confirming no new errors are introduced. These changes are tied directly to the perplexity gains observed in our LLM evaluations. revision: yes

Circularity Check

0 steps flagged

No circularity; claims rest on empirical interventions, brute-force search, and LLM validation

full rationale

The paper's argument proceeds by diagnosing the block-size paradox via tensor distribution analysis, proposing fixes (underflow prevention, 4-over-6 geometry correction), running brute-force search to establish an optimal baseline that empirically confirms theoretical MSE improvement with finer blocks, and validating downstream perplexity gains across LLMs. No equations or derivations are presented that equate any prediction to a fitted parameter or input by construction. No self-citations appear as load-bearing uniqueness theorems, no ansatzes are smuggled, and no known results are merely renamed. The resolution is demonstrated through external experimental benchmarks rather than reducing to the paper's own assumptions or data fits.

Axiom & Free-Parameter Ledger

0 free parameters · 0 axioms · 0 invented entities

Abstract provides no explicit free parameters, axioms, or invented entities; the work relies on empirical observations of tensor distributions and algorithmic adjustments to existing quantization formats.

pith-pipeline@v0.9.0 · 5521 in / 1051 out tokens · 52857 ms · 2026-05-12T01:06:21.808389+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

11 extracted references · 11 canonical work pages · 1 internal anchor

  1. [1]

    arXiv preprint arXiv:2601.19026 , year=

    Is Finer Better? The Limits of Microscaling Formats in Large Language Models , author=. arXiv preprint arXiv:2601.19026 , year=

  2. [2]

    Four Over Six: More Accurate NVFP4 Quantization with Adaptive Block Scaling

    Four Over Six: More Accurate NVFP4 Quantization with Adaptive Block Scaling , author=. arXiv preprint arXiv:2512.02010 , year=

  3. [3]

    Scaling Learning Algorithms Towards

    Bengio, Yoshua and LeCun, Yann , booktitle =. Scaling Learning Algorithms Towards

  4. [4]

    and Osindero, Simon and Teh, Yee Whye , journal =

    Hinton, Geoffrey E. and Osindero, Simon and Teh, Yee Whye , journal =. A Fast Learning Algorithm for Deep Belief Nets , volume =

  5. [5]

    2016 , publisher=

    Deep learning , author=. 2016 , publisher=

  6. [6]

    2023 , month =

    Bita Darvish Rouhani and Nitin Garegrat and Tom Savell and Ankit More and Kyung-Nam Han and Ritchie Zhao and Mathew Hall and Jasmine Klar and Eric Chung and Yuan Yu and Michael Schulte and Ralph Wittig and Ian Bratt and Nigel Stephens and Jelena Milanovic and John Brothers and Pradeep Dubey and Marius Cornea and Alexander Heinecke and Andres Rodriguez and...

  7. [7]

    arXiv preprint arXiv:2509.25149 , year=

    Pretraining large language models with nvfp4 , author=. arXiv preprint arXiv:2509.25149 , year=

  8. [8]

    Unveiling the potential of quantization with mxfp4: Strategies for quantization error reduction

    Unveiling the Potential of Quantization with MXFP4: Strategies for Quantization Error Reduction , author=. arXiv preprint arXiv:2603.08713 , year=

  9. [9]

    2023 , month =

    Micikevicius, Paulius and Oberman, Stuart and Dubey, Pradeep and Cornea, Marius and Rodriguez, Andres and Bratt, Ian and others , institution =. 2023 , month =

  10. [10]

    2025 , note=

    Introducing. 2025 , note=

  11. [11]

    arXiv preprint arXiv:2509.23202 , year=

    Bridging the gap between promise and performance for microscaling FP4 quantization , author=. arXiv preprint arXiv:2509.23202 , year=