pith. machine review for the scientific record. sign in

arxiv: 2604.18673 · v1 · submitted 2026-04-20 · ✦ hep-th

Recognition: unknown

Neural Networks Reveal a Universal Bias in Conformal Correlators

Authors on Pith no claims yet

Pith reviewed 2026-05-10 04:27 UTC · model grok-4.3

classification ✦ hep-th
keywords neural networksconformal field theorycrossing symmetrycorrelatorsspectral biasnon-perturbative QFTvariational methods
0
0 comments X

The pith

Simple neural networks trained on crossing symmetry reconstruct conformal correlators from minimal input data.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper establishes that basic neural networks, when trained to enforce crossing symmetry, can recover conformal correlators restricted to a line with high accuracy using only an external scaling dimension, a spectral gap, and the correlator value at one point. This holds across many conformal theories and spacetime dimensions for both four-point functions and thermal two-point functions. The authors link the result to a match between the spectral bias of gradient descent training and an intrinsic smoothness property of conformal field theory, which in turn points toward a variational principle that could serve as a new computational tool for non-perturbative quantum field theory.

Core claim

We propose that simple neural networks (NNs) trained on crossing symmetry can reconstruct conformal correlators restricted to a line to remarkable accuracy. The input is minimal: an external scaling dimension, a spectral gap, and the value of the correlator at a single point. We present evidence across a wide range of conformal theories and dimensions, for both four-point and thermal two-point functions. We attribute these observations to the spectral bias of gradient-based NN training, which appears to align with an intrinsic smoothness property of conformal field theory. This suggests a novel variational principle for conformal correlators and opens a path towards a powerful new framework.

What carries the argument

Gradient-based neural network training whose spectral bias aligns with the smoothness property of conformal correlators, allowing reconstruction from crossing symmetry plus minimal data.

If this is right

  • The method provides a variational principle that computes conformal correlators without solving the full crossing equations directly.
  • Reconstruction accuracy persists across different spacetime dimensions and for both four-point and thermal two-point functions.
  • Only an external scaling dimension, spectral gap, and one correlator value are required as input.
  • The approach opens a computational route to non-perturbative quantities in quantum field theory.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The observed alignment may indicate that conformal data possess a hidden low-frequency structure that gradient descent naturally discovers.
  • Similar networks could be tested on other constraints such as unitarity bounds or modular invariance to see whether the same bias operates.
  • If the smoothness match is universal, the technique might extend to correlators in non-conformal theories with analogous analytic properties.

Load-bearing premise

The spectral bias that arises in gradient-based neural network training aligns with an intrinsic smoothness property of conformal field theory.

What would settle it

Train the network on crossing symmetry with the stated minimal inputs and check whether reconstruction error remains low for a known exact correlator such as the Ising four-point function; a sharp rise in error would falsify the claim.

Figures

Figures reproduced from arXiv: 2604.18673 by Andreas Stergiou, Kausik Ghosh, Sidhaarth Kumar, Vasilis Niarchos.

Figure 1
Figure 1. Figure 1: FIG. 1 [PITH_FULL_IMAGE:figures/full_fig_p004_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: FIG. 2: Thermal two-point function [PITH_FULL_IMAGE:figures/full_fig_p005_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: FIG. 3: Collected plots of anchored neural network predictions against analytic results. [PITH_FULL_IMAGE:figures/full_fig_p007_3.png] view at source ↗
read the original abstract

We propose that simple neural networks (NNs) trained on crossing symmetry can reconstruct conformal correlators restricted to a line to remarkable accuracy. The input is minimal: an external scaling dimension, a spectral gap, and the value of the correlator at a single point. We present evidence across a wide range of conformal theories and dimensions, for both four-point and thermal two-point functions. We attribute these observations to the spectral bias of gradient-based NN training, which appears to align with an intrinsic smoothness property of conformal field theory. This suggests a novel variational principle for conformal correlators and opens a path towards a powerful new computational framework for non-perturbative quantum field theory.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The manuscript proposes that simple neural networks trained to enforce crossing symmetry can reconstruct conformal correlators restricted to a line to high accuracy. The inputs consist of an external scaling dimension, a spectral gap, and the correlator value at a single point. Evidence is claimed across a range of CFTs and dimensions for both four-point functions and thermal two-point functions. The success is attributed to the spectral bias of gradient-based NN training aligning with an intrinsic smoothness property of CFTs, suggesting a novel variational principle for conformal correlators.

Significance. If the central claims are substantiated, the work could introduce a new variational framework for computing non-perturbative CFT correlators with minimal input data, potentially offering computational advantages in regimes where bootstrap or perturbative methods are limited. The interdisciplinary link between NN training dynamics and CFT structure would be noteworthy if isolated from generic approximation effects.

major comments (3)
  1. [Abstract] Abstract: The claims of 'remarkable accuracy' and 'evidence across a wide range' are not accompanied by quantitative metrics (e.g., mean squared error, maximum deviation, or error bars), training protocols, data selection criteria, or network hyperparameters, preventing verification of the central reconstruction results.
  2. [Results] Results section: The attribution of accuracy to spectral bias in gradient-based training lacks isolating controls; no comparisons are presented to other approximators (such as Chebyshev polynomials or Fourier bases) that minimize the identical crossing-violation loss on the same low-dimensional inputs, leaving open whether the outcome stems from NN-specific bias or from the smoothness of crossing-constrained 1D functions.
  3. [Methods] Methods: The formulation of the loss function used to enforce crossing symmetry during training is not specified in sufficient detail (e.g., how the crossing equation is discretized or weighted), which is load-bearing for reproducing the reported reconstructions and for assessing whether the procedure is parameter-free as implied.
minor comments (2)
  1. Clarify the precise neural network architecture (number of layers, neurons per layer, activation functions) and the range of external dimensions and gaps explored in the numerical experiments.
  2. [Discussion] The discussion of the 'universal bias' would benefit from explicit comparison to the established spectral bias literature cited, to distinguish the CFT-specific alignment from general ML observations.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for their careful reading and constructive feedback. We address each major comment below and will revise the manuscript to improve reproducibility, add controls, and clarify methodological details.

read point-by-point responses
  1. Referee: [Abstract] Abstract: The claims of 'remarkable accuracy' and 'evidence across a wide range' are not accompanied by quantitative metrics (e.g., mean squared error, maximum deviation, or error bars), training protocols, data selection criteria, or network hyperparameters, preventing verification of the central reconstruction results.

    Authors: We agree that quantitative metrics and experimental details are necessary for verification. In the revised manuscript we will add a new table in the Results section reporting mean squared errors, maximum deviations, and error bars (from multiple random seeds) for each CFT and dimension considered. We will also include a supplementary section listing the exact network architectures, learning rates, epoch counts, data selection criteria, and training protocols used throughout the study. revision: yes

  2. Referee: [Results] Results section: The attribution of accuracy to spectral bias in gradient-based training lacks isolating controls; no comparisons are presented to other approximators (such as Chebyshev polynomials or Fourier bases) that minimize the identical crossing-violation loss on the same low-dimensional inputs, leaving open whether the outcome stems from NN-specific bias or from the smoothness of crossing-constrained 1D functions.

    Authors: This criticism is well-taken; the present manuscript does not contain such isolating controls. In the revision we will add direct comparisons in which Chebyshev polynomials and Fourier bases are optimized to minimize the identical crossing-violation loss on the same minimal inputs. These controls will quantify whether the reported accuracy is attributable to the spectral bias of gradient-based NN training or is a generic consequence of smoothness under crossing constraints. revision: yes

  3. Referee: [Methods] Methods: The formulation of the loss function used to enforce crossing symmetry during training is not specified in sufficient detail (e.g., how the crossing equation is discretized or weighted), which is load-bearing for reproducing the reported reconstructions and for assessing whether the procedure is parameter-free as implied.

    Authors: We acknowledge the omission. The loss is the mean-squared violation of the crossing equation, evaluated on a uniform grid of 200 points in z ∈ [0,1] with equal weighting. The procedure depends on this discretization choice and is therefore not strictly parameter-free. In the revised Methods section we will state the precise loss functional, the grid size, the weighting scheme, and a brief sensitivity analysis with respect to grid resolution. revision: yes

Circularity Check

0 steps flagged

No significant circularity; empirical NN reconstruction uses external crossing constraint as independent input.

full rationale

The paper's derivation chain consists of training NNs to minimize a crossing-violation loss on functions parameterized by external scaling dimension, spectral gap, and one-point value, then observing accurate reconstruction of known CFT correlators. This is not self-definitional or a fitted input renamed as prediction, because crossing symmetry is an external CFT axiom enforced during training rather than derived from the NN outputs. Attribution to spectral bias references independent machine-learning literature on gradient descent preferences, without reducing to a self-citation chain or ansatz smuggled from prior author work. No uniqueness theorem or renaming of known results is invoked as load-bearing. The central claim remains an empirical observation verifiable against external CFT data and is self-contained against benchmarks.

Axiom & Free-Parameter Ledger

0 free parameters · 2 axioms · 0 invented entities

The central claim rests on two standard domain assumptions from CFT and machine learning; no new free parameters or invented entities are introduced in the abstract.

axioms (2)
  • domain assumption Crossing symmetry is a fundamental constraint on conformal correlators
    Standard axiom of conformal field theory invoked to train the networks.
  • domain assumption Gradient-based neural network training exhibits spectral bias toward smooth functions
    Established property of neural networks used to explain the observed accuracy.

pith-pipeline@v0.9.0 · 5413 in / 1285 out tokens · 52614 ms · 2026-05-10T04:27:10.225794+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 2 Pith papers

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Descending into the Modular Bootstrap

    hep-th 2026-04 unverdicted novelty 7.0

    Machine-learning optimization produces candidate truncated modular-invariant partition functions for 2d CFTs in the central-charge window 1 to 8/7, indicating a continuous solution space and a stricter spectral-gap bo...

  2. Neural Networks, Dispersion Relations and the Thermal Bootstrap

    hep-th 2026-05 unverdicted novelty 4.0

    A neural-network approach with dispersion relations handles infinite OPE towers in thermal conformal correlators without positivity.

Reference graph

Works this paper leans on

44 extracted references · 30 canonical work pages · cited by 2 Pith papers · 1 internal anchor

  1. [1]

    Bubble diagram of scalarϕ 4 in AdS 2 The exact one-loop Witten diagram in scalarϕ 4 theory in AdS2 for ∆ϕ = 1 is given by [38] Gbubble(z) = 9 250 (z−1) 2 " π4z2 6−(z−2)z + 15 4z3 arctanh(1−2z) + (2z−1) log 4(1−z) + 2z2 logz+ 4(1−2z) log 3(1−z) logz+ 2 log(1−z) z−2z 2 +π 2(z−1) 2(2 +z 2) logz + 2 log2(1−z) π2(2z−1)−3(z−1) 2(1 +z 2) log2 z −180(z−1) 2 (z2 −...

  2. [2]

    2d minimal model correlators on a line The reduced four-point function of the primary field ϕ1,2 in the 2d minimal modelM(p,p ′), with scaling di- mension ∆ϕ = 3 2 p p′ −1, is [39] G(z) = (G1(z))2 +N(∆ ϕ) (G2(z))2 , (A2) with G1(z) =(1−z) −∆ϕ × 2F1 1 3(1−2∆ ϕ),−2∆ ϕ; 2 3(1−2∆ ϕ);z , G2(z) =(1−z) 1+∆ϕ 3 z 1+4∆ϕ 3 × 2F1 2 3(1 + ∆ϕ), 1 + 2∆ϕ; 4 3(1 + ∆ϕ);z ,...

  3. [3]

    De- noting the spacetime coordinates as (τ,⃗ x), the thermal two-point function of a scalar primary operator depends only onτand|x|= √ τ 2 +⃗ x2: g(τ,|x|) =⟨ϕ(τ,⃗ x)ϕ(0, 0)⟩ β

    Thermal two-point functions For a finite-temperatured-dimensional CFT onS 1 β × Rd−1 the Euclidean timeτ∈[0,β) is compactified. De- noting the spacetime coordinates as (τ,⃗ x), the thermal two-point function of a scalar primary operator depends only onτand|x|= √ τ 2 +⃗ x2: g(τ,|x|) =⟨ϕ(τ,⃗ x)ϕ(0, 0)⟩ β . (A4) Using theSO(d−1) rotational symmetry of the CF...

  4. [4]

    The superconformal Ward identities can be used to express the four-point correlator in terms of a single functionH(z, ¯z) of the conformal cross-ratios

    Correlators of half-BPS operators in 4dN= 4 SYM In the main text we considered four-point correlation functions of the half-BPS superconformal primaries in the20 ′. The superconformal Ward identities can be used to express the four-point correlator in terms of a single functionH(z, ¯z) of the conformal cross-ratios. In a large- cexpansion,H(z) =H (0)(z)+ ...

  5. [5]

    The Conformal Bootstrap: Theory, Numerical Techniques, and Applications

    D. Poland, S. Rychkov, and A. Vichi, The Confor- mal Bootstrap: Theory, Numerical Techniques, and Applications, Rev. Mod. Phys.91, 015002 (2019), 8 arXiv:1805.04405 [hep-th]

  6. [6]

    The analytic functional bootstrap. Part I: 1D CFTs and 2D S- matrices

    D. Mazac and M. F. Paulos, The analytic functional bootstrap. Part I: 1D CFTs and 2D S-matrices, JHEP 02, 162, arXiv:1803.10233 [hep-th]

  7. [7]

    Ghosh, M

    K. Ghosh, M. F. Paulos, and N. Suchel, Solving 1D cross- ing and QFT2/CFT1, (2025), arXiv:2503.22798 [hep-th]

  8. [8]

    M. F. Paulos, Dispersion relations and exact bounds on CFT correlators, JHEP08, 166, arXiv:2012.10454 [hep- th]

  9. [9]

    Ghosh, S

    K. Ghosh, S. Kumar, V. Niarchos, and A. Stergiou, Neu- ral Spectral Bias and Conformal Correlators I: Introduc- tion and Applications (2026), in preparation

  10. [10]

    Ghosh, S

    K. Ghosh, S. Kumar, V. Niarchos, and A. Stergiou, Neu- ral Spectral Bias and Conformal Correlators III: Boot- strability and Holomorphic Bootstrap, (2026), in prepa- ration

  11. [11]

    Kubo, Statistical-mechanical theory of irreversible processes

    R. Kubo, Statistical-mechanical theory of irreversible processes. i. general theory and simple applications to magnetic and conduction problems, Journal of the Phys- ical Society of Japan12, 570 (1957)

  12. [12]

    P. C. Martin and J. Schwinger, Theory of many-particle systems. i, Physical Review115, 1342 (1959)

  13. [13]

    Barrat, E

    J. Barrat, E. Marchetto, A. Miscioscia, and E. Pomoni, Thermal Bootstrap for the Critical O(N) Model, Phys. Rev. Lett.134, 211604 (2025), arXiv:2411.00978 [hep- th]

  14. [14]

    Barrat, D

    J. Barrat, D. N. Bozkurt, E. Marchetto, A. Miscioscia, and E. Pomoni, The analytic bootstrap at finite temper- ature, (2025), arXiv:2506.06422 [hep-th]

  15. [15]

    The Conformal Bootstrap at Finite Temperature

    L. Iliesiu, M. Kologlu, R. Mahajan, E. Perlmutter, and D. Simmons-Duffin, The Conformal Bootstrap at Finite Temperature, JHEP10, 070, arXiv:1802.10266 [hep-th]

  16. [16]

    A. C. Petkou and A. Stergiou, Dynamics of Finite- Temperature Conformal Field Theories from Operator Product Expansion Inversion Formulas, Phys. Rev. Lett. 121, 071602 (2018), arXiv:1806.02340 [hep-th]

  17. [17]

    Ghosh, S

    K. Ghosh, S. Kumar, V. Niarchos, and A. Stergiou, Neu- ral Spectral Bias and Conformal Correlators II: Modular and Annulus Bootstrap, (2026), in preparation

  18. [18]

    Niarchos, C

    V. Niarchos, C. Papageorgakis, A. Stratoudakis, and M. Woolley, Deep finite temperature bootstrap, Phys. Rev. D112, 126012 (2025), arXiv:2508.08560 [hep-th]

  19. [19]

    Cybenko, Approximation by superpositions of a sig- moidal function, Mathematics of Control, Signals, and Systems2, 303 (1989)

    G. Cybenko, Approximation by superpositions of a sig- moidal function, Mathematics of Control, Signals, and Systems2, 303 (1989)

  20. [20]

    Hornik, M

    K. Hornik, M. Stinchcombe, and H. White, Multilayer feedforward networks are universal approximators, Neu- ral Networks2, 359 (1989)

  21. [21]

    Funahashi, On the approximate realization of con- tinuous mappings by neural networks, Neural Networks 2, 183 (1989)

    K.-I. Funahashi, On the approximate realization of con- tinuous mappings by neural networks, Neural Networks 2, 183 (1989)

  22. [22]

    Leshno, V

    M. Leshno, V. Y. Lin, A. Pinkus, and S. Schocken, Mul- tilayer feedforward networks with a nonpolynomial ac- tivation function can approximate any function, Neural Networks6, 861 (1993)

  23. [23]

    D. P. Kingma and J. Ba, Adam: A Method for Stochastic Optimization,3rd International Conference on Learning Representations (ICLR 2015), arXiv:1412.6980 [cs.LG]

  24. [24]

    On the Spectral Bias of Neural Networks

    N. Rahaman, A. Baratin, D. Arpit, F. Draxler, M. Lin, F. A. Hamprecht, Y. Bengio, and A. Courville, On the Spectral Bias of Neural Networks,Proceedings of the 36th International Conference on Machine Learning, PMLR, 97, 5301 (2019), arXiv:1806.08734 [stat.ML]

  25. [25]

    Z.-Q. J. Xu, Y. Zhang, T. Luo, Y. Xiao, and Z. Ma, Frequency Principle: Fourier Analysis Sheds Light on Deep Neural Networks, Communications in Computa- tional Physics28, 1746 (2020), arXiv:1901.06523 [cs.LG]

  26. [26]

    Neural tangent kernel: Convergence and generalization in neural networks

    A. Jacot, F. Gabriel, and C. Hongler, Neural Tangent Kernel: Convergence and Generalization in Neural Net- works,Advances in Neural Information Processing Sys- tems 31, , 8571 (2018), arXiv:1806.07572 [cs.LG]

  27. [27]

    On the Inductive Bias of Neural Tangent Kernels

    A. Bietti and J. Mairal, On the Inductive Bias of Neural Tangent Kernels,Advances in Neural Information Pro- cessing Systems 32, , 12897 (2019), arXiv:1905.12173 [stat.ML]

  28. [28]

    Gunasekar, J

    S. Gunasekar, J. Lee, D. Soudry, and N. Srebro, Char- acterizing implicit bias in terms of optimization geome- try, inProceedings of the 35th International Conference on Machine Learning, Proceedings of Machine Learn- ing Research, Vol. 80 (PMLR, 2018) pp. 1832–1841, arXiv:1802.08246

  29. [29]

    Arora, S

    S. Arora, S. S. Du, W. Hu, Z. Li, and R. Wang, Fine- grained analysis of optimization and generalization for overparameterized two-layer neural networks, inProceed- ings of the 36th International Conference on Machine Learning, Proceedings of Machine Learning Research, Vol. 97 (PMLR, 2019) pp. 322–332, arXiv:1901.08584

  30. [30]

    King’s College London, King’s Computational Research, Engineering and Technology Environment (CREATE), (2022)

  31. [31]

    F. Kos, D. Poland, and D. Simmons-Duffin, Boot- strapping theO(N) vector models, JHEP06, 091, arXiv:1307.6856 [hep-th]

  32. [32]

    F. Kos, D. Poland, D. Simmons-Duffin, and A. Vichi, Precision Islands in the Ising andO(N) Models, JHEP 08, 036, arXiv:1603.04436 [hep-th]

  33. [33]

    The Lightcone Bootstrap and the Spectrum of the 3d Ising CFT

    D. Simmons-Duffin, The Lightcone Bootstrap and the Spectrum of the 3d Ising CFT, JHEP03, 086, arXiv:1612.08471 [hep-th]

  34. [34]

    Chang, V

    C.-H. Chang, V. Dommes, R. S. Erramilli, A. Homrich, P. Kravchuk, A. Liu, M. S. Mitchell, D. Poland, and D. Simmons-Duffin, Bootstrapping the 3d Ising stress tensor, JHEP03, 136, arXiv:2411.15300 [hep-th]

  35. [35]

    W. Zhu, C. Han, E. Huffman, J. S. Hofmann, and Y.-C. He, Uncovering Conformal Symmetry in the 3D Ising Transition: State-Operator Correspondence from a Quantum Fuzzy Sphere Regularization, Phys. Rev. X13, 021009 (2023), arXiv:2210.13482 [cond-mat.stat-mech]

  36. [36]

    C. Han, L. Hu, W. Zhu, and Y.-C. He, Conformal four- point correlators of the three-dimensional Ising transition via the quantum fuzzy sphere, Phys. Rev. B108, 235123 (2023), arXiv:2306.04681 [cond-mat.stat-mech]

  37. [37]

    Operator Product Expansion Coefficients of the 3D Ising Criti- cality via Quantum Fuzzy Spheres

    L. Hu, Y.-C. He, and W. Zhu, Operator Product Ex- pansion Coefficients of the 3D Ising Criticality via Quan- tum Fuzzy Spheres, Phys. Rev. Lett.131, 031601 (2023), arXiv:2303.08844 [cond-mat.stat-mech]

  38. [38]

    Halverson, A

    J. Halverson, A. Maiti, and K. Stoner, Neural Networks and Quantum Field Theory, Mach. Learn. Sci. Tech.2, 035002 (2021), arXiv:2008.08601 [cs.LG]

  39. [39]

    Demirtas, J

    M. Demirtas, J. Halverson, A. Maiti, M. D. Schwartz, and K. Stoner, Neural network field theories: non- Gaussianity, actions, and locality, Mach. Learn. Sci. Tech.5, 015002 (2024), arXiv:2307.03223 [hep-th]

  40. [40]

    Halverson, J

    J. Halverson, J. Naskar, and J. Tian, Conformal fields from neural networks, JHEP10, 039, arXiv:2409.12222 [hep-th]

  41. [41]

    Ferko, J

    C. Ferko, J. Halverson, and A. Mutchler, Universality of Neural Network Field Theory, (2026), arXiv:2601.14453 9 [hep-th]

  42. [42]

    The analytic functional bootstrap. Part II. Natural bases for the crossing equation,

    D. Mazac and M. F. Paulos, The analytic functional bootstrap. Part II. Natural bases for the crossing equa- tion, JHEP02, 163, arXiv:1811.10646 [hep-th]

  43. [43]

    The bootstrap program for boundary CFTd,

    P. Liendo, L. Rastelli, and B. C. van Rees, The Boot- strap Program for Boundary CFT d, JHEP07, 113, arXiv:1210.4258 [hep-th]

  44. [44]

    F. A. Dolan, M. Nirschl, and H. Osborn, Conjectures for large N superconformal N=4 chiral primary four point functions, Nucl. Phys. B749, 109 (2006), arXiv:hep- th/0601148