pith. machine review for the scientific record. sign in

arxiv: 2604.18012 · v1 · submitted 2026-04-20 · 💻 cs.LG · cs.NA· math.NA

Recognition: unknown

Neural Shape Operator Surrogates -- Expression Rate Bounds

Authors on Pith no claims yet

Pith reviewed 2026-05-10 04:50 UTC · model grok-4.3

classification 💻 cs.LG cs.NAmath.NA
keywords neural operatorsparametric PDEsshape surrogatesoperator learningholomorphic approximationsdomain variationsurrogate models
0
0 comments X

The pith

Neural and spectral operators approximate shape-to-solution maps for PDEs on varying domains with uniform error bounds and explicit convergence rates.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper shows how to construct neural and spectral operator surrogates for the map from domain shape to PDE solution that work uniformly across an entire family of admissible shapes. The approach pulls every shape-dependent problem back to one fixed reference domain through an affine-parametric encoding of the geometry. This produces a holomorphic family of parametric PDEs on the reference domain whose solutions can be approximated at rates governed by the number of shape parameters. The resulting bounds hold for both neural networks and spectral methods and apply to elliptic, parabolic, and boundary-integral problems.

Core claim

Pulling the PDE back to a fixed reference domain via affine-parametric shape encoding turns the shape-to-solution operator into a holomorphic parametric map on that domain. Under uniform well-posedness, quantified parametric holomorphy then guarantees finite-parametric approximations whose error decays at explicit rates in the number of parameters. These rates transfer directly to neural and spectral operator surrogates, yielding constructive existence proofs with bounds that remain uniform over the whole collection of admissible shapes.

What carries the argument

Affine-parametric pullback of the PDE to a fixed reference domain, which converts geometry variation into a holomorphic parameter dependence whose approximation rates control the neural surrogate error.

If this is right

  • Error bounds and rates for the neural surrogates hold uniformly over all shapes in the admissible family.
  • The same rates apply to spectral operator surrogates and to both elliptic and parabolic problems as well as boundary integral equations.
  • Principal-component encodings of shape and frame-based decoders are admissible within the theory.
  • The results give a theoretical basis for the observed generalization of neural operators across parametric families of geometries.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The same holomorphy argument could be used to derive rates for other operator-learning architectures that admit parametric expansions.
  • Training data drawn from the holomorphic parameter space should in principle suffice for generalization, reducing the need for exhaustive sampling of shapes.
  • The framework suggests a way to certify neural-operator performance a priori once the PDE well-posedness constants and holomorphy radius are known.

Load-bearing premise

The PDE family must remain uniformly well-posed on the reference domain and the solution must be holomorphic in the shape parameters with a quantifiable decay rate for the coefficients.

What would settle it

A concrete family of diffeomorphic domains and an associated elliptic PDE for which no neural operator with a fixed number of parameters can achieve the predicted uniform error bound across the family.

read the original abstract

We prove error bounds for operator surrogates of solution operators for partial differential and boundary integral equations on families of domains which are diffeomorphic to one common reference (or latent) domain $D_{ref}$. The pullback of the PDE to $D_{ref}$ via affine-parametric shape encoding produces a collection of holomorphic parametric PDEs on $D_{ref}$. Sufficient conditions for (uniformly with respect to the parameter) well-posedness are given, implying existence, uniqueness and stability of parametric solution families on $D_{ref}$. We illustrate the abstract hypotheses by reviewing recent holomorphy results for a suite of elliptic and parabolic PDEs. Quantified parametric holomorphy implies existence of finite-parametric, discrete approximations of the parametric solution families with convergence rates in terms of the number $N$ of parameters. We obtain constructive proofs of existence of Neural and Spectral Operator surrogates for the shape-to-solution maps with error bounds and convergence rate guarantees uniform on the collection of admissible shapes. We admit principal-component shape encoders and frame decoders. Our results support in particular the (empirically reported) ability of neural operators to realize data-to-solution maps for elliptic and parabolic PDEs and BIEs that generalize across parametric families of shapes.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The paper proves error bounds and convergence rate guarantees for neural and spectral operator surrogates of the shape-to-solution maps for PDEs and boundary integral equations on families of domains diffeomorphic to a fixed reference domain. Affine-parametric shape encodings pull the problems back to the reference domain, yielding a family of holomorphic parametric PDEs; sufficient conditions for uniform well-posedness are stated and illustrated via existing holomorphy results for elliptic and parabolic problems. These yield constructive existence proofs for the surrogates with rates uniform over admissible shapes, allowing principal-component encoders and frame decoders.

Significance. If the central claims hold, the work supplies the first uniform, constructive approximation-theoretic justification for neural operators that generalize across parametric families of shapes. The explicit rates in terms of the number of parameters and the emphasis on holomorphy-based constructions are strengths that could directly inform architecture choices and sample complexity in data-driven PDE solvers.

major comments (2)
  1. [Abstract] The abstract asserts 'constructive proofs of existence' with 'error bounds and convergence rate guarantees,' yet supplies no explicit derivation of the neural-network or spectral approximation rates, no concrete error estimates, and no verification that the holomorphy radii remain uniform under the chosen affine-parametric encoding. Without these steps the central claim that the surrogates achieve the stated rates remains unverifiable from the given text.
  2. [Abstract (paragraph 1)] The weakest assumption—'sufficient conditions for (uniformly with respect to the parameter) well-posedness' together with 'quantified parametric holomorphy'—is illustrated only by reference to prior results for specific PDEs. No self-contained check is provided that the pullback operators preserve the required holomorphy constants uniformly over the admissible shape collection, which is load-bearing for the uniformity of the surrogate rates.
minor comments (2)
  1. [Abstract] Notation for the reference domain D_ref and the admissible shape collection is introduced without a dedicated preliminary section or table summarizing the standing assumptions on the diffeomorphisms.
  2. [Abstract (last sentence)] The statement that results 'support in particular the (empirically reported) ability of neural operators' would benefit from a brief pointer to the specific empirical works being referenced.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the careful reading and constructive comments on our manuscript. We address each major comment point by point below.

read point-by-point responses
  1. Referee: [Abstract] The abstract asserts 'constructive proofs of existence' with 'error bounds and convergence rate guarantees,' yet supplies no explicit derivation of the neural-network or spectral approximation rates, no concrete error estimates, and no verification that the holomorphy radii remain uniform under the chosen affine-parametric encoding. Without these steps the central claim that the surrogates achieve the stated rates remains unverifiable from the given text.

    Authors: The abstract is a concise summary; the explicit derivations of the neural-network and spectral approximation rates, the concrete error estimates (in terms of parameter count N), and the verification that holomorphy radii remain uniform under the affine-parametric encoding are all contained in the main text. Specifically, Section 2 establishes the uniform holomorphy via the pullback construction and Proposition 2.2, while Sections 3 and 4 derive the rates from standard holomorphic approximation theory, yielding algebraic or exponential convergence uniform over admissible shapes. To improve readability we have added one sentence to the abstract directing readers to these sections for the detailed estimates. revision: partial

  2. Referee: [Abstract (paragraph 1)] The weakest assumption—'sufficient conditions for (uniformly with respect to the parameter) well-posedness' together with 'quantified parametric holomorphy'—is illustrated only by reference to prior results for specific PDEs. No self-contained check is provided that the pullback operators preserve the required holomorphy constants uniformly over the admissible shape collection, which is load-bearing for the uniformity of the surrogate rates.

    Authors: Section 2 formulates the sufficient conditions for uniform well-posedness and quantified parametric holomorphy in an abstract setting that applies directly to the pulled-back problems. The affine-parametric encoding ensures that the pullback operators act continuously on the shape parameter, so that the holomorphy constants (in particular the radius) remain bounded below uniformly over the admissible shape family; this follows from the assumed regularity of the diffeomorphisms and is stated in the proof of Lemma 2.4. The references to prior holomorphy results for concrete elliptic and parabolic PDEs serve only to confirm that the abstract hypotheses are satisfied in practice. To address the request for a more self-contained presentation we have inserted a short remark in Section 2 that summarizes the uniform bound on the holomorphy radius without requiring the reader to consult the cited works. revision: yes

Circularity Check

0 steps flagged

No significant circularity identified

full rationale

The derivation begins from stated sufficient conditions on uniform well-posedness of the pullback parametric PDEs, invokes reviewed (externally established) holomorphy results for concrete elliptic and parabolic problems to obtain quantified parametric holomorphy, and then constructs independent existence proofs and uniform error bounds for the neural and spectral operator surrogates. No equation or claim reduces by definition to a fitted quantity, no prediction is statistically forced by an input fit, and no load-bearing premise collapses to an unverified self-citation chain. The central rate guarantees for the shape-to-solution maps are derived from the holomorphy assumptions without circular reduction to the paper's own inputs.

Axiom & Free-Parameter Ledger

0 free parameters · 2 axioms · 0 invented entities

The central claims rest on standard background results from functional analysis and complex analysis together with domain-specific well-posedness assumptions; no new free parameters or invented entities are introduced.

axioms (2)
  • domain assumption The pullback of the PDE to D_ref via affine-parametric shape encoding produces a collection of holomorphic parametric PDEs on D_ref.
    Invoked as the starting point that converts shape variation into parametric dependence on a fixed domain.
  • domain assumption Sufficient conditions exist for uniform (with respect to the parameter) well-posedness of the parametric PDEs on D_ref.
    Stated to guarantee existence, uniqueness and stability of the parametric solution families.

pith-pipeline@v0.9.0 · 5518 in / 1463 out tokens · 38929 ms · 2026-05-10T04:50:06.918589+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

73 extracted references · 10 canonical work pages · 1 internal anchor

  1. [1]

    Br¨ ugger and H

    R. Br¨ ugger and H. Harbrecht. On the reformulation of the classical Stefan problem as a shape optimization problem.SIAM Journal on Control Optimization, 60(1):310–329, 2022

  2. [2]

    Br¨ ugger, H

    R. Br¨ ugger, H. Harbrecht, and J. Tausch. On the numerical solution of a time-dependent shape optimization problem for the heat equation.SIAM Journal on Control Optimization, 59(2):931–953, 2021

  3. [3]

    Burton and G.F

    A.J. Burton and G.F. Miller. The application of integral equation methods to the numeri- cal solution of some exterior boundary-value problems.Proceedings of the Royal Society A, 323:201–210, 1971

  4. [4]

    Castrillon-Candas, F

    J.E. Castrillon-Candas, F. Nobile, and R.F. Tempone. Analytic regularity and collocation approximation for PDEs with random domain deformations.Computers & Mathematics with Applications, 71(6):1173–1197, 2016. SHAPE OPERATOR SURROGATES 25

  5. [5]

    Castrillon-Candas and J

    J.E. Castrillon-Candas and J. Xu. A stochastic collocation approach for parabolic PDEs with random domain deformations.Computers & Mathematics with Applications, 93:32–49, 2021

  6. [6]

    Cheng, Z

    Z. Cheng, Z. Hao, X. Wang, J. Huang, Y. Wu, X. Liu, Y. Zhao, S. Liu, and H. Su. Refer- ence neural operators: Learning the smooth dependence of solutions of PDEs on geometric deformations.Proceedings of Machine Learning Research, 235:8060–8076, 2024. Proceedings of ICML 2024

  7. [7]

    Chernov and T

    A. Chernov and T. Lˆ e. Analytic and Gevrey class regularity for parametric elliptic eigenvalue problems and applications.SIAM J. Numer. Anal., 62(4):1874–1900, 2024

  8. [8]

    Chernov and T

    A. Chernov and T. Lˆ e. Gevrey class regularity for steady-state incompressible Navier-Stokes equations in parametric domains and related models.arXiv-Preprint, arXiv:2504.13753, 2025

  9. [9]

    Christensen.An Introduction to Frames and Riesz Bases

    O. Christensen.An Introduction to Frames and Riesz Bases. Applied and Numerical Har- monic Analysis. Birkh¨ auser Boston, Inc., Boston, MA, 2003

  10. [10]

    Christensen.Frames and Bases

    O. Christensen.Frames and Bases. An Introductory Course. Applied and Numerical Har- monic Analysis. Birkh¨ auser Boston, Inc., Boston, MA, 2008

  11. [11]

    Cohen, C

    A. Cohen, C. Schwab, and J. Zech. Shape holomorphy of the stationary Navier-Stokes equa- tions.SIAM Journal on Mathematical Analaysis, 50:1720–1752, 2018

  12. [12]

    Colton and R

    D. Colton and R. Kress.Inverse Acoustic and Electromagnetic Scattering. Springer, Berlin- Heidelberg-New York, 2nd edition edition, 1997

  13. [13]

    Dahmen and R

    W. Dahmen and R. Stevenson. Element-by-element construction of wavelets satisfying sta- bility and moment conditions.SIAM J. Numer. Anal., 37(1):319–352, 1999

  14. [14]

    Dalla Riva, M

    M. Dalla Riva, M. Lanza de Cristoforis, and P. Musolino. Analytic dependence of volume potentials corresponding to parametric families of fundamental solutions.Integral Equations Operator Theory, 82(3):371–393, 2015

  15. [15]

    Dalla Riva, P

    M. Dalla Riva, P. Luzzini, and P. Musolino. Shape analyticity and singular perturbations for layer potential operators.ESAIM Math. Model. Numer. Anal., 56(6):1889–1910, 2022

  16. [16]

    Dautray and J.-L

    R. Dautray and J.-L. Lions.Mathematical analysis and numerical methods for science and technology. Vol. 5. Springer, Berlin, 1992. Evolution problems I

  17. [17]

    Davydov and R

    O. Davydov and R. Stevenson. Hierarchical Riesz bases forH spΩq,1ăsă 5 2 .Constr. Approx., 22(3):365–394, 2005

  18. [18]

    Delfour and J.-P

    M.C. Delfour and J.-P. Zol´ esio.Shapes and Geometries: Metric, Analysis, Differential Cal- culus, and Optimization. Society of Industrial and Applied Mathematics, Philadelphia, PA, second edition, 2011

  19. [19]

    Uncertainty quantification for stationary and time-dependent PDEs subject to Gevrey regular random domain deformations

    A. Djurdjevac, V. Kaarnioja, C. Schillings, and A.-A. Zepernick. Uncertainty quantification for stationary and time-dependent PDEs subject to Gevrey regular random domain deforma- tions.arXiv-Preprint, arXiv:2502.12345, 2025

  20. [20]

    D¨ olz and F

    J. D¨ olz and F. Henr´ ıquez. Parametric shape holomorphy of boundary integral operators with applications.SIAM J. Math. Anal., 56(5):6731–6767, 2024

  21. [21]

    D. Dung, V. Nguyen, D. Pham, and Ch. Schwab. Approximation of pde solution manifolds: Sparse-grid interpolation and quadrature. Technical Report 2026-07, Seminar for Applied Mathematics, ETH Z¨ urich, Switzerland, 2026

  22. [22]

    Duprez, V

    M. Duprez, V. Lleras, A. Lozinski, V. Vigon, and K. Vuillemot. Φ-FEM-FNO: A new approach to train a Neural Operator as a fast PDE solver for variable geometries.Communications in Nonlinear Science and Numerical Simulation, 152:109131, 2026

  23. [23]

    M.B. Giles. Multilevel Monte Carlo methods.Acta Numer., 24:259–328, 2015

  24. [24]

    X. Gong, D. Luo, T. O’Leary-Roseberry, R. Nicholson, and O. Ghattas. Shape Derivative- Informed Neural Operators with Application to Risk-Averse Shape Optimization.arXiv- Preprint, arXiv:2603.03211, 2026

  25. [25]

    Goswami, A

    S. Goswami, A. Bora, Y Yu, and G.E. Karniadakis. Physics-informed deep neural operator networks. InMachine Learning in Modeling and Simulation. Methods and Applications, pages 219–254. Springer Nature Switzerland AG, Cham, 2023

  26. [26]

    Gr¨ aßle, R

    B. Gr¨ aßle, R. Hiptmair, and S. Sauter. Stable skeleton integral equations for general- coefficient Helmholtz transmission problems.SAM Reports, ETHZ, 2025-20, 2025

  27. [27]

    Grisvard.Elliptic problems in nonsmooth domains, volume 69 ofClassics in Applied Math- ematics

    P. Grisvard.Elliptic problems in nonsmooth domains, volume 69 ofClassics in Applied Math- ematics. Society for Industrial and Applied Mathematics (SIAM), Philadelphia, PA, 2011. Reprint of the 1985 original. With a foreword by Susanne C. Brenner

  28. [28]

    Harbrecht, V

    H. Harbrecht, V. Karnaev, and M. Schmidlin. Quantifying domain uncertainty in linear elasticity.SIAM/ASA Journal on Uncertainty Quantification, 12(2):503–523, 2024. 26 HELMUT HARBRECHT AND CHRISTOPH SCHWAB

  29. [29]

    Harbrecht, M

    H. Harbrecht, M. Peters, and M. Siebenmorgen. Analysis of the domain mapping method for elliptic diffusion problems on random domains.Numerische Mathematik, 134(4):823–856, 2016

  30. [30]

    Harbrecht, M

    H. Harbrecht, M. Schmidlin, and Ch. Schwab. The Gevrey class implicit mapping theorem with application to UQ of semilinear elliptic PDEs.Mathematical Models and Methods in Applied Sciences, 34(5):881–917, 2024

  31. [31]

    Harbrecht, R

    H. Harbrecht, R. Schneider, and Ch. Schwab. Multilevel frames for sparse tensor product spaces.Numer. Math., 110(2):199–220, 2008

  32. [32]

    Heil.A Basis Theory Primer

    Ch. Heil.A Basis Theory Primer. Applied and Numerical Harmonic Analysis. Birkh¨ auser/Springer, New York, expanded edition, 2011

  33. [33]

    Henr´ ıquez and Ch

    F. Henr´ ıquez and Ch. Schwab. Shape holomorphy of the Calder´ on projector for the Laplacian inR 2.Integral Equations Operator Theory, 93(4):Paper No. 43, 40, 2021

  34. [34]

    Herrmann, Ch

    L. Herrmann, Ch. Schwab, and J. Zech. Neural and spectral operator surrogates: unified construction and expression rate bounds.Adv. Comput. Math., 50(4):Paper No. 72, 43, 2024

  35. [35]

    Hiptmair

    R. Hiptmair. Operator preconditioning.Comput. Math. Appl., 52(5):699–706, 2006

  36. [36]

    Hiptmair, L

    R. Hiptmair, L. Scarabosio, C. Schillings, and C. Schwab. Large deformation shape un- certainty quantification in acoustic scattering.Advances ind Computation Mathematics, 44:1475–1518, 2018

  37. [37]

    Hiptmair, Ch

    R. Hiptmair, Ch. Schwab, and E.A. Spence. Frequency-explicit shape holomorphy in uncer- tainty quantification for acoustic scattering.SIAM/ASA Journal on Uncertainty Quantifi- cation, 13(4):904–1949, 2025

  38. [38]

    Ilse, J.M

    M. Ilse, J.M. Tomczak, C. Louizos, and M. Welling. DIVA: Domain invariant variational autoencoders.Proceedings of Machine Learning Research, 21:322–348, 2020

  39. [39]

    Jerez-Hanckes, Ch

    C. Jerez-Hanckes, Ch. Schwab, and J. Zech. Electromagnetic wave scattering by random surfaces: Shape holomorphy.Math. Mod. Meth. Appl. Sci., 27(12):2229–2259, 2017

  40. [40]

    Kelley.General Topology

    J. Kelley.General Topology. University Series in Higher Mathematics, Van Nostrand, NY, 1955

  41. [41]

    Kovachki, S

    N.B. Kovachki, S. Lanthaler, and A.M. Stuart. Operator learning: Algorithms and analysis. InHandbook of Numerical Analysis, volume 25, pages 419–467. North-Holland, Amsterdam, 2024

  42. [42]

    Kussmaul

    R. Kussmaul. Ein numerisches Verfahren zur L¨ osung des Neumannschen Außenraumproblems f¨ ur die Helmholtzsche Schwingungsgleichung.Computing, pages 246–273, 1969

  43. [43]

    Geometric operator learning with optimal transport.Journal of Machine Learning Research, 26:1–38, 2025

    Xinyi Li, Zongyi Li, Nikola Kovachki, and Anima Anandkumar. Geometric operator learning with optimal transport.Journal of Machine Learning Research, 26:1–38, 2025

  44. [44]

    Z. Li, N.B. Kovachki, Ch. Choy, B. Li, J. Kossaifi, S.P. Otta, M.A. Nabian, M. Stadler, Ch. Hundt, K. Azizzadenesheli, and A. Anandkumar. Geometry-Informed Neural Operator for Large-Scale 3D PDEs. In A. Oh, T. Naumann, A. Globerson, K. Saenko, M. Hardt, and S. Levine, editors,Advances in Neural Information Processing Systems, volume 36, pages 35836–35854....

  45. [45]

    Fourier neural operator with learned deformations for PDEs on general geometries.J

    Zongyi Li, Daniel Zhengyu Huang, Burigede Liu, and Anima Anandkumar. Fourier neural operator with learned deformations for PDEs on general geometries.J. Mach. Learn. Res., 24:Paper No. [388], 26, 2023

  46. [46]

    N. Liu, X. Li, M.R. Rajanna, E.W. Reutzel, B. Sawyer, P. Rao, J. Lua, N. Phan, and Y. Yu. Deep neural operator enabled digital twin modeling for additive manufacturing.Advances in Computational Science and Engineering, 2(3):174–201, 2024

  47. [47]

    Loeffler, Z

    S.E. Loeffler, Z. Ahmad, S.Y. Ali, C. Yamamoto, D.M. Popescu, A. Yee, Y. Lal, N. Trayanova, and M. Maggioni. Graph Fourier Neural Kernels (G-FuNK): Learning solutions of nonlinear diffusive parametric PDEs on multiple domains.arXiv-Preprint, arXiv:2410.04655, 2024

  48. [48]

    H. Luo, H. Wu, H. Zhou, L. Xing, Y. Di, J. Wang, and M. Long. Transolver++: An accurate neural solver for PDEs on million-scale geometries.arXiv-Preprint, arXiv:2502.02414v2, 2025

  49. [49]

    Rigno: A graph-based framework for robust and accurate operator learning for pdes on arbitrary domains.arXiv preprint arXiv:2501.19205,

    S. Mousavi, S. Wen, L. Lingsch, M. Herde, B. Raoni´ c, and S. Mishra. RIGNO: A graph- based framework for robust and accurate operator learning for PDEs on arbitrary domains. arXiv-Preprint, arXiv:2501.19205, 2025

  50. [50]

    Mu˜ noz, Y

    G.A. Mu˜ noz, Y. Sarantopoulos, and A. Tonge. Complexifications of real Banach spaces, polynomials and multilinear maps.Studia Math., 134(1):1–33, 1999

  51. [51]

    P. Oswald. On a BPX-preconditioner for P1 elements.Computing, 51(2):125–133, 1993

  52. [52]

    Oswald.Multilevel Finite Element Approximation

    P. Oswald.Multilevel Finite Element Approximation. Teubner Skripten zur Numerik. B.G. Teubner, Stuttgart, 1994. Theory and applications. SHAPE OPERATOR SURROGATES 27

  53. [53]

    Padula, F

    G. Padula, F. Romor, G. Stabile, and G. Rozza. Generative models for the deformation of industrial shapes with linear geometric constraints: Model order and parameter space reductions.Computer Methods in Applied Mechanics and Engineering, 423:116823, 2024

  54. [54]

    Pazy.Semigroups of linear operators and applications to partial differential equations, volume 44 ofApplied Mathematical Sciences

    A. Pazy.Semigroups of linear operators and applications to partial differential equations, volume 44 ofApplied Mathematical Sciences. Springer, New York, 1983

  55. [55]

    Schwab and R

    Ch. Schwab and R. Stevenson. Space-time adaptive wavelet methods for parabolic evolution problems.Math. Comp., 78(267):1293–1318, 2009

  56. [56]

    Sokolowski and J.-P

    J. Sokolowski and J.-P. Zol´ esio.Introduction to Shape Optimization. Shape Sensitivity Anal- ysis, volume 16 ofSpringer Series in Computational Mathematics. Springer, Berlin, 1992

  57. [57]

    Steinwart and C

    I. Steinwart and C. Scovel. Mercer’s theorem on general domains: on the interaction between measures, kernels, and RKHSs.Constr. Approx., 35(3):363–417, 2012

  58. [58]

    Tartakovsky and D

    D.M. Tartakovsky and D. Xiu. Stochastic analysis of transport in tubes with rough walls. Journal of Computational Physics, 217(1):248–259, 2006

  59. [59]

    Taylor, A

    S. Taylor, A. Bihlo, and J.-Ch. Nave. Diffeomorphic neural operator learning.arXiv-Preprint, arxive:2508.06690, 2025

  60. [60]

    Temam.Navier-Stokes Equations, volume 2 ofTheory and Numerical Analysis, Stud

    R. Temam.Navier-Stokes Equations, volume 2 ofTheory and Numerical Analysis, Stud. Math. Appa.North-Holland, Amsterdam, 1977

  61. [61]

    Triebel.Function Spaces and Wavelets on Domains, volume 7 ofEMS Tracts in Mathe- matics

    H. Triebel.Function Spaces and Wavelets on Domains, volume 7 ofEMS Tracts in Mathe- matics. European Mathematical Society (EMS), Z¨ urich, 2008

  62. [62]

    Truesdell.A first course in rational continuum mechanics

    C.A. Truesdell.A first course in rational continuum mechanics. Vol. 1, volume 71 ofPure and Applied Mathematics. Academic Press, Inc., Boston, MA, second edition, 1991. General concepts

  63. [63]

    Valent.Boundary Value Problems of Finite Elasticity

    T. Valent.Boundary Value Problems of Finite Elasticity. Local Theorems on Existence, Uniqueness, and Analytic Dependence on Data, volume 31 ofSpringer Tracts in Natural Philosophy. Springer, New York, 1988

  64. [64]

    van Harten and L

    W.G. van Harten and L. Scarabosio. Exploiting locality in sparse polynomial approximation of parametric elliptic PDEs and application to parameterized domains.ESAIM Math. Model. Numer. Anal., 58(5):1581–1613, 2024

  65. [65]

    S. Wen, A. Kumbhat, L. Lingsch, S. Mousavi, Y. Zhao, P. Chandrashekar, and S. Mishra. Geometry aware operator transformer as an efficient and accurate neural surrogate for PDEs on arbitrary domains.arXiv-Preprint, arXiv:2505.18781, 2025

  66. [66]

    Westermann, B

    J. Westermann, B. Huber, T. O’Leary-Roseberry, and J. Zech. Performance of neural and polynomial operator surrogates, 2026

  67. [67]

    Wloka.Partielle Differentialgleichungen

    J. Wloka.Partielle Differentialgleichungen. B.G. Teubner, Stuttgart, 1982. Sobolevr¨ aume und Randwertaufgaben

  68. [68]

    S. Xiao, P. Jin, and Y. Tang. Learning solution operators of PDEs defined on varying domains via MIONet.arXiv-Preprint, arXiv:2402.15097v2, 2024

  69. [69]

    Xiu and D.M

    D. Xiu and D.M. Tartakovsky. Numerical methods for differential equations in random do- mains.SIAM Journal on Scientific Computing, 28(3):1167–1185, 2006

  70. [70]

    M. Yin, N. Charon, R. Brody, L. Lu, N. Trayanova, and M. Maggioni. Dimon: Learning solution operators of partial differential equations on a diffeomorphic family of domains. arXiv-Preprint, arXiv:2402.07250, 2024

  71. [71]

    C. Zeng, Y. Zhang, J. Zhou, Y. Wang, Z. Wang, Y. Liu, L. Wu, and D.Z. Huang. Point cloud neural operator for parametric pdes on complex and variable geometries.Computer Methods in Applied Mechanics and Engineering, 443:118022, 2025

  72. [72]

    Zhang, S

    L. Zhang, S. Pagani, J. Zhang, and F. Regazzoni. Shape-informed surrogate models based on signed distance function domain encoding.Journal of Computational Physics, 538:114178, 2025

  73. [73]

    Z. Zhao, C. Liu, Y. Li, Z. Chen, and X. Liu. Diffeomorphism neural operator for various domains and parameters of partial differential equations.Communications Physics, 8:15, 2025. 28 HELMUT HARBRECHT AND CHRISTOPH SCHWAB Helmut Harbrecht, Departement Mathematik und Informatik, Universit ¨at Basel, Spiegelgasse 1, 4051 Basel, Switzerland Email address:hel...