pith. machine review for the scientific record. sign in

arxiv: 2605.09782 · v1 · submitted 2026-05-10 · 💻 cs.DS · stat.ME

Recognition: no theorem link

Near-Linear Time Generalized Sinkhorn Algorithms for Bounded Genus Graphs

Authors on Pith no claims yet

Pith reviewed 2026-05-12 03:05 UTC · model grok-4.3

classification 💻 cs.DS stat.ME
keywords Sinkhorn algorithmoptimal transportbounded genus graphsplanar graphsgeodesic distancenear-linear timegraph decompositionmatrix multiplication
0
0 comments X

The pith

GenusSink delivers near-linear time approximate Sinkhorn algorithms for bounded-genus graphs with shortest-path costs.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper develops GenusSink to perform generalized Sinkhorn iterations for optimal transport problems where the cost is the shortest-path distance on a bounded-genus graph. It achieves this efficiency through graph separator decompositions that reduce the problem to fast operations on low-treewidth structures. A reader would care because this makes computing transport plans on large planar meshes or 3D object approximations practical at scales where the standard quadratic-time method fails. The work also establishes that the approximation becomes exact in a numerical sense for graphs with treewidth growing very slowly with n.

Core claim

GenusSink provides near-linear time preprocessing, iteration steps, and transport plan querying with near-linear memory for approximate generalized Sinkhorn algorithms using shortest-path-distance costs on bounded genus graphs. It relies on separator-based decompositions, computational geometry methods, and fast matrix-vector multiplications for generalized distance matrices via Fourier analysis and low displacement rank theory. The algorithm is numerically equivalent to the brute-force geodesic Sinkhorn on n-vertex graphs with treewidth O(log log n), such as trees.

What carries the argument

The separator-based decomposition of the graph paired with separation graph field integrators (S-GFIs) that support fast multiplications involving generalized distance matrices while keeping iteration errors controlled.

Load-bearing premise

Bounded-genus graphs admit separator decompositions that approximate their shortest-path metrics with small-treewidth metrics while preserving enough structure for accurate fast multiplications in the Sinkhorn process.

What would settle it

Measure the runtime on a family of planar graphs with increasing vertex count n and verify whether preprocessing, iteration, and querying times all scale as O(n polylog n); alternatively, compare the output matrix from GenusSink to the brute-force result on a tree graph with n vertices to confirm numerical equivalence within floating-point precision.

Figures

Figures reproduced from arXiv: 2605.09782 by Ananya Parashar, Derek Long, Dwaipayan Saha, Krzysztof Choromanski.

Figure 1
Figure 1. Figure 1: Left: Seven-vertex planar graph. Right: A pictorial description of its treewidth decomposition, with the treewidth parameter upper-bounding the sizes of the clusters of nodes above (that we will refer to as bags, see: Sec. 2.1). Two graph pieces, obtained with the separator {B, C, D}, are highlighted via red dashed boxes. The treewidth decomposition provides a gateway for designing efficient algorithms for… view at source ↗
Figure 2
Figure 2. Figure 2: Left to right: 2D-surfaces of 3D objects with genus values: 0, 1, 2 and 3. In this paper, we are especially interested in families of graphs with bounded genus that can be used in particular to provide discretized mesh representations of the above surfaces. distributions defined on the manifolds approximated by weighted graphs and with cost functions given by geodesic distances. We conduct rigorous theoret… view at source ↗
Figure 3
Figure 3. Figure 3: Left: The pictorial representation of the S-GFI data structure. In that example, the corresponding tree has a root with two children and four grandchildren. Right: The visualization of the cross_compute function with the medians depicted in red/green and their corresponding splits highlighted as dashed lines. ξi = X j∈RB f(ai [k] + bj [k])I[wk i ≺ z k j ]vj (11) We can then continue the calculations recurs… view at source ↗
Figure 4
Figure 4. Figure 4: ). In contrast, most approximate baselines exhibit a clear accuracy gap. 0 5000 10000 15000 20000 Number of nodes 10 13 10 9 10 5 10 1 10 3 |O Tf ull O Tm e t h o d| 0 5000 10000 15000 20000 Number of nodes 10 13 10 9 10 5 10 1 10 3 |O Tf ull O Tm e t h o d| 0 5000 10000 15000 20000 Number of nodes 10 13 10 9 10 5 10 1 10 3 |O Tf ull O Tm e t h o d| 0 5000 10000 15000 20000 Number of nodes 10 2 10 1 10 0 1… view at source ↗
Figure 5
Figure 5. Figure 5: Left: examples of Thingi10K meshes. Middle: absolute OT-cost error relative to dense geodesic Sinkhorn for several efficient Sinkhorn methods. As before, GenusSink stays numerically exact while providing significant computational gains. Right: runtime scaling for dense geodesic kernel construction versus two most accurate efficient methods: GenusSink and Greenkhorn (see: left), with error bars over repeate… view at source ↗
Figure 6
Figure 6. Figure 6: Bronx road graph (left), mean response times across different Sinkhorn methods (middle), and tail response time gaps relative to GenusSink at quantiles from 0.95 onwards (right). 5 Conclusion We proposed in this paper GenusSink, a new class of efficient Sinkhorn algorithms operating on graph metric spaces for bounded genus graphs and characterized by near-linear time (1) pre-processing, (2) iteration step,… view at source ↗
Figure 7
Figure 7. Figure 7: Planar dumbbell graphs after the root separator split. The two lobes form the left and right subgraphs, while the narrow bridge induces a small separator; the examples above correspond to two different handle widths. w ∈ {1, 2, 3}, and initial/target distributions given by two-component Gaussian mixtures localized on the left and right lobes, respectively, and normalized over the graph vertices. The entrop… view at source ↗
Figure 8
Figure 8. Figure 8: Planar dumbbell experiments for handle widths w = 1, 2, 3 (left to right). Top: absolute Sinkhorn￾cost error relative to the dense geodesic full-kernel baseline. Bottom: total runtime for the baseline and two most accurate efficient methods: Sparse Sinkhorn and GenusSink. Across all widths, GenusSink remains numerically exact while providing significant computational gains. In addition to the dumbbell grap… view at source ↗
Figure 9
Figure 9. Figure 9: Pseudo-genus surface family used in our synthetic experiments. We show planar sphere-topology surface meshes embedded in R 3 for pseudo-genus orders 1, 2, and 3, colored according to the root separator split used to initialize the separation tree. mixtures in the surface parameter coordinates and normalized over the graph vertices. We again compare GenusSink against the dense geodesic full-kernel Sinkhorn … view at source ↗
read the original abstract

We present GenusSink, a new class of approximate generalized Sinkhorn algorithms with shortest-path-distance costs for bounded genus (e.g. planar) graphs, providing near-linear time: (1) pre-processing, (2) iteration step, (3) final transport plan matrix querying and near-linear memory. Graphs handled by GenusSink include in particular planar graphs and bounded-genus meshes approximating 3D objects. GenusSink addresses total quadratic time complexity of its brute-force counterpart by leveraging separator-based decomposition of graphs, computational geometry techniques, and new results on fast matrix-vector multiplications with generalized distance matrices, using, in particular, Fourier analysis and low displacement rank theory. It is inspired by recent breakthroughs in graph theory on approximating bounded genus metrics with small treewidth metrics \citep{minor-free-paper}. The graph-centric approach enables us to target optimal transport problem with the corresponding distributions defined on the manifolds approximated by weighted graphs and with cost functions given by geodesic distances. We conduct rigorous theoretical analysis of GenusSink, provide practical implementations, leveraging newly introduced in this paper \textit{separation graph field integrators} (S-GFIs) data structures and present empirical verification. GenusSink provides orders of magnitude more accurate computations than other efficient Sinkhorn algorithms, while still guaranteeing significant computational improvements, as compared to the baseline. As a by-product of the developed methods, we show that GenusSink is \textbf{numerically equivalent} to the brute-force geodesic Sinkhorn algorithm on $n$-vertex graphs with treewidth $O(\log \log (n))$ (e.g. on trees).

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The manuscript introduces GenusSink, an approximate generalized Sinkhorn algorithm for optimal transport with shortest-path (geodesic) costs on bounded-genus graphs. It claims near-linear time for (1) preprocessing, (2) each iteration step, and (3) final transport-plan matrix querying, together with near-linear memory. The approach combines separator-based graph decompositions, metric approximations by small-treewidth graphs (citing minor-free results), fast generalized distance matrix-vector multiplications via Fourier analysis and low-displacement rank, and newly introduced separation graph field integrators (S-GFIs). The authors assert rigorous theoretical analysis, practical implementations, empirical verification showing orders-of-magnitude higher accuracy than other efficient Sinkhorn methods, and a by-product result that GenusSink is numerically equivalent to brute-force geodesic Sinkhorn on n-vertex graphs of treewidth O(log log n).

Significance. If the error bounds are shown to control accumulation across iterations while preserving the stated runtime and numerical equivalence, the result would meaningfully advance scalable OT on planar graphs and bounded-genus meshes that approximate 3D manifolds. The near-linear claims directly target the quadratic bottleneck of standard Sinkhorn with geodesic costs, and the low-treewidth equivalence is a clean theoretical corollary. The work also introduces new data structures (S-GFIs) that may have broader utility for fast matrix operations on graphs with good separators.

major comments (3)
  1. [Abstract] Abstract: the claim of 'rigorous theoretical analysis' is load-bearing for the central near-linear-time guarantee, yet the abstract supplies no explicit approximation error bounds for the S-GFI multiplications, no propagation analysis through the O(log(1/ε)) Sinkhorn iterations, and no statement of how separator-induced metric distortions affect the final transport plan. Without these, it is impossible to verify that per-step error δ remains controlled after iteration (e.g., via a contraction or Lipschitz argument on the Sinkhorn map).
  2. [Theoretical analysis] Theoretical analysis section (implied by the abstract's description of S-GFIs, Fourier analysis, and low-displacement rank): the manuscript must demonstrate that the compounded error after k = O(log(1/ε)) iterations stays o(ε) rather than growing as Ω(kδ). If only single-multiplication error is bounded, the final transport plan can deviate by Ω(log(1/ε)·δ), violating the assertion that GenusSink remains a faithful approximate geodesic Sinkhorn algorithm.
  3. [By-product result] By-product result on numerical equivalence: the claim that GenusSink is numerically equivalent to brute-force Sinkhorn on graphs of treewidth O(log log n) requires an explicit argument that all approximations become exact (zero error) in that regime. The abstract states the result but does not indicate where or how this exactness is proved.
minor comments (2)
  1. [Abstract] The abstract introduces the acronym 'S-GFIs' without an immediate parenthetical expansion or forward reference; this should be clarified on first use.
  2. [Abstract] The citation 'minor-free-paper' should be replaced by a full bibliographic entry so readers can locate the cited separator and metric-approximation results.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for the careful reading and constructive comments. We address each major comment point by point below, clarifying the existing analysis and indicating revisions to improve explicitness.

read point-by-point responses
  1. Referee: [Abstract] Abstract: the claim of 'rigorous theoretical analysis' is load-bearing for the central near-linear-time guarantee, yet the abstract supplies no explicit approximation error bounds for the S-GFI multiplications, no propagation analysis through the O(log(1/ε)) Sinkhorn iterations, and no statement of how separator-induced metric distortions affect the final transport plan. Without these, it is impossible to verify that per-step error δ remains controlled after iteration (e.g., via a contraction or Lipschitz argument on the Sinkhorn map).

    Authors: We agree the abstract is too concise on these points. The full manuscript (Sections 4.2 and 5) bounds the S-GFI multiplication error by δ = O(ε / log(1/ε)), shows the generalized Sinkhorn map is a contraction with Lipschitz constant ρ < 1 (via standard analysis on the scaling vectors), and absorbs separator distortions from the cited minor-free metric approximations into the overall (1+η) factor with η chosen small enough to fit within ε. The accumulated error is then bounded by δ/(1-ρ) rather than linear in the iteration count. We will revise the abstract to include a one-sentence summary of these controls. revision: yes

  2. Referee: [Theoretical analysis] Theoretical analysis section (implied by the abstract's description of S-GFIs, Fourier analysis, and low-displacement rank): the manuscript must demonstrate that the compounded error after k = O(log(1/ε)) iterations stays o(ε) rather than growing as Ω(kδ). If only single-multiplication error is bounded, the final transport plan can deviate by Ω(log(1/ε)·δ), violating the assertion that GenusSink remains a faithful approximate geodesic Sinkhorn algorithm.

    Authors: Section 5.3 already contains the required propagation argument. Because the Sinkhorn iteration map is contractive (Lipschitz constant ρ < 1), the total error after k steps satisfies a geometric-series bound ||e_k|| ≤ δ/(1-ρ) + ρ^k · initial, which remains o(ε) when δ = o(ε(1-ρ)). This is derived from the standard Lipschitz analysis of the generalized Sinkhorn operator on the space of positive scaling vectors. We will add an explicit corollary stating the accumulation bound and the choice of δ to make the dependence on k transparent. revision: yes

  3. Referee: [By-product result] By-product result on numerical equivalence: the claim that GenusSink is numerically equivalent to brute-force Sinkhorn on graphs of treewidth O(log log n) requires an explicit argument that all approximations become exact (zero error) in that regime. The abstract states the result but does not indicate where or how this exactness is proved.

    Authors: The equivalence is proved in Section 7: when treewidth is O(log log n), the separator decomposition produces subproblems whose distance matrices admit exact (zero-error) low-displacement-rank and Fourier representations, so the S-GFI integrators incur no truncation or approximation error and the metric approximation from the minor-free result is exact. Consequently every step of GenusSink coincides numerically with brute-force geodesic Sinkhorn. We will expand the abstract's by-product sentence to reference this section and note the exactness condition. revision: yes

Circularity Check

0 steps flagged

No significant circularity; derivation relies on external citations and new constructions

full rationale

The paper's central claims rest on separator decompositions and small-treewidth metric approximations drawn from externally cited prior work (minor-free results), combined with newly introduced S-GFIs, Fourier analysis, and low-displacement-rank matrix techniques. The stated numerical equivalence on O(log log n)-treewidth graphs follows directly from the approximations becoming exact in that regime rather than from any fitted parameter or self-referential definition. No load-bearing step reduces by construction to the paper's own inputs, self-citations, or ansatzes smuggled via prior author work. Error accumulation concerns raised in the skeptic note pertain to correctness analysis rather than circularity.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 1 invented entities

The central efficiency claims rest on the domain assumption that bounded-genus graphs permit separator decompositions preserving fast matrix structure, plus the newly introduced S-GFI data structures; no explicit free parameters are stated.

axioms (1)
  • domain assumption Bounded genus graphs admit separator-based decompositions that approximate their metrics with small-treewidth metrics while enabling fast generalized distance matrix operations
    Directly invoked via citation to the minor-free-paper and used to justify the near-linear time claims.
invented entities (1)
  • separation graph field integrators (S-GFIs) no independent evidence
    purpose: Data structures supporting the practical fast matrix-vector multiplications inside the GenusSink iterations
    Newly introduced in the paper for implementation; no independent evidence outside this work is provided.

pith-pipeline@v0.9.0 · 5598 in / 1509 out tokens · 81530 ms · 2026-05-12T03:05:08.325874+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

46 extracted references · 46 canonical work pages · 1 internal anchor

  1. [1]

    Alaya, Maxime Bérar, Gilles Gasso, and Alain Rakotomamonjy

    Mokhtar Z. Alaya, Maxime Bérar, Gilles Gasso, and Alain Rakotomamonjy. Screening sinkhorn algorithm for regularized optimal transport. InAdvances in Neural Information Processing Systems, volume 32, 2019

  2. [2]

    Altschuler, Jonathan Weed, and Philippe Rigollet

    Jason M. Altschuler, Jonathan Weed, and Philippe Rigollet. Near-linear time approxima- tion algorithms for optimal transport via sinkhorn iteration. In Isabelle Guyon, Ulrike von Luxburg, Samy Bengio, Hanna M. Wallach, Rob Fergus, S. V . N. Vishwanathan, and Roman Garnett, editors,Advances in Neural Information Processing Systems 30: Annual Conference on ...

  3. [3]

    Information geometry for regularized optimal transport and barycenters of patterns.Neural Comput., 31(5):827–848,

    Shun-ichi Amari, Ryo Karakida, Masafumi Oizumi, and Marco Cuturi. Information geometry for regularized optimal transport and barycenters of patterns.Neural Comput., 31(5):827–848,

  4. [4]

    Neural Computation , volume=

    doi: 10.1162/NECO\_A\_01178. URLhttps://doi.org/10.1162/neco_a_01178

  5. [5]

    Lectures on optimal transport.UNITEXT,

    Luigi Ambrosio, Elia Brué, and Daniele Semola. Lectures on optimal transport.UNITEXT,

  6. [6]

    URLhttps://api.semanticscholar.org/CorpusID:244769124

  7. [7]

    Wasserstein generative adversarial networks

    Martín Arjovsky, Soumith Chintala, and Léon Bottou. Wasserstein generative adversarial networks. In Doina Precup and Yee Whye Teh, editors,Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, NSW, Australia, 6-11 August 2017, Proceedings of Machine Learning Research, pages 214–223. PMLR, 2017. URL http:// proceedings.ml...

  8. [8]

    Bodlaender

    Hans L. Bodlaender. A linear time algorithm for finding tree-decompositions of small treewidth. Proceedings of the twenty-fifth annual ACM symposium on Theory of Computing, 1993. URL https://api.semanticscholar.org/CorpusID:2028181

  9. [9]

    Fast tree-field integrators: From low displacement rank to topological transformers

    Krzysztof Marcin Choromanski, Arijit Sehanobish, Somnath Basu Roy Chowdhury, Han Lin, Kumar Avinava Dubey, Tamás Sarlós, and Snigdha Chaturvedi. Fast tree-field integrators: From low displacement rank to topological transformers. In Amir Globersons, Lester Mackey, Danielle Belgrave, Angela Fan, Ulrich Paquet, Jakub M. Tomczak, and Cheng Zhang, edi- tors,A...

  10. [10]

    Cormen, Charles E

    Thomas H. Cormen, Charles E. Leiserson, Ronald L. Rivest, and Clifford Stein.Introduction to Algorithms, 3rd Edition. MIT Press, 2009. ISBN 978-0-262-03384-8. URL http://mitpress. mit.edu/books/introduction-algorithms

  11. [11]

    Sinkhorn distances: Lightspeed computation of optimal transport

    Marco Cuturi. Sinkhorn distances: Lightspeed computation of optimal transport. In Christopher J. C. Burges, Léon Bottou, Zoubin Ghahramani, and Kilian Q. Weinberger, editors,Advances in Neural Information Processing Systems 26: 27th Annual Conference on Neural Information Processing Systems 2013. Proceedings of a meeting held December 5-8, 2013, Lake Taho...

  12. [12]

    Differentiable ranking and sort- ing using optimal transport

    Marco Cuturi, Olivier Teboul, and Jean-Philippe Vert. Differentiable ranking and sort- ing using optimal transport. In Hanna M. Wallach, Hugo Larochelle, Alina Beygelz- imer, Florence d’Alché-Buc, Emily B. Fox, and Roman Garnett, editors,Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, ...

  13. [13]

    Fomin and Lukasz Kowalik and Daniel Lokshtanov and D

    Marek Cygan, Fedor V . Fomin, Lukasz Kowalik, Daniel Lokshtanov, Dániel Marx, Marcin Pilipczuk, Michal Pilipczuk, and Saket Saurabh.Parameterized Algorithms. Springer, 2015. ISBN 978-3-319-21274-6. doi: 10.1007/978-3-319-21275-3. URL https://doi.org/10. 1007/978-3-319-21275-3. 10

  14. [14]

    Springer, 2012

    Reinhard Diestel.Graph Theory, 4th Edition, volume 173 ofGraduate texts in mathematics. Springer, 2012. ISBN 978-3-642-14278-9

  15. [15]

    The optimal partial transport problem.Archive for Rational Mechanics and Analysis, 195:533–560, 2010

    Alessio Figalli. The optimal partial transport problem.Archive for Rational Mechanics and Analysis, 195:533–560, 2010. URL https://api.semanticscholar.org/CorpusID: 15899562

  16. [16]

    Kim, Vinod Vaikuntanathan, and Or Zamir

    Arnold Filtser and Hung Le. Low treewidth embeddings of planar and minor-free metrics. In63rd IEEE Annual Symposium on Foundations of Computer Science, FOCS 2022, Denver, CO, USA, October 31 - November 3, 2022, pages 1081–1092. IEEE, 2022. doi: 10.1109/ FOCS54457.2022.00105. URLhttps://doi.org/10.1109/FOCS54457.2022.00105

  17. [17]

    Charlie Frogner, Chiyuan Zhang, Hossein Mobahi, Mauricio Araya-Polo, and Tomaso A. Poggio. Learning with a wasserstein loss. In Corinna Cortes, Neil D. Lawrence, Daniel D. Lee, Masashi Sugiyama, and Roman Garnett, editors,Advances in Neural Information Processing Systems 28: Annual Conference on Neural Information Processing Systems 2015, December 7-12, 2...

  18. [18]

    Aude Genevay, Marco Cuturi, Gabriel Peyré, and Francis R. Bach. Stochastic optimization for large-scale optimal transport. In Daniel D. Lee, Masashi Sugiyama, Ulrike von Luxburg, Isabelle Guyon, and Roman Garnett, editors,Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, December 5- 10, ...

  19. [19]

    Courville

    Ishaan Gulrajani, Faruk Ahmed, Martín Arjovsky, Vincent Dumoulin, and Aaron C. Courville. Improved training of wasserstein gans. In Isabelle Guyon, Ulrike von Luxburg, Samy Bengio, Hanna M. Wallach, Rob Fergus, S. V . N. Vishwanathan, and Roman Garnett, edi- tors,Advances in Neural Information Processing Systems 30: Annual Conference on Neu- ral Informati...

  20. [20]

    Tape, Guy Wolf, and Smita Krishnaswamy

    Guillaume Huguet, Alexander Tong, María Ramos Zapatero, Christopher J. Tape, Guy Wolf, and Smita Krishnaswamy. Geodesic sinkhorn for fast and accurate optimal transport on manifolds. In2023 IEEE 33rd International Workshop on Machine Learning for Signal Processing (MLSP), pages 1–6. IEEE, 2023

  21. [21]

    Weiss, Gianluca Janka, Ann M

    Tomasz Kacprzak, Francois Kamper, Michael W. Weiss, Gianluca Janka, Ann M. Dillner, and Satoshi Takahama. Scalable approximate algorithms for optimal transport linear mod- els.ArXiv, abs/2504.04609, 2025. URL https://api.semanticscholar.org/CorpusID: 277621015

  22. [22]

    Kelvin Kan, Xingjian Li, and Stanley J. Osher. Ot-transformer: A continuous-time transformer architecture with optimal transport regularization.CoRR, abs/2501.18793, 2025. doi: 10.48550/ ARXIV .2501.18793. URLhttps://doi.org/10.48550/arXiv.2501.18793

  23. [23]

    Kantorovitch

    L. Kantorovitch. On the translocation of masses.Management Science, 5:1–4, 1958. URL https://api.semanticscholar.org/CorpusID:214798034

  24. [24]

    Metric learning in optimal transport for domain adaptation

    Tanguy Kerdoncuff, Rémi Emonet, and Marc Sebban. Metric learning in optimal transport for domain adaptation. In Christian Bessiere, editor,Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI 2020, pages 2162–2168. ijcai.org, 2020. doi: 10.24963/IJCAI.2020/299. URLhttps://doi.org/10.24963/ijcai.2020/299

  25. [25]

    Kolkin, Jason Salavon, and Gregory Shakhnarovich

    Nicholas I. Kolkin, Jason Salavon, and Gregory Shakhnarovich. Style transfer by relaxed optimal transport and self-similarity. InIEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, June 16-20, 2019, pages 10051–10060. Computer Vision Foundation / IEEE, 2019. doi: 10.1109/CVPR.2019.01029

  26. [26]

    Importance sparsification for sinkhorn algo- rithm.J

    Mengyu Li, Jun Yu, Tao Li, and Cheng Meng. Importance sparsification for sinkhorn algo- rithm.J. Mach. Learn. Res., 24:247:1–247:44, 2023. URL https://jmlr.org/papers/ v24/22-1311.html. 11

  27. [27]

    Recent advances in optimal transport for machine learning, 2024

    Eduardo Fernandes Montesuma, Fred Ngolè Mboula, and Antoine Souloumiac. Recent advances in optimal transport for machine learning, 2024. URL https://arxiv.org/abs/2306. 16156

  28. [28]

    Cambridge University Press, 2014

    Yann Ollivier, Herve Pajot, and Cédric Villani, editors.Optimal Transport - Theory and Applications, volume 413 ofLondon Mathematical Society lec- ture note series. Cambridge University Press, 2014. ISBN 978-1-10-768949-

  29. [29]

    URL http://www.cambridge.org/de/academic/subjects/mathematics/ geometry-and-topology/optimal-transport-theory-and-applications

  30. [30]

    Fatima Minhas, R

    Luiz Manella Pereira and M. Hadi Amini. A survey on optimal transport for machine learning: Theory and applications.IEEE Access, 13:26506–26526, 2025. doi: 10.1109/ACCESS.2025. 3539926. URLhttps://doi.org/10.1109/ACCESS.2025.3539926

  31. [31]

    Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks

    Gabriel Peyré. Optimal transport for machine learners.CoRR, abs/2505.06589, 2025. doi: 10.48550/ARXIV .2505.06589. URLhttps://doi.org/10.48550/arXiv.2505.06589

  32. [32]

    Computational optimal transport.Found

    Gabriel Peyré and Marco Cuturi. Computational optimal transport.Found. Trends Mach. Learn., 11(5-6):355–607, 2019. doi: 10.1561/2200000073. URL https://doi.org/10. 1561/2200000073

  33. [33]

    Gromov-wasserstein averaging of kernel and distance matrices

    Gabriel Peyré, Marco Cuturi, and Justin Solomon. Gromov-wasserstein averaging of kernel and distance matrices. In Maria-Florina Balcan and Kilian Q. Weinberger, editors,Proceedings of the 33nd International Conference on Machine Learning, ICML 2016, New York City, NY, USA, June 19-24, 2016, JMLR Workshop and Conference Proceedings, pages 2664–2672. JMLR.org,

  34. [34]

    URLhttp://proceedings.mlr.press/v48/peyre16.html

  35. [35]

    Wasserstein barycenter and its application to texture mixing

    Julien Rabin, Gabriel Peyré, Julie Delon, and Marc Bernot. Wasserstein barycenter and its application to texture mixing. In Alfred M. Bruckstein, Bart M. ter Haar Romeny, Alexander M. Bronstein, and Michael M. Bronstein, editors,Scale Space and Variational Methods in Computer Vision - Third International Conference, SSVM 2011, Ein-Gedi, Is- rael, May 29 -...

  36. [36]

    Mason, and Andrea Raith

    Samuel Ridler, Andrew J. Mason, and Andrea Raith. A simulation and optimisation package for emergency medical services.European Journal of Operational Research, 298(3):1101– 1113, 2022. ISSN 0377-2217. doi: https://doi.org/10.1016/j.ejor.2021.07.038. URL https: //www.sciencedirect.com/science/article/pii/S0377221721006330

  37. [37]

    Tim Salimans, Han Zhang, Alec Radford, and Dimitris N. Metaxas. Improving gans using optimal transport. In6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings. OpenRe- view.net, 2018. URLhttps://openreview.net/forum?id=rkQkBnJAb

  38. [38]

    Linear time sinkhorn divergences using positive features

    Meyer Scetbon and Marco Cuturi. Linear time sinkhorn divergences using positive features. In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin, editors,Advances in Neural Information Processing Systems, volume 33, pages 13468–13480. Curran Associates, Inc., 2020. URL https://proceedings.neurips.cc/paper_files/paper/2020/file/ 9bde76f262285bb1...

  39. [39]

    Stabilized sparse scaling algorithms for entropy regularized transport problems.SIAM Journal on Scientific Computing, 41(3):A1443–A1481, 2019

    Bernhard Schmitzer. Stabilized sparse scaling algorithms for entropy regularized transport problems.SIAM Journal on Scientific Computing, 41(3):A1443–A1481, 2019. doi: 10.1137/ 16M1106018

  40. [40]

    Espformer: Doubly-stochastic attention with expected sliced transport plans

    Ashkan Shahbazi, Elaheh Akbari, Darian Salehi, Xinran Liu, Navid Naderializadeh, and Soheil Kolouri. Espformer: Doubly-stochastic attention with expected sliced transport plans. In Aarti Singh, Maryam Fazel, Daniel Hsu, Simon Lacoste-Julien, Felix Berkenkamp, Tegan Maharaj, Kiri Wagstaff, and Jerry Zhu, editors,Forty-second International Conference on Mac...

  41. [41]

    Sainath, and Sanjiv Kumar

    Vikas Sindhwani, Tara N. Sainath, and Sanjiv Kumar. Structured transforms for small-footprint deep learning. In Corinna Cortes, Neil D. Lawrence, Daniel D. Lee, Masashi Sugiyama, and Roman Garnett, editors,Advances in Neural Information Processing Systems 28: Annual Confer- ence on Neural Information Processing Systems 2015, December 7-12, 2015, Montreal,...

  42. [42]

    Accelerating sinkhorn algorithm with sparse newton iterations

    Xun Tang, Michael Shavlovsky, Holakou Rahmanian, Elisa Tardini, Kiran Koshy Thekumpara- mpil, Tesi Xiao, and Lexing Ying. Accelerating sinkhorn algorithm with sparse newton iterations. InThe Twelfth International Conference on Learning Representations, ICLR 2024, Vienna, Austria, May 7-11, 2024. OpenReview.net, 2024. URL https://openreview.net/forum? id=K...

  43. [43]

    Optimal transport for structured data with application on graphs

    Titouan Vayer, Nicolas Courty, Romain Tavenard, Laetitia Chapel, and Rémi Flamary. Optimal transport for structured data with application on graphs. In Kamalika Chaudhuri and Ruslan Salakhutdinov, editors,Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA, Proceedings of Machine Lea...

  44. [44]

    Gromov-wasserstein learning for graph matching and node embedding

    Hongteng Xu, Dixin Luo, Hongyuan Zha, and Lawrence Carin. Gromov-wasserstein learning for graph matching and node embedding. In Kamalika Chaudhuri and Ruslan Salakhutdinov, editors,Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9- 15 June 2019, Long Beach, California, USA, Proceedings of Machine Learning Research, pages 6...

  45. [45]

    pseudo-genus

    Qingnan Zhou and Alec Jacobson. Thingi10k: A dataset of 10,000 3d-printing models.arXiv preprint arXiv:1605.04797, 2016. A Efficient computation of cross_compute for|S|= 1and generalf The problem of computing efficientlycross_computeis equivalent to the problem of the efficient (near-linear) multiplication with matrices M= [f(x i +y j)]j=1,...,|B| i=1,......

  46. [46]

    Guidelines: • The answer [N/A] means that the paper does not involve crowdsourcing nor research with human subjects

    Institutional review board (IRB) approvals or equivalent for research with human subjects Question: Does the paper describe potential risks incurred by study participants, whether such risks were disclosed to the subjects, and whether Institutional Review Board (IRB) approvals (or an equivalent approval/review based on the requirements of your country or ...