Recognition: unknown
On Model-Based Clustering With Entropic Optimal Transport
Pith reviewed 2026-05-07 15:03 UTC · model grok-4.3
The pith
Entropic optimal transport loss shares the log-likelihood global optimum but avoids its spurious local optima in model-based clustering.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
The entropic optimal transport loss shares the same global optimum as the log-likelihood but has a much better-behaved landscape, thereby avoiding spurious local-optima configurations that are pervasive with the log-likelihood. Similar to the EM algorithm for the log-likelihood, this new loss can be optimized by the Sinkhorn-EM algorithm, which converges at a rate comparable to that of EM. Extensive numerical experiments and real-world applications show that this new loss outperforms log-likelihood optimization.
What carries the argument
The entropic optimal transport loss between the empirical data distribution and the parametric mixture model, minimized via the Sinkhorn-EM algorithm.
If this is right
- The Sinkhorn-EM algorithm reaches solutions of comparable quality to multiple-restart EM but with far fewer initializations.
- Clustering performance improves on image segmentation tasks such as C. elegans microscopy.
- Spatial transcriptomics datasets yield higher-quality cell-type groupings than likelihood-based methods.
- Convergence occurs at rates comparable to the classical EM algorithm for the same models.
Where Pith is reading between the lines
- The same loss construction could be applied to other mixture families or latent-variable models whose likelihood surfaces contain many local optima.
- In high-dimensional or noisy data regimes the improved landscape may reduce sensitivity to initialization more than is shown here.
- The method may combine naturally with regularization or constraints that further stabilize the optimization path.
Load-bearing premise
The entropic optimal transport loss must possess exactly the same global optimum as the log-likelihood while possessing a meaningfully smoother landscape that the Sinkhorn-EM algorithm can reliably reach.
What would settle it
On a mixture model instance whose log-likelihood surface is known to contain spurious local optima, repeated independent runs of Sinkhorn-EM should reach the global solution in nearly every trial while standard EM reaches it only with multiple random starts.
Figures
read the original abstract
We develop a new methodology for model-based clustering. Optimizing the log-likelihood provides a principled statistical framework for clustering, with solutions found via the EM algorithm. However, because the log-likelihood is nonconvex, only convergence to stationary points can be guaranteed, and practitioners often use multiple starting points in the hope that one will converge to the global solution. We consider a new loss function based on entropic optimal transport that shares the same global optimum as the log-likelihood but has a much better-behaved landscape, thereby avoiding spurious local-optima configurations that are pervasive with the log-likelihood. Similar to the EM algorithm for the log-likelihood, this new loss can be optimized by the Sinkhorn-EM algorithm, which we show converges at a rate comparable to that of EM. By analyzing extensive numerical experiments and two real-world applications in image segmentation in C. elegans microscopy and clustering in spatial transcriptomics, we show that this new loss outperforms log-likelihood optimization, indicating that it represents a valuable clustering methodology for practitioners.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The paper proposes a new loss function for model-based clustering derived from entropic optimal transport. It claims this loss shares the exact global optimum with the standard Gaussian mixture log-likelihood (including label-switching invariance) while having a smoother landscape that avoids the spurious local optima common to EM. The authors introduce the Sinkhorn-EM algorithm for optimization, establish comparable convergence rates, and report superior empirical performance over log-likelihood optimization in extensive simulations plus two real applications (C. elegans image segmentation and spatial transcriptomics clustering).
Significance. If the claimed equivalence of global optima and the landscape improvement hold rigorously, the method would address a long-standing practical limitation of EM for mixture models, reducing reliance on multiple random initializations. The empirical results on real biological data suggest potential value for practitioners, and the provision of reproducible code and parameter-free derivations (if present) would strengthen the contribution.
major comments (3)
- [§3.2, Theorem 1] §3.2 and Theorem 1: the asserted exact equivalence of global optima between the entropic OT loss and the log-likelihood is load-bearing for the central claim, yet the provided argument matches the objectives only at the true parameters without a full characterization of the critical-point sets or an explicit proof that the entropy term and Sinkhorn approximation introduce no additional stationary points.
- [§4] §4: the convergence-rate analysis for Sinkhorn-EM assumes a fixed entropy regularization strength that yields rates comparable to EM, but the paper does not quantify how sensitive the rate is to the choice of this hyperparameter or demonstrate that the chosen value preserves the claimed landscape advantage across the simulation settings in §5.
- [§5.3–5.4] §5.3–5.4 (real-data applications): the reported outperformance in C. elegans segmentation and spatial transcriptomics relies on post-hoc selection of the regularization parameter and number of clusters; without a pre-specified protocol or cross-validation details, it is unclear whether the gains are robust or could be matched by careful tuning of standard EM with multiple starts.
minor comments (3)
- [§2–3] Notation for the transport plan and marginal constraints is introduced inconsistently between §2 and §3; a single unified definition would improve readability.
- [Figure 2] Figure 2 (landscape visualization) lacks axis labels on the contour plots and does not indicate the location of the true parameters, making it difficult to verify the claimed absence of spurious minima.
- [Abstract and §4] The abstract states that Sinkhorn-EM 'converges at a rate comparable to that of EM,' but the main text provides only asymptotic statements; a finite-sample bound or numerical timing table would be helpful.
Simulated Author's Rebuttal
We thank the referee for their constructive and detailed comments, which have helped clarify several aspects of our work. We provide point-by-point responses below and describe the revisions we will implement.
read point-by-point responses
-
Referee: [§3.2, Theorem 1] §3.2 and Theorem 1: the asserted exact equivalence of global optima between the entropic OT loss and the log-likelihood is load-bearing for the central claim, yet the provided argument matches the objectives only at the true parameters without a full characterization of the critical-point sets or an explicit proof that the entropy term and Sinkhorn approximation introduce no additional stationary points.
Authors: We appreciate the referee highlighting this important point. Theorem 1 establishes that the global minimum of the entropic OT loss coincides with that of the log-likelihood by showing that at the true parameters, the two objectives differ by a constant independent of the parameters. However, we acknowledge that a complete characterization of all critical points is not provided. In the revision, we will add a lemma showing that any stationary point of the OT loss corresponds to a stationary point of the log-likelihood under the Sinkhorn approximation, and discuss the conditions under which no spurious stationary points are introduced. This will strengthen the theoretical foundation. revision: yes
-
Referee: [§4] §4: the convergence-rate analysis for Sinkhorn-EM assumes a fixed entropy regularization strength that yields rates comparable to EM, but the paper does not quantify how sensitive the rate is to the choice of this hyperparameter or demonstrate that the chosen value preserves the claimed landscape advantage across the simulation settings in §5.
Authors: We agree that sensitivity analysis would be valuable. In the revised manuscript, we will include additional experiments varying the regularization parameter ε over a range and report the convergence rates and clustering performance. We will also provide theoretical bounds showing how the rate depends on ε, under the assumption that ε is chosen sufficiently small to maintain the landscape advantage. revision: yes
-
Referee: [§5.3–5.4] §5.3–5.4 (real-data applications): the reported outperformance in C. elegans segmentation and spatial transcriptomics relies on post-hoc selection of the regularization parameter and number of clusters; without a pre-specified protocol or cross-validation details, it is unclear whether the gains are robust or could be matched by careful tuning of standard EM with multiple starts.
Authors: The referee correctly identifies a limitation in the presentation of the real-data results. While we selected parameters based on domain knowledge and visual inspection in the original submission, we will revise §5.3 and §5.4 to include a cross-validation procedure for selecting the number of clusters and regularization strength. We will also compare against EM with multiple random initializations using the same selection protocol to demonstrate robustness. This will clarify that the gains are not due to post-hoc tuning. revision: yes
Circularity Check
Entropic OT loss introduced as independent construction; equivalence to log-likelihood optimum derived, not definitional.
full rationale
The paper defines the new loss directly from entropic optimal transport (independent of the mixture log-likelihood) and states that it shares the same global optimum as a derived property rather than by re-expressing the loss in terms of the log-likelihood or fitting parameters to force the match. The Sinkhorn-EM algorithm is presented as an analogous but distinct optimizer. No equations reduce the claimed optimum or landscape improvement to a tautology or to a self-citation chain; the central claims rest on the OT formulation plus empirical validation in experiments and applications. This meets the criteria for a self-contained derivation with no load-bearing circular steps.
Axiom & Free-Parameter Ledger
free parameters (1)
- entropy regularization strength
axioms (1)
- domain assumption The entropic OT loss shares exactly the same global optimum as the log-likelihood
Reference graph
Works this paper leans on
-
[1]
Altschuler, J., Weed, J., and Rigollet, P. (2017). Near-linear time approximation algorithms for optimal transport via S inkhorn iteration. In Advances in Neural Information Processing Systems , pages 1961--1971
2017
-
[2]
and Vassilvitskii, S
Arthur, D. and Vassilvitskii, S. (2007). K-means++ the advantages of careful seeding. In Proceedings of the eighteenth annual ACM-SIAM symposium on Discrete algorithms , pages 1027--1035
2007
-
[3]
J., and Yu, B
Balakrishnan, S., Wainwright, M. J., and Yu, B. (2017). Statistical guarantees for the EM algorithm: from population to sample-based analysis. Ann. Statist. , 45(1):77--120
2017
-
[4]
E., Gerber, M., and Robert, C
Bernton, E., Jacob, P. E., Gerber, M., and Robert, C. P. (2019). On parameter estimation with the wasserstein distance. Information and Inference: A Journal of the IMA , 8(4):657--676
2019
-
[5]
Biernacki, C., Celeux, G., and Govaert, G. (2003). Choosing starting values for the em algorithm for getting the highest likelihood in multivariate gaussian mixture models. Computational Statistics & Data Analysis , 41(3-4):561--575
2003
-
[6]
B., and Raftery, A
Bouveyron, C., Celeux, G., Murphy, T. B., and Raftery, A. E. (2019). Model-based clustering and classification for data science: with applications in R , volume 50. Cambridge University Press
2019
-
[7]
Chen, Y., Song, D., Xi, X., and Zhang, Y. (2024). Local minima structures in gaussian mixture models. IEEE Transactions on Information Theory , 70(6):4218--4257
2024
-
[8]
and Church, G
Cheng, Y. and Church, G. M. (2000). Biclustering of expression data. In Ismb , volume 8, pages 93--103
2000
-
[9]
Chewi, S., Niles-Weed, J., and Rigollet, P. (2024). Statistical optimal transport
2024
-
[10]
and Tusn \'a dy, G
Csisz \'a r, I. and Tusn \'a dy, G. (1984). Information geonetry and alternating minimization procedures. Statistics and decisions , 1:205--237
1984
-
[11]
Cuturi, M. (2013). Sinkhorn distances: Lightspeed computation of optimal transport. In Advances in neural information processing systems , pages 2292--2300
2013
-
[12]
and Doucet, A
Cuturi, M. and Doucet, A. (2014). Fast computation of wasserstein barycenters. In International conference on machine learning , pages 685--693. PMLR
2014
-
[13]
P., Laird, N
Dempster, A. P., Laird, N. M., and Rubin, D. B. (1977). Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society: Series B (Methodological) , 39(1):1--22
1977
-
[14]
Dhillon, I. S. (2001). Co-clustering documents and words using bipartite spectral graph partitioning. In Proceedings of the seventh ACM SIGKDD international conference on Knowledge discovery and data mining , pages 269--274
2001
-
[15]
J., Jordan, M
Dwivedi, R., Ho, N., Khamaru, K., Wainwright, M. J., Jordan, M. I., and Yu, B. (2020). Singularity, misspecification and the convergence rate of em. The Annals of Statistics , 48(6):3161--3182
2020
-
[16]
and Sieranoja, S
Fr \"a nti, P. and Sieranoja, S. (2019). How much can k-means be improved by using better initialization and repeats? Pattern Recognition , 93:95--112
2019
- [17]
-
[18]
and Nadif, M
Govaert, G. and Nadif, M. (2013). Co-clustering: models, algorithms and applications . John Wiley & Sons
2013
-
[19]
and Hundrieser, S
Groppe, M. and Hundrieser, S. (2024). Lower complexity adaptation for empirical entropic optimal transport. Journal of Machine Learning Research , 25(344):1--55
2024
-
[20]
S., Raftery, A
Handcock, M. S., Raftery, A. E., and Tantrum, J. M. (2007). Model-based clustering for social networks. Journal of the Royal Statistical Society: Series A (Statistics in Society) , 170(2):301--354
2007
-
[21]
Hartigan, J. A. (1972). Direct clustering of a data matrix. Journal of the american statistical association , 67(337):123--129
1972
-
[22]
and Arabie, P
Hubert, L. and Arabie, P. (1985). Comparing partitions. Journal of classification , 2:193--218
1985
-
[23]
Hundrieser, S., Klatt, M., and Munk, A. (2024). Limit distributions and sensitivity analysis for empirical entropic optimal transport on countable spaces. The Annals of Applied Probability , 34(1B):1403--1468
2024
-
[24]
J., and Jordan, M
Jin, C., Zhang, Y., Balakrishnan, S., Wainwright, M. J., and Jordan, M. I. (2016). Local maxima in the likelihood of gaussian mixture models: Structural results and algorithmic consequences. In NeuRIPS , pages 4116--4124
2016
-
[25]
and Warmuth, M
Kivinen, J. and Warmuth, M. K. (1997). Exponentiated gradient versus gradient descent for linear predictors. information and computation , 132(1):1--63
1997
-
[26]
Klatt, M., Tameling, C., and Munk, A. (2020). Empirical regularized optimal transport: Statistical theory and applications. SIAM Journal on Mathematics of Data Science , 2(2):419--443
2020
-
[27]
K., and Hoffmann, H
Kolouri, S., Rohde, G. K., and Hoffmann, H. (2018). Sliced wasserstein distance for learning gaussian mixture models. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition , pages 3427--3436
2018
-
[28]
B., Werenski, M., Murphy, J
Masud, S. B., Werenski, M., Murphy, J. M., and Aeron, S. (2023). Multivariate soft rank via entropy-regularized optimal transport: Sample efficiency and generative modeling. Journal of Machine Learning Research , 24(160):1--65
2023
-
[29]
R., Collado-Torres, L., Weber, L
Maynard, K. R., Collado-Torres, L., Weber, L. M., Uytingco, C., Barry, B. K., Williams, S. R., Catallini, J. L., Tran, M. N., Besich, Z., Tippani, M., et al. (2021). Transcriptome-scale spatial gene expression in the human dorsolateral prefrontal cortex. Nature neuroscience , 24(3):425--436
2021
-
[30]
McLachlan, G. J. (1982). 9 the classification and mixture maximum likelihood approaches to cluster analysis. Handbook of statistics , 2:199--208
1982
-
[32]
and Niles-Weed, J
Mena, G. and Niles-Weed, J. (2019). Statistical bounds for entropic optimal transport: sample complexity and the central limit theorem. In Advances in Neural Information Processing Systems , pages 4543--4553
2019
-
[33]
and Pachter, L
Moses, L. and Pachter, L. (2022). Museum of spatial transcriptomics. Nature methods , 19(5):534--546
2022
-
[34]
and Govaert, G
Nadif, M. and Govaert, G. (2008). Algorithms for model-based block gaussian clustering. DMIN , 8:14--17
2008
-
[35]
Neal, R. M. and Hinton, G. E. (1998). A view of the EM algorithm that justifies incremental, sparse, and other variants. In Learning in graphical models , pages 355--368. Springer
1998
-
[36]
Nejatbakhsh, A., Varol, E., Yemini, E., Hobert, O., and Paninski, L. (2020). Probabilistic joint segmentation and labeling of c. elegans neurons. In Medical Image Computing and Computer Assisted Intervention--MICCAI 2020: 23rd International Conference, Lima, Peru, October 4--8, 2020, Proceedings, Part V 23 , pages 130--140. Springer
2020
-
[37]
Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., Blondel, M., Prettenhofer, P., Weiss, R., Dubourg, V., et al. (2011). Scikit-learn: Machine learning in python. the Journal of machine Learning research , 12:2825--2830
2011
-
[38]
Peyr \'e , G., Cuturi, M., et al. (2019). Computational optimal transport. Foundations and Trends in Machine Learning , 11(5-6):355--607
2019
-
[39]
Pooladian, A.-A., Divol, V., and Niles-Weed, J. (2023). Minimax estimation of discontinuous optimal transport maps: The semi-discrete case. In International Conference on Machine Learning , pages 28128--28150. PMLR
2023
-
[40]
Redner, R. A. and Walker, H. F. (1984). Mixture densities, maximum likelihood and the EM algorithm. SIAM review , 26(2):195--239
1984
-
[41]
and Stromme, A
Rigollet, P. and Stromme, A. J. (2025). On the sample complexity of entropic optimal transport. The Annals of Statistics , 53(1):61--90
2025
-
[42]
and Weed, J
Rigollet, P. and Weed, J. (2018). Entropic optimal transport is maximum-likelihood deconvolution. Comptes rendus Math\'ematique , 356(11--12)
2018
-
[43]
Sadhu, R., Goldfeld, Z., and Kato, K. (2024). Stability and statistical inference for semidiscrete optimal transport maps. The Annals of Applied Probability , 34(6):5694--5736
2024
-
[44]
and Knopp, P
Sinkhorn, R. and Knopp, P. (1967). Concerning nonnegative matrices and doubly stochastic matrices. Pacific Journal of Mathematics , 21(2):343--348
1967
-
[45]
and Risso, D
Sottosanti, A. and Risso, D. (2023). Co-clustering of spatially resolved transcriptomic data. The Annals of Applied Statistics , 17(2):1444--1468
2023
-
[46]
Steinley, D. (2003). Local optima in k-means clustering: what you don't know may hurt you. Psychological methods , 8(3):294
2003
-
[47]
Tan, K. M. and Witten, D. M. (2014). Sparse biclustering of transposable data. Journal of Computational and Graphical Statistics , 23(4):985--1008
2014
-
[48]
Varol, E., Nejatbakhsh, A., Sun, R., Mena, G., Yemini, E., Hobert, O., and Paninski, L. (2020). Statistical atlas of c. elegans neurons. In International Conference on Medical Image Computing and Computer-Assisted Intervention , pages 119--129. Springer
2020
-
[49]
Villani, C. (2008). Optimal transport: old and new , volume 338. Springer Science & Business Media
2008
-
[50]
Wu, C. J. (1983). On the convergence properties of the em algorithm. The Annals of statistics , pages 95--103
1983
-
[51]
and Zhou, H
Wu, Y. and Zhou, H. H. (2021). Randomly initialized em algorithm for two-component gaussian mixture achieves near optimality in O( n ) iterations. Mathematical Statistics & Learning , 4
2021
-
[52]
J., and Maleki, A
Xu, J., Hsu, D. J., and Maleki, A. (2018). Benefits of over-parameterization with EM . In Advances in Neural Information Processing Systems , pages 10662--10672
2018
-
[53]
and Jordan, M
Xu, L. and Jordan, M. I. (1996). On convergence properties of the EM algorithm for gaussian mixtures. Neural computation , 8(1):129--151
1996
- [54]
-
[55]
E., Samuel, A
Yemini, E., Lin, A., Nejatbakhsh, A., Varol, E., Sun, R., Mena, G. E., Samuel, A. D., Paninski, L., Venkatachalam, V., and Hobert, O. (2021). Neuropal: a multicolor atlas for whole-brain neuronal identification in c. elegans. Cell , 184(1):272--288
2021
-
[56]
Lincoln laboratory journal , volume=
A survey of spectral unmixing algorithms , author=. Lincoln laboratory journal , volume=. 2003 , publisher=
2003
-
[57]
The Annals of Statistics , volume=
CHIME: Clustering of high-dimensional Gaussian mixtures with EM algorithm and its optimality , author=. The Annals of Statistics , volume=. 2019 , publisher=
2019
-
[58]
Journal of Multivariate Analysis , volume=
Law of log determinant of sample covariance matrix and optimal estimation of differential entropy for high-dimensional Gaussian distributions , author=. Journal of Multivariate Analysis , volume=. 2015 , publisher=
2015
-
[59]
IEEE Transactions on Image Processing , volume=
A Gaussian mixture model representation of endmember variability in hyperspectral unmixing , author=. IEEE Transactions on Image Processing , volume=. 2018 , publisher=
2018
-
[60]
ISPRS Journal of Photogrammetry and Remote Sensing , volume=
Blind spectral unmixing based on sparse component analysis for hyperspectral remote sensing imagery , author=. ISPRS Journal of Photogrammetry and Remote Sensing , volume=. 2016 , publisher=
2016
-
[61]
Plos one , volume=
Robust blind spectral unmixing for fluorescence microscopy using unsupervised learning , author=. Plos one , volume=. 2019 , publisher=
2019
-
[62]
Computer methods and programs in biomedicine , volume=
A review of atlas-based segmentation for magnetic resonance brain images , author=. Computer methods and programs in biomedicine , volume=. 2011 , publisher=
2011
-
[63]
NeuroImage , volume=
Automatic anatomical brain MRI segmentation combining label propagation and decision fusion , author=. NeuroImage , volume=. 2006 , publisher=
2006
-
[64]
elegans , author=
NeuroPAL: A Neuronal Polychromatic Atlas of Landmarks for Whole-Brain Imaging in C. elegans , author=. BioRxiv , pages=. 2019 , publisher=
2019
-
[65]
arXiv preprint arXiv:1810.00828 , year=
Singularity, misspecification, and the convergence rate of EM , author=. arXiv preprint arXiv:1810.00828 , year=
-
[66]
Proceedings of the 32nd International Conference on Neural Information Processing Systems , pages=
Theoretical guarantees for the EM algorithm when applied to mis-specified Gaussian mixture models , author=. Proceedings of the 32nd International Conference on Neural Information Processing Systems , pages=
-
[67]
Cell , volume=
Global brain dynamics embed the motor command sequence of Caenorhabditis elegans , author=. Cell , volume=. 2015 , publisher=
2015
-
[68]
arXiv preprint arXiv:1908.10935 , year=
Randomly initialized EM algorithm for two-component Gaussian mixture achieves near optimality in O ( sqrt \ n \ ) iterations , author=. arXiv preprint arXiv:1908.10935 , year=
-
[69]
Advances in Neural Information Processing Systems , pages=
Global analysis of expectation maximization for mixtures of two gaussians , author=. Advances in Neural Information Processing Systems , pages=
-
[70]
NeuRIPS , pages=
Local maxima in the likelihood of gaussian mixture models: Structural results and algorithmic consequences , author=. NeuRIPS , pages=
-
[71]
Developmental biology , volume=
The embryonic cell lineage of the nematode Caenorhabditis elegans , author=. Developmental biology , volume=
-
[72]
Ten Steps of
Constantinos Daskalakis and Christos Tzamos and Manolis Zampetakis , Bibsource =. Ten Steps of. Proceedings of the 30th Conference on Learning Theory,. 2017 , Bdsk-Url-1 =
2017
-
[73]
2017 , Bdsk-Url-1 =
Proceedings of the 30th Conference on Learning Theory,. 2017 , Bdsk-Url-1 =
2017
-
[74]
Balakrishnan, Sivaraman and Wainwright, Martin J. and Yu, Bin , Date-Added =. Statistical guarantees for the. Ann. Statist. , Mrclass =. 2017 , Bdsk-Url-1 =. doi:10.1214/16-AOS1435 , Fjournal =
-
[75]
On-line expectation-maximization algorithm for latent data models , Url =
Capp\'. On-line expectation-maximization algorithm for latent data models , Url =. J. R. Stat. Soc. Ser. B Stat. Methodol. , Mrclass =. 2009 , Bdsk-Url-1 =. doi:10.1111/j.1467-9868.2009.00698.x , Fjournal =
-
[76]
The variational approximation for
Tzikas, Dimitris G and Likas, Aristidis C and Galatsanos, Nikolaos P , Date-Added =. The variational approximation for. IEEE Signal Processing Magazine , Number =
-
[77]
and Krishnan, Thriyambakam , Date-Added =
McLachlan, Geoffrey J. and Krishnan, Thriyambakam , Date-Added =. The. 2008 , Bdsk-Url-1 =. doi:10.1002/9780470191613 , Edition =
-
[78]
elegans Neurons , author=
Probabilistic Joint Segmentation and Labeling of C. elegans Neurons , author=. International Conference on Medical Image Computing and Computer-Assisted Intervention , pages=. 2020 , organization=
2020
-
[79]
Information geonetry and alternating minimization procedures , Volume =
Csisz. Information geonetry and alternating minimization procedures , Volume =. Statistics and decisions , Pages =
-
[80]
Semidual regularized optimal transport , Volume =
Cuturi, Marco and Peyr. Semidual regularized optimal transport , Volume =. SIAM Review , Number =
-
[81]
Mixture densities, maximum likelihood and the
Redner, Richard A and Walker, Homer F , Journal =. Mixture densities, maximum likelihood and the
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.