pith. machine review for the scientific record. sign in

arxiv: 1405.5505 · v3 · submitted 2014-05-21 · 📊 stat.ML · cs.LG

Recognition: unknown

Kernel Mean Shrinkage Estimators

Authors on Pith no claims yet
classification 📊 stat.ML cs.LG
keywords kernelmeanestimatorsempiricalestimatormethodsshrinkagestandard
0
0 comments X
read the original abstract

A mean function in a reproducing kernel Hilbert space (RKHS), or a kernel mean, is central to kernel methods in that it is used by many classical algorithms such as kernel principal component analysis, and it also forms the core inference step of modern kernel methods that rely on embedding probability distributions in RKHSs. Given a finite sample, an empirical average has been used commonly as a standard estimator of the true kernel mean. Despite a widespread use of this estimator, we show that it can be improved thanks to the well-known Stein phenomenon. We propose a new family of estimators called kernel mean shrinkage estimators (KMSEs), which benefit from both theoretical justifications and good empirical performance. The results demonstrate that the proposed estimators outperform the standard one, especially in a "large d, small n" paradigm.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 1 Pith paper

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Variance Matters: Improving Domain Adaptation via Stratified Sampling

    cs.LG 2025-12 unverdicted novelty 6.0

    VaRDASS improves unsupervised domain adaptation by using stratified sampling to reduce variance in discrepancy estimation for measures like correlation alignment and MMD, with derived error bounds, an optimality proof...