pith. machine review for the scientific record. sign in

arxiv: 1102.5750 · v1 · submitted 2011-02-28 · 📊 stat.ML · cs.LG· math.ST· stat.TH

Recognition: unknown

Neyman-Pearson classification, convexity and stochastic constraints

Authors on Pith no claims yet
classification 📊 stat.ML cs.LGmath.STstat.TH
keywords probabilityclassificationclassifierempiricalerrorneyman-pearsonproblemstype
0
0 comments X
read the original abstract

Motivated by problems of anomaly detection, this paper implements the Neyman-Pearson paradigm to deal with asymmetric errors in binary classification with a convex loss. Given a finite collection of classifiers, we combine them and obtain a new classifier that satisfies simultaneously the two following properties with high probability: (i) its probability of type I error is below a pre-specified level and (ii), it has probability of type II error close to the minimum possible. The proposed classifier is obtained by solving an optimization problem with an empirical objective and an empirical constraint. New techniques to handle such problems are developed and have consequences on chance constrained programming.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.