pith. machine review for the scientific record. sign in

arxiv: 1806.01768 · v3 · submitted 2018-06-05 · 💻 cs.LG · stat.ML

Recognition: unknown

Evidential Deep Learning to Quantify Classification Uncertainty

Authors on Pith no claims yet
classification 💻 cs.LG stat.ML
keywords neuralpredictionuncertaintyclassificationdeterministicdirichletdistributionfunction
0
0 comments X
read the original abstract

Deterministic neural nets have been shown to learn effective predictors on a wide range of machine learning problems. However, as the standard approach is to train the network to minimize a prediction loss, the resultant model remains ignorant to its prediction confidence. Orthogonally to Bayesian neural nets that indirectly infer prediction uncertainty through weight uncertainties, we propose explicit modeling of the same using the theory of subjective logic. By placing a Dirichlet distribution on the class probabilities, we treat predictions of a neural net as subjective opinions and learn the function that collects the evidence leading to these opinions by a deterministic neural net from data. The resultant predictor for a multi-class classification problem is another Dirichlet distribution whose parameters are set by the continuous output of a neural net. We provide a preliminary analysis on how the peculiarities of our new loss function drive improved uncertainty estimation. We observe that our method achieves unprecedented success on detection of out-of-distribution queries and endurance against adversarial perturbations.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 4 Pith papers

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Rethinking Vacuity for OOD Detection in Evidential Deep Learning

    cs.AI 2026-05 accept novelty 7.0

    Vacuity-based OOD detection in evidential deep learning is highly sensitive to class cardinality differences between ID and OOD, which can artificially inflate AUROC and AUPR without any change in model predictions.

  2. Ensemble-Based Dirichlet Modeling for Predictive Uncertainty and Selective Classification

    stat.ML 2026-04 unverdicted novelty 6.0

    Ensemble-based method of moments on softmax outputs produces stable Dirichlet predictive distributions that improve uncertainty-guided tasks like selective classification over evidential deep learning.

  3. MedFormer-UR: Uncertainty-Routed Transformer for Medical Image Classification

    eess.IV 2026-04 unverdicted novelty 5.0

    MedFormer-UR integrates evidential uncertainty from Dirichlet distributions and class-specific prototypes into a transformer to improve calibration and selective prediction on medical images across four modalities.

  4. PriorNet: Prior-Guided Engagement Estimation from Face Video

    cs.CV 2026-05 unverdicted novelty 4.0

    PriorNet improves engagement estimation from face videos by injecting priors into preprocessing, adaptation, and objective design, showing improvements on multiple benchmarks.