pith. machine review for the scientific record. sign in

arxiv: 1112.5745 · v1 · submitted 2011-12-24 · 📊 stat.ML · cs.LG

Recognition: unknown

Bayesian Active Learning for Classification and Preference Learning

Authors on Pith no claims yet
classification 📊 stat.ML cs.LG
keywords learninginformationactiveclassificationmodelspreferencetheoreticapproach
0
0 comments X
read the original abstract

Information theoretic active learning has been widely studied for probabilistic models. For simple regression an optimal myopic policy is easily tractable. However, for other tasks and with more complex models, such as classification with nonparametric models, the optimal solution is harder to compute. Current approaches make approximations to achieve tractability. We propose an approach that expresses information gain in terms of predictive entropies, and apply this method to the Gaussian Process Classifier (GPC). Our approach makes minimal approximations to the full information theoretic objective. Our experimental performance compares favourably to many popular active learning algorithms, and has equal or lower computational complexity. We compare well to decision theoretic approaches also, which are privy to more information and require much more computational time. Secondly, by developing further a reformulation of binary preference learning to a classification problem, we extend our algorithm to Gaussian Process preference learning.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 17 Pith papers

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. The Minimax Rate of Second-Order Calibration

    cs.LG 2026-05 unverdicted novelty 8.0

    The minimax rate of estimating second-order calibration error is Õ(1/√n) with a matching Ω(1/√n) lower bound, enabled by analyticity from the sech kernel and yielding the first finite-sample guarantee for second-order...

  2. Active Learning for Gaussian Process Regression Under Self-Induced Boltzmann Weights

    cs.LG 2026-05 unverdicted novelty 7.0

    AB-SID-iVAR enables Gaussian process active learning for self-induced Boltzmann distributions by closed-form approximation of the target, with high-probability error vanishing guarantees and empirical gains on PES and...

  3. LLM-guided Semi-Supervised Approaches for Social Media Crisis Data Classification

    cs.AI 2026-05 conditional novelty 7.0

    LG-CoTrain, an LLM-guided co-training method, outperforms classical semi-supervised baselines for crisis tweet classification in low-resource settings with 5-25 labeled examples per class.

  4. Active Learning for Communication Structure Optimization in LLM-Based Multi-Agent Systems

    cs.MA 2026-05 unverdicted novelty 7.0

    An ensemble-based information-theoretic active learning method with ensemble Kalman inversion selects valuable tasks to optimize communication structures in LLM multi-agent systems under constrained budgets.

  5. InfoChess: A Game of Adversarial Inference and a Laboratory for Quantifiable Information Control

    cs.MA 2026-04 unverdicted novelty 7.0

    InfoChess proposes a symmetric adversarial game focused purely on information control and probabilistic king-location inference, with RL agents outperforming heuristic baselines and gameplay dissected via belief entro...

  6. Do We Still Need Humans in the Loop? Comparing Human and LLM Annotation in Active Learning for Hostility Detection

    cs.CL 2026-04 conditional novelty 7.0

    LLM annotation can replace human labels for hostility detection with comparable F1 at much lower cost, but active learning adds little value and error structures differ systematically.

  7. Epistemic Uncertainty for Test-Time Discovery

    cs.LG 2026-05 unverdicted novelty 6.0

    UG-TTT adds epistemic uncertainty measured by adapter disagreement as an exploration bonus in RL for LLMs, raising maximum reward and diversity on scientific discovery benchmarks.

  8. Decoupled PFNs: Identifiable Epistemic-Aleatoric Decomposition via Structured Synthetic Priors

    stat.ML 2026-05 conditional novelty 6.0

    Decoupled PFNs use controllable synthetic priors to train separate latent-signal and noise heads, making epistemic-aleatoric decomposition identifiable and improving acquisition in noisy settings.

  9. Active Learning for Communication Structure Optimization in LLM-Based Multi-Agent Systems

    cs.MA 2026-05 unverdicted novelty 6.0

    An ensemble-based information-theoretic active learning method using ensemble Kalman inversion selects valuable tasks to optimize communication structures in LLM multi-agent systems more reliably than random sampling ...

  10. ProEval: Proactive Failure Discovery and Efficient Performance Estimation for Generative AI Evaluation

    cs.LG 2026-04 unverdicted novelty 6.0

    ProEval is a proactive framework using pre-trained GPs, Bayesian quadrature, and superlevel set sampling to estimate performance and find failures in generative AI with 8-65x fewer samples than baselines.

  11. Vibrotactile Preference Learning: Uncertainty-Aware Preference Learning for Personalized Vibration Feedback

    cs.HC 2026-04 unverdicted novelty 6.0

    VPL learns individualized vibrotactile preferences efficiently via uncertainty-aware Gaussian process models and active query selection in a 13-participant user study on an Xbox controller.

  12. Boundary-Centric Active Learning for Temporal Action Segmentation

    cs.CV 2026-04 unverdicted novelty 6.0

    B-ACT improves label efficiency in temporal action segmentation by selecting only boundary frames for annotation via a two-stage uncertainty-driven process that fuses neighborhood uncertainty, class ambiguity, and tem...

  13. Agentic Discovery with Active Hypothesis Exploration for Visual Recognition

    cs.CV 2026-04 unverdicted novelty 6.0

    HypoExplore uses LLMs for hypothesis-driven evolutionary search with a Trajectory Tree and Hypothesis Memory Bank to discover lightweight vision architectures, reaching 94.11% accuracy on CIFAR-10 from an 18.91% basel...

  14. RADS: Reinforcement Learning-Based Sample Selection Improves Transfer Learning in Low-resource and Imbalanced Clinical Settings

    cs.CL 2026-04 unverdicted novelty 5.0

    RADS applies reinforcement learning to pick informative samples for transfer learning, improving performance over uncertainty and diversity sampling in low-resource imbalanced clinical settings.

  15. When Active Learning Falls Short: An Empirical Study on Chemical Reaction Extraction

    cs.LG 2026-04 unverdicted novelty 5.0

    Active learning for chemical reaction extraction frequently produces non-monotonic learning curves and fails to deliver stable gains over random sampling because of strong pretraining, structured CRF decoding, and lab...

  16. Testing the Assumptions of Active Learning for Translation Tasks with Few Samples

    cs.CL 2026-04 unverdicted novelty 5.0

    Informativeness and diversity of samples selected by active learning show no correlation with test performance on translation tasks using few samples; ordering and pre-training effects dominate instead.

  17. Feature Weighting Improves Pool-Based Sequential Active Learning for Regression

    cs.LG 2026-04 conditional novelty 4.0

    Feature weighting derived from ridge regression coefficients improves sample selection in pool-based sequential active learning for both single-task and multi-task regression.