pith. machine review for the scientific record. sign in

arxiv: 1703.01488 · v1 · submitted 2017-03-04 · 📊 stat.ML

Recognition: unknown

Autoencoding Variational Inference For Topic Models

Authors on Pith no claims yet
classification 📊 stat.ML
keywords topicinferencemodelmodelsvariationalaevbavitmfind
0
0 comments X
read the original abstract

Topic models are one of the most popular methods for learning representations of text, but a major challenge is that any change to the topic model requires mathematically deriving a new inference algorithm. A promising approach to address this problem is autoencoding variational Bayes (AEVB), but it has proven diffi- cult to apply to topic models in practice. We present what is to our knowledge the first effective AEVB based inference method for latent Dirichlet allocation (LDA), which we call Autoencoded Variational Inference For Topic Model (AVITM). This model tackles the problems caused for AEVB by the Dirichlet prior and by component collapsing. We find that AVITM matches traditional methods in accuracy with much better inference time. Indeed, because of the inference network, we find that it is unnecessary to pay the computational cost of running variational optimization on test data. Because AVITM is black box, it is readily applied to new topic models. As a dramatic illustration of this, we present a new topic model called ProdLDA, that replaces the mixture model in LDA with a product of experts. By changing only one line of code from LDA, we find that ProdLDA yields much more interpretable topics, even if LDA is trained via collapsed Gibbs sampling.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 2 Pith papers

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Many Wrongs Make a Right: Leveraging Biased Simulations Towards Unbiased Parameter Inference

    hep-ph 2026-04 unverdicted novelty 7.0

    Template-Adapted Mixture Model uses many biased simulations for data-driven estimates of signal and background distributions, yielding unbiased signal fraction estimates with well-calibrated uncertainties.

  2. HuggingFace's Transformers: State-of-the-art Natural Language Processing

    cs.CL 2019-10 accept novelty 6.0

    Hugging Face releases an open-source Python library that supplies a unified API and pretrained weights for major Transformer architectures used in natural language processing.