Recognition: unknown
Automatic Posterior Transformation for Likelihood-Free Inference
read the original abstract
How can one perform Bayesian inference on stochastic simulators with intractable likelihoods? A recent approach is to learn the posterior from adaptively proposed simulations using neural network-based conditional density estimators. However, existing methods are limited to a narrow range of proposal distributions or require importance weighting that can limit performance in practice. Here we present automatic posterior transformation (APT), a new sequential neural posterior estimation method for simulation-based inference. APT can modify the posterior estimate using arbitrary, dynamically updated proposals, and is compatible with powerful flow-based density estimators. It is more flexible, scalable and efficient than previous simulation-based inference techniques. APT can operate directly on high-dimensional time series and image data, opening up new applications for likelihood-free inference.
This paper has not been read by Pith yet.
Forward citations
Cited by 3 Pith papers
-
Neural Posterior Estimation for UHECR source inference from 3D propagation simulations
A Deep Set encoder plus normalizing flow model trained on five million CRPropa 3 events recovers UHECR source parameters without bias and classifies primary composition at over 98 percent accuracy.
-
Probing the IMF in the Early Universe -- Direct measurements in the Bo\"otes I UFD with JWST/NIRCam
The low-mass IMF in Boötes I is consistent with the Milky Way within 68% confidence for broken power-law and lognormal forms, indicating universality at low metallicity.
-
Spectroscopy of analogue black holes using simulation-based inference
Simulation-based inference reliably extracts physical parameters from noisy spectra of analogue black holes.
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.