Recognition: unknown
NIPS 2016 Tutorial: Generative Adversarial Networks
read the original abstract
This report summarizes the tutorial presented by the author at NIPS 2016 on generative adversarial networks (GANs). The tutorial describes: (1) Why generative modeling is a topic worth studying, (2) how generative models work, and how GANs compare to other generative models, (3) the details of how GANs work, (4) research frontiers in GANs, and (5) state-of-the-art image models that combine GANs with other methods. Finally, the tutorial contains three exercises for readers to complete, and the solutions to these exercises.
This paper has not been read by Pith yet.
Forward citations
Cited by 3 Pith papers
-
Differentiable free energy surface: a variational approach to directly observing rare events using generative deep-learning models
VaFES constructs a latent space from reversible collective variables and variationally optimizes a tractable-density generative model to produce a continuous free energy surface from which rare events are directly sampled.
-
GAN-based Domain Adaptation for Image-aware Layout Generation in Advertising Poster Design
PDA-GAN with pixel discriminator bridges domain gap from inpainted posters to generate SOTA image-aware layouts on a new 60k-pair CGL-Dataset.
-
Conditional Wasserstein GAN for Simulating Neutrino Event Summaries using Incident Energy of Electron Neutrinos
A conditional Wasserstein GAN generates complete kinematic event summaries for IBD-CC, NC, and NuEElastic electron neutrino interactions that match GENIE distributions in 1D marginals and correlations.
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.