Hamiltonian Variational Auto-Encoder

Variational Auto-Encoders (VAEs) have become very popular techniques to perform inference and learning in latent variable models as they allow us to leverage the rich representational power of neural networks to obtain flexible approximations of the posterior of latent variables as well as tight evi...

Full description

Bibliographic Details
Main Authors: Caterini, A, Doucet, A, Sejdinovic, D
Format: Journal article
Published: Massachusetts Institute of Technology Press 2019
_version_ 1797062329046663168
author Caterini, A
Doucet, A
Sejdinovic, D
author_facet Caterini, A
Doucet, A
Sejdinovic, D
author_sort Caterini, A
collection OXFORD
description Variational Auto-Encoders (VAEs) have become very popular techniques to perform inference and learning in latent variable models as they allow us to leverage the rich representational power of neural networks to obtain flexible approximations of the posterior of latent variables as well as tight evidence lower bounds (ELBOs). Combined with stochastic variational inference, this provides a methodology scaling to large datasets. However, for this methodology to be practically efficient, it is necessary to obtain low-variance unbiased estimators of the ELBO and its gradients with respect to the parameters of interest. While the use of Markov chain Monte Carlo (MCMC) techniques such as Hamiltonian Monte Carlo (HMC) has been previously suggested to achieve this [23, 26], the proposed methods require specifying reverse kernels which have a large impact on performance. Additionally, the resulting unbiased estimator of the ELBO for most MCMC kernels is typically not amenable to the reparameterization trick. We show here how to optimally select reverse kernels in this setting and, by building upon Hamiltonian Importance Sampling (HIS) [17], we obtain a scheme that provides low-variance unbiased estimators of the ELBO and its gradients using the reparameterization trick. This allows us to develop a Hamiltonian Variational Auto-Encoder (HVAE). This method can be reinterpreted as a target-informed normalizing flow [20] which, within our context, only requires a few evaluations of the gradient of the sampled likelihood and trivial Jacobian calculations at each iteration.
first_indexed 2024-03-06T20:43:59Z
format Journal article
id oxford-uuid:353d868e-af18-466c-94bb-76238febcf01
institution University of Oxford
last_indexed 2024-03-06T20:43:59Z
publishDate 2019
publisher Massachusetts Institute of Technology Press
record_format dspace
spelling oxford-uuid:353d868e-af18-466c-94bb-76238febcf012022-03-26T13:30:47ZHamiltonian Variational Auto-EncoderJournal articlehttp://purl.org/coar/resource_type/c_dcae04bcuuid:353d868e-af18-466c-94bb-76238febcf01Symplectic Elements at OxfordMassachusetts Institute of Technology Press2019Caterini, ADoucet, ASejdinovic, DVariational Auto-Encoders (VAEs) have become very popular techniques to perform inference and learning in latent variable models as they allow us to leverage the rich representational power of neural networks to obtain flexible approximations of the posterior of latent variables as well as tight evidence lower bounds (ELBOs). Combined with stochastic variational inference, this provides a methodology scaling to large datasets. However, for this methodology to be practically efficient, it is necessary to obtain low-variance unbiased estimators of the ELBO and its gradients with respect to the parameters of interest. While the use of Markov chain Monte Carlo (MCMC) techniques such as Hamiltonian Monte Carlo (HMC) has been previously suggested to achieve this [23, 26], the proposed methods require specifying reverse kernels which have a large impact on performance. Additionally, the resulting unbiased estimator of the ELBO for most MCMC kernels is typically not amenable to the reparameterization trick. We show here how to optimally select reverse kernels in this setting and, by building upon Hamiltonian Importance Sampling (HIS) [17], we obtain a scheme that provides low-variance unbiased estimators of the ELBO and its gradients using the reparameterization trick. This allows us to develop a Hamiltonian Variational Auto-Encoder (HVAE). This method can be reinterpreted as a target-informed normalizing flow [20] which, within our context, only requires a few evaluations of the gradient of the sampled likelihood and trivial Jacobian calculations at each iteration.
spellingShingle Caterini, A
Doucet, A
Sejdinovic, D
Hamiltonian Variational Auto-Encoder
title Hamiltonian Variational Auto-Encoder
title_full Hamiltonian Variational Auto-Encoder
title_fullStr Hamiltonian Variational Auto-Encoder
title_full_unstemmed Hamiltonian Variational Auto-Encoder
title_short Hamiltonian Variational Auto-Encoder
title_sort hamiltonian variational auto encoder
work_keys_str_mv AT caterinia hamiltonianvariationalautoencoder
AT douceta hamiltonianvariationalautoencoder
AT sejdinovicd hamiltonianvariationalautoencoder