Expanding the capabilities of normalizing flows in deep generative models and variational inference

<p>Deep generative models and variational Bayesian inference are two frameworks for reasoning about observed high-dimensional data, which may even be combined. A fundamental requirement of either approach is the parametrization of an expressive family of density models. Normalizing flows, some...

Full description

Bibliographic Details
Main Author: Caterini, AL
Other Authors: Doucet, A
Format: Thesis
Language:English
Published: 2021
Subjects:
_version_ 1826312266886152192
author Caterini, AL
author2 Doucet, A
author_facet Doucet, A
Caterini, AL
author_sort Caterini, AL
collection OXFORD
description <p>Deep generative models and variational Bayesian inference are two frameworks for reasoning about observed high-dimensional data, which may even be combined. A fundamental requirement of either approach is the parametrization of an expressive family of density models. Normalizing flows, sometimes also referred to as invertible neural networks, are one class of models providing this: they are formulated to be bijective and differentiable, and thus produce a tractable density model via the change-of-variable formula. Beyond just deep generative modelling and variational inference, normalizing flows have shown promise as a plug-in density model in other settings such as approximate Bayesian computation and lossless compression. However, the bijectivity constraint can pose quite a restriction on the expressiveness of these approaches, and forces the learned distribution to have full support over the ambient space which is not well-aligned with the common assumption that low-dimensional manifold structure is embedded within high-dimensional data.</p> <p>In this thesis, we challenge this requirement of strict bijectivity over the space of interest to modify normalizing flow models. The first work focuses on the setting of variational inference, defining a normalizing flow based on a discretized time-inhomogeneous Hamiltonian dynamics over an extended position-momentum space. This enables the flow to be guided by the true posterior unlike baseline flow-based models, thus requiring fewer parameters in the inference model to achieve comparable improvements in inference. The next chapter proposes a new deep generative model which relaxes the bijectivity requirement of normalizing flows by injecting learned noise at each layer, sacrificing easy evaluation of the density for expressiveness. We show, theoretically and empirically, the benefits of these models in density estimation over baseline flows. We then demonstrate in the next chapter that the benefits of this model class extend to the setting of variational inference, relying on auxiliary methods to train our models. Finally, the last paper in this thesis addresses the issue of full support in the ambient space and proposes injective flow models directly embedding low-dimensional structure into high dimensions. Our method is the first to optimize the injective change-of-variable term and produces promising results on out-of-distribution detection, which had previous eluded deep generative models. We conclude with some directions for future work and a broader perspective on the field.</p>
first_indexed 2024-03-07T08:24:59Z
format Thesis
id oxford-uuid:9ed6cfab-be3d-4c6b-bbbd-99ff2d169ab5
institution University of Oxford
language English
last_indexed 2024-03-07T08:24:59Z
publishDate 2021
record_format dspace
spelling oxford-uuid:9ed6cfab-be3d-4c6b-bbbd-99ff2d169ab52024-02-12T11:44:48ZExpanding the capabilities of normalizing flows in deep generative models and variational inferenceThesishttp://purl.org/coar/resource_type/c_db06uuid:9ed6cfab-be3d-4c6b-bbbd-99ff2d169ab5variational inferencegenerative modellingnormalizing flowsdensity estimationEnglishHyrax Deposit2021Caterini, ALDoucet, ASejdinovic, DTeh, YBarber, D<p>Deep generative models and variational Bayesian inference are two frameworks for reasoning about observed high-dimensional data, which may even be combined. A fundamental requirement of either approach is the parametrization of an expressive family of density models. Normalizing flows, sometimes also referred to as invertible neural networks, are one class of models providing this: they are formulated to be bijective and differentiable, and thus produce a tractable density model via the change-of-variable formula. Beyond just deep generative modelling and variational inference, normalizing flows have shown promise as a plug-in density model in other settings such as approximate Bayesian computation and lossless compression. However, the bijectivity constraint can pose quite a restriction on the expressiveness of these approaches, and forces the learned distribution to have full support over the ambient space which is not well-aligned with the common assumption that low-dimensional manifold structure is embedded within high-dimensional data.</p> <p>In this thesis, we challenge this requirement of strict bijectivity over the space of interest to modify normalizing flow models. The first work focuses on the setting of variational inference, defining a normalizing flow based on a discretized time-inhomogeneous Hamiltonian dynamics over an extended position-momentum space. This enables the flow to be guided by the true posterior unlike baseline flow-based models, thus requiring fewer parameters in the inference model to achieve comparable improvements in inference. The next chapter proposes a new deep generative model which relaxes the bijectivity requirement of normalizing flows by injecting learned noise at each layer, sacrificing easy evaluation of the density for expressiveness. We show, theoretically and empirically, the benefits of these models in density estimation over baseline flows. We then demonstrate in the next chapter that the benefits of this model class extend to the setting of variational inference, relying on auxiliary methods to train our models. Finally, the last paper in this thesis addresses the issue of full support in the ambient space and proposes injective flow models directly embedding low-dimensional structure into high dimensions. Our method is the first to optimize the injective change-of-variable term and produces promising results on out-of-distribution detection, which had previous eluded deep generative models. We conclude with some directions for future work and a broader perspective on the field.</p>
spellingShingle variational inference
generative modelling
normalizing flows
density estimation
Caterini, AL
Expanding the capabilities of normalizing flows in deep generative models and variational inference
title Expanding the capabilities of normalizing flows in deep generative models and variational inference
title_full Expanding the capabilities of normalizing flows in deep generative models and variational inference
title_fullStr Expanding the capabilities of normalizing flows in deep generative models and variational inference
title_full_unstemmed Expanding the capabilities of normalizing flows in deep generative models and variational inference
title_short Expanding the capabilities of normalizing flows in deep generative models and variational inference
title_sort expanding the capabilities of normalizing flows in deep generative models and variational inference
topic variational inference
generative modelling
normalizing flows
density estimation
work_keys_str_mv AT caterinial expandingthecapabilitiesofnormalizingflowsindeepgenerativemodelsandvariationalinference