Disentangling disentanglement in variational autoencoders

We develop a generalisation of disentanglement in variational autoencoders (VAEs)—decomposition of the latent representation—characterising it as the fulfilment of two factors: a) the latent encodings of the data having an appropriate level of overlap, and b) the aggregate encoding of the data confo...

全面介绍

书目详细资料
Main Authors: Mathieu, E, Rainforth, T, Siddharth, N, Teh, Y
格式: Conference item
出版: PMLR 2019
_version_ 1826282660861837312
author Mathieu, E
Rainforth, T
Siddharth, N
Teh, Y
author_facet Mathieu, E
Rainforth, T
Siddharth, N
Teh, Y
author_sort Mathieu, E
collection OXFORD
description We develop a generalisation of disentanglement in variational autoencoders (VAEs)—decomposition of the latent representation—characterising it as the fulfilment of two factors: a) the latent encodings of the data having an appropriate level of overlap, and b) the aggregate encoding of the data conforming to a desired structure, represented through the prior. Decomposition permits disentanglement, i.e. explicit independence between latents, as a special case, but also allows for a much richer class of properties to be imposed on the learnt representation, such as sparsity, clustering, independent subspaces, or even intricate hierarchical dependency relationships. We show that the β-VAE varies from the standard VAE predominantly in its control of latent overlap and that for the standard choice of an isotropic Gaussian prior, its objective is invariant to rotations of the latent representation. Viewed from the decomposition perspective, breaking this invariance with simple manipulations of the prior can yield better disentanglement with little or no detriment to reconstructions. We further demonstrate how other choices of prior can assist in producing different decompositions and introduce an alternative training objective that allows the control of both decomposition factors in a principled manner.
first_indexed 2024-03-07T00:47:12Z
format Conference item
id oxford-uuid:851a88f4-fb16-4349-a729-edb51eba8782
institution University of Oxford
last_indexed 2024-03-07T00:47:12Z
publishDate 2019
publisher PMLR
record_format dspace
spelling oxford-uuid:851a88f4-fb16-4349-a729-edb51eba87822022-03-26T21:55:13ZDisentangling disentanglement in variational autoencodersConference itemhttp://purl.org/coar/resource_type/c_5794uuid:851a88f4-fb16-4349-a729-edb51eba8782Symplectic Elements at OxfordPMLR2019Mathieu, ERainforth, TSiddharth, NTeh, YWe develop a generalisation of disentanglement in variational autoencoders (VAEs)—decomposition of the latent representation—characterising it as the fulfilment of two factors: a) the latent encodings of the data having an appropriate level of overlap, and b) the aggregate encoding of the data conforming to a desired structure, represented through the prior. Decomposition permits disentanglement, i.e. explicit independence between latents, as a special case, but also allows for a much richer class of properties to be imposed on the learnt representation, such as sparsity, clustering, independent subspaces, or even intricate hierarchical dependency relationships. We show that the β-VAE varies from the standard VAE predominantly in its control of latent overlap and that for the standard choice of an isotropic Gaussian prior, its objective is invariant to rotations of the latent representation. Viewed from the decomposition perspective, breaking this invariance with simple manipulations of the prior can yield better disentanglement with little or no detriment to reconstructions. We further demonstrate how other choices of prior can assist in producing different decompositions and introduce an alternative training objective that allows the control of both decomposition factors in a principled manner.
spellingShingle Mathieu, E
Rainforth, T
Siddharth, N
Teh, Y
Disentangling disentanglement in variational autoencoders
title Disentangling disentanglement in variational autoencoders
title_full Disentangling disentanglement in variational autoencoders
title_fullStr Disentangling disentanglement in variational autoencoders
title_full_unstemmed Disentangling disentanglement in variational autoencoders
title_short Disentangling disentanglement in variational autoencoders
title_sort disentangling disentanglement in variational autoencoders
work_keys_str_mv AT mathieue disentanglingdisentanglementinvariationalautoencoders
AT rainfortht disentanglingdisentanglementinvariationalautoencoders
AT siddharthn disentanglingdisentanglementinvariationalautoencoders
AT tehy disentanglingdisentanglementinvariationalautoencoders