Tighter variational bounds are not necessarily better

We provide theoretical and empirical evidence that using tighter evidence lower bounds (ELBOs) can be detrimental to the process of learning an inference network by reducing the signal-to-noise ratio of the gradient estimator. Our results call into question common implicit assumptions that tighter E...

Full description

Bibliographic Details
Main Authors: Rainforth, T, Kosiorek, A, Le, T, Maddison, C, Igl, M, Wood, F, Teh, Y
Format: Conference item
Language:English
Published: Proceedings of Machine Learning Research 2018
_version_ 1797053824312016896
author Rainforth, T
Kosiorek, A
Le, T
Maddison, C
Igl, M
Wood, F
Teh, Y
author_facet Rainforth, T
Kosiorek, A
Le, T
Maddison, C
Igl, M
Wood, F
Teh, Y
author_sort Rainforth, T
collection OXFORD
description We provide theoretical and empirical evidence that using tighter evidence lower bounds (ELBOs) can be detrimental to the process of learning an inference network by reducing the signal-to-noise ratio of the gradient estimator. Our results call into question common implicit assumptions that tighter ELBOs are better variational objectives for simultaneous model learning and inference amortization schemes. Based on our insights, we introduce three new algorithms: the partially importance weighted auto-encoder (PIWAE), the multiply importance weighted auto-encoder (MIWAE), and the combination importance weighted autoencoder (CIWAE), each of which includes the standard importance weighted auto-encoder (IWAE) as a special case. We show that each can deliver improvements over IWAE, even when performance is measured by the IWAE target itself. Furthermore, our results suggest that PIWAE may be able to deliver simultaneous improvements in the training of both the inference and generative networks.
first_indexed 2024-03-06T18:48:59Z
format Conference item
id oxford-uuid:0f862a36-5288-422c-b39a-46feaf717a9c
institution University of Oxford
language English
last_indexed 2024-03-06T18:48:59Z
publishDate 2018
publisher Proceedings of Machine Learning Research
record_format dspace
spelling oxford-uuid:0f862a36-5288-422c-b39a-46feaf717a9c2022-03-26T09:51:40ZTighter variational bounds are not necessarily betterConference itemhttp://purl.org/coar/resource_type/c_5794uuid:0f862a36-5288-422c-b39a-46feaf717a9cEnglishSymplectic Elements at OxfordProceedings of Machine Learning Research2018Rainforth, TKosiorek, ALe, TMaddison, CIgl, MWood, FTeh, YWe provide theoretical and empirical evidence that using tighter evidence lower bounds (ELBOs) can be detrimental to the process of learning an inference network by reducing the signal-to-noise ratio of the gradient estimator. Our results call into question common implicit assumptions that tighter ELBOs are better variational objectives for simultaneous model learning and inference amortization schemes. Based on our insights, we introduce three new algorithms: the partially importance weighted auto-encoder (PIWAE), the multiply importance weighted auto-encoder (MIWAE), and the combination importance weighted autoencoder (CIWAE), each of which includes the standard importance weighted auto-encoder (IWAE) as a special case. We show that each can deliver improvements over IWAE, even when performance is measured by the IWAE target itself. Furthermore, our results suggest that PIWAE may be able to deliver simultaneous improvements in the training of both the inference and generative networks.
spellingShingle Rainforth, T
Kosiorek, A
Le, T
Maddison, C
Igl, M
Wood, F
Teh, Y
Tighter variational bounds are not necessarily better
title Tighter variational bounds are not necessarily better
title_full Tighter variational bounds are not necessarily better
title_fullStr Tighter variational bounds are not necessarily better
title_full_unstemmed Tighter variational bounds are not necessarily better
title_short Tighter variational bounds are not necessarily better
title_sort tighter variational bounds are not necessarily better
work_keys_str_mv AT rainfortht tightervariationalboundsarenotnecessarilybetter
AT kosioreka tightervariationalboundsarenotnecessarilybetter
AT let tightervariationalboundsarenotnecessarilybetter
AT maddisonc tightervariationalboundsarenotnecessarilybetter
AT iglm tightervariationalboundsarenotnecessarilybetter
AT woodf tightervariationalboundsarenotnecessarilybetter
AT tehy tightervariationalboundsarenotnecessarilybetter