Concrete dropout

Dropout is used as a practical tool to obtain uncertainty estimates in large vision models and reinforcement learning (RL) tasks. But to obtain well-calibrated uncertainty estimates, a grid-search over the dropout probabilities is necessary— a prohibitive operation with large models, and an impossib...

Disgrifiad llawn

Manylion Llyfryddiaeth
Prif Awduron: Gal, Y, Hron, J, Kendall, A
Fformat: Conference item
Cyhoeddwyd: NIPS Foundation 2018
_version_ 1826264049770299392
author Gal, Y
Hron, J
Kendall, A
author_facet Gal, Y
Hron, J
Kendall, A
author_sort Gal, Y
collection OXFORD
description Dropout is used as a practical tool to obtain uncertainty estimates in large vision models and reinforcement learning (RL) tasks. But to obtain well-calibrated uncertainty estimates, a grid-search over the dropout probabilities is necessary— a prohibitive operation with large models, and an impossible one with RL. We propose a new dropout variant which gives improved performance and better calibrated uncertainties. Relying on recent developments in Bayesian deep learning, we use a continuous relaxation of dropout’s discrete masks. Together with a principled optimisation objective, this allows for automatic tuning of the dropout probability in large models, and as a result faster experimentation cycles. In RL this allows the agent to adapt its uncertainty dynamically as more data is observed. We analyse the proposed variant extensively on a range of tasks, and give insights into common practice in the field where larger dropout probabilities are often used in deeper model layers.
first_indexed 2024-03-06T20:01:34Z
format Conference item
id oxford-uuid:2779c391-6a38-4c70-be60-7b0c8c88a1a2
institution University of Oxford
last_indexed 2024-03-06T20:01:34Z
publishDate 2018
publisher NIPS Foundation
record_format dspace
spelling oxford-uuid:2779c391-6a38-4c70-be60-7b0c8c88a1a22022-03-26T12:07:10ZConcrete dropoutConference itemhttp://purl.org/coar/resource_type/c_5794uuid:2779c391-6a38-4c70-be60-7b0c8c88a1a2Symplectic Elements at OxfordNIPS Foundation2018Gal, YHron, JKendall, ADropout is used as a practical tool to obtain uncertainty estimates in large vision models and reinforcement learning (RL) tasks. But to obtain well-calibrated uncertainty estimates, a grid-search over the dropout probabilities is necessary— a prohibitive operation with large models, and an impossible one with RL. We propose a new dropout variant which gives improved performance and better calibrated uncertainties. Relying on recent developments in Bayesian deep learning, we use a continuous relaxation of dropout’s discrete masks. Together with a principled optimisation objective, this allows for automatic tuning of the dropout probability in large models, and as a result faster experimentation cycles. In RL this allows the agent to adapt its uncertainty dynamically as more data is observed. We analyse the proposed variant extensively on a range of tasks, and give insights into common practice in the field where larger dropout probabilities are often used in deeper model layers.
spellingShingle Gal, Y
Hron, J
Kendall, A
Concrete dropout
title Concrete dropout
title_full Concrete dropout
title_fullStr Concrete dropout
title_full_unstemmed Concrete dropout
title_short Concrete dropout
title_sort concrete dropout
work_keys_str_mv AT galy concretedropout
AT hronj concretedropout
AT kendalla concretedropout