<inline-formula> <tex-math notation="LaTeX">$\beta$ </tex-math></inline-formula>-Dropout: A Unified Dropout

Dropout is an effective regularization method for deep learning tasks. Several variants of dropout based on sampling with different distributions have been proposed individually and have shown good generalization performance on various learning tasks. Among these variants, the canonical Bernoulli dr...

Full description

Bibliographic Details
Main Authors: Lei Liu, Yuhao Luo, Xu Shen, Mingzhai Sun, Bin Li
Format: Article
Language:English
Published: IEEE 2019-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/8666975/
_version_ 1831805960551661568
author Lei Liu
Yuhao Luo
Xu Shen
Mingzhai Sun
Bin Li
author_facet Lei Liu
Yuhao Luo
Xu Shen
Mingzhai Sun
Bin Li
author_sort Lei Liu
collection DOAJ
description Dropout is an effective regularization method for deep learning tasks. Several variants of dropout based on sampling with different distributions have been proposed individually and have shown good generalization performance on various learning tasks. Among these variants, the canonical Bernoulli dropout is a discrete method, while the uniform dropout and the Gaussian dropout are continuous dropout methods. When facing a new learning task, one must make a decision on which method is more suitable, which is somehow unnatural and inconvenient. In this paper, we attempt to change the selection problem to a parameter tuning problem by proposing a general form of dropout, &#x03B2;-dropout, to unify the discrete dropout with continuous dropout. We show that by adjusting the shape parameter &#x03B2;, the &#x03B2;-dropout can yield the Bernoulli dropout, uniform dropout, and approximate Gaussian dropout. Furthermore, it can obtain continuous regularization strength, which paves the way for self-adaptive dropout regularization. As a first attempt, we propose a self-adaptive &#x03B2;-dropout, in which the parameter &#x03B2; is tuned automatically following a pre-designed strategy. The &#x03B2;-dropout is tested extensively on the MNIST, CIFAR-10, SVHN, NORB, and ILSVRC-12 datasets to investigate its superior performance. The results show that the &#x03B2;-dropout can conduct finer control of its regularization strength, therefore obtaining better performance.
first_indexed 2024-12-22T19:32:38Z
format Article
id doaj.art-d999cb8aaefe40ac8709562b7391ccd6
institution Directory Open Access Journal
issn 2169-3536
language English
last_indexed 2024-12-22T19:32:38Z
publishDate 2019-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj.art-d999cb8aaefe40ac8709562b7391ccd62022-12-21T18:15:03ZengIEEEIEEE Access2169-35362019-01-017361403615310.1109/ACCESS.2019.29048818666975<inline-formula> <tex-math notation="LaTeX">$\beta$ </tex-math></inline-formula>-Dropout: A Unified DropoutLei Liu0Yuhao Luo1Xu Shen2Mingzhai Sun3https://orcid.org/0000-0002-0522-5511Bin Li4https://orcid.org/0000-0002-2332-3959CAS Key Laboratory of Technology in Geo-spatial Information Processing and Application System, University of Science and Technology of China, Hefei, ChinaDepartment of Precision Machinery and Precision Instrumentation, University of Science and Technology of China, Hefei, ChinaCAS Key Laboratory of Technology in Geo-spatial Information Processing and Application System, University of Science and Technology of China, Hefei, ChinaDepartment of Precision Machinery and Precision Instrumentation, University of Science and Technology of China, Hefei, ChinaCAS Key Laboratory of Technology in Geo-spatial Information Processing and Application System, University of Science and Technology of China, Hefei, ChinaDropout is an effective regularization method for deep learning tasks. Several variants of dropout based on sampling with different distributions have been proposed individually and have shown good generalization performance on various learning tasks. Among these variants, the canonical Bernoulli dropout is a discrete method, while the uniform dropout and the Gaussian dropout are continuous dropout methods. When facing a new learning task, one must make a decision on which method is more suitable, which is somehow unnatural and inconvenient. In this paper, we attempt to change the selection problem to a parameter tuning problem by proposing a general form of dropout, &#x03B2;-dropout, to unify the discrete dropout with continuous dropout. We show that by adjusting the shape parameter &#x03B2;, the &#x03B2;-dropout can yield the Bernoulli dropout, uniform dropout, and approximate Gaussian dropout. Furthermore, it can obtain continuous regularization strength, which paves the way for self-adaptive dropout regularization. As a first attempt, we propose a self-adaptive &#x03B2;-dropout, in which the parameter &#x03B2; is tuned automatically following a pre-designed strategy. The &#x03B2;-dropout is tested extensively on the MNIST, CIFAR-10, SVHN, NORB, and ILSVRC-12 datasets to investigate its superior performance. The results show that the &#x03B2;-dropout can conduct finer control of its regularization strength, therefore obtaining better performance.https://ieeexplore.ieee.org/document/8666975/Regularizationdropoutdeep learningGaussian dropoutBernoulli dropout
spellingShingle Lei Liu
Yuhao Luo
Xu Shen
Mingzhai Sun
Bin Li
<inline-formula> <tex-math notation="LaTeX">$\beta$ </tex-math></inline-formula>-Dropout: A Unified Dropout
IEEE Access
Regularization
dropout
deep learning
Gaussian dropout
Bernoulli dropout
title <inline-formula> <tex-math notation="LaTeX">$\beta$ </tex-math></inline-formula>-Dropout: A Unified Dropout
title_full <inline-formula> <tex-math notation="LaTeX">$\beta$ </tex-math></inline-formula>-Dropout: A Unified Dropout
title_fullStr <inline-formula> <tex-math notation="LaTeX">$\beta$ </tex-math></inline-formula>-Dropout: A Unified Dropout
title_full_unstemmed <inline-formula> <tex-math notation="LaTeX">$\beta$ </tex-math></inline-formula>-Dropout: A Unified Dropout
title_short <inline-formula> <tex-math notation="LaTeX">$\beta$ </tex-math></inline-formula>-Dropout: A Unified Dropout
title_sort inline formula tex math notation latex beta tex math inline formula dropout a unified dropout
topic Regularization
dropout
deep learning
Gaussian dropout
Bernoulli dropout
url https://ieeexplore.ieee.org/document/8666975/
work_keys_str_mv AT leiliu inlineformulatexmathnotationlatexbetatexmathinlineformuladropoutaunifieddropout
AT yuhaoluo inlineformulatexmathnotationlatexbetatexmathinlineformuladropoutaunifieddropout
AT xushen inlineformulatexmathnotationlatexbetatexmathinlineformuladropoutaunifieddropout
AT mingzhaisun inlineformulatexmathnotationlatexbetatexmathinlineformuladropoutaunifieddropout
AT binli inlineformulatexmathnotationlatexbetatexmathinlineformuladropoutaunifieddropout