Bayesian Optimization Based Efficient Layer Sharing for Incremental Learning

Incremental learning is a methodology that continuously uses the sequential input data to extend the existing network’s knowledge. The layer sharing algorithm is one of the representative methods which leverages general knowledge by sharing some initial layers of the existing network. To determine t...

Full description

Bibliographic Details
Main Authors: Bomi Kim, Taehyeon Kim, Yoonsik Choe
Format: Article
Language:English
Published: MDPI AG 2021-03-01
Series:Applied Sciences
Subjects:
Online Access:https://www.mdpi.com/2076-3417/11/5/2171
_version_ 1797416520315305984
author Bomi Kim
Taehyeon Kim
Yoonsik Choe
author_facet Bomi Kim
Taehyeon Kim
Yoonsik Choe
author_sort Bomi Kim
collection DOAJ
description Incremental learning is a methodology that continuously uses the sequential input data to extend the existing network’s knowledge. The layer sharing algorithm is one of the representative methods which leverages general knowledge by sharing some initial layers of the existing network. To determine the performance of the incremental network, it is critical to estimate how much the initial convolutional layers in the existing network can be shared as the fixed feature extractors. However, the existing algorithm selects the sharing configuration through improper optimization strategy but a brute force manner such as searching for all possible sharing layers case. This is a non-convex and non-differential problem. Accordingly, this can not be solved using powerful optimization techniques such as the gradient descent algorithm or other convex optimization problem, and it leads to high computational complexity. To solve this problem, we firstly define this as a discrete combinatorial optimization problem, and propose a novel efficient incremental learning algorithm-based Bayesian optimization, which guarantees the global convergence in a non-convex and non-differential optimization. Additionally, our proposed algorithm can adaptively find the optimal number of sharing layers via adjusting the threshold accuracy parameter in the proposed loss function. With the proposed method, the global optimal sharing layer can be found in only six or eight iterations without searching for all possible layer cases. Hence, the proposed method can find the global optimal sharing layers by utilizing Bayesian optimization, which achieves both high combined accuracy and low computational complexity.
first_indexed 2024-03-09T06:05:24Z
format Article
id doaj.art-6c24c7a9be324cf0b78a3d47fa729395
institution Directory Open Access Journal
issn 2076-3417
language English
last_indexed 2024-03-09T06:05:24Z
publishDate 2021-03-01
publisher MDPI AG
record_format Article
series Applied Sciences
spelling doaj.art-6c24c7a9be324cf0b78a3d47fa7293952023-12-03T12:04:41ZengMDPI AGApplied Sciences2076-34172021-03-01115217110.3390/app11052171Bayesian Optimization Based Efficient Layer Sharing for Incremental LearningBomi Kim0Taehyeon Kim1Yoonsik Choe2Department of Electrical and Electronic Engineering, Yonsei University, Seoul 03722, KoreaDepartment of Electrical and Electronic Engineering, Yonsei University, Seoul 03722, KoreaDepartment of Electrical and Electronic Engineering, Yonsei University, Seoul 03722, KoreaIncremental learning is a methodology that continuously uses the sequential input data to extend the existing network’s knowledge. The layer sharing algorithm is one of the representative methods which leverages general knowledge by sharing some initial layers of the existing network. To determine the performance of the incremental network, it is critical to estimate how much the initial convolutional layers in the existing network can be shared as the fixed feature extractors. However, the existing algorithm selects the sharing configuration through improper optimization strategy but a brute force manner such as searching for all possible sharing layers case. This is a non-convex and non-differential problem. Accordingly, this can not be solved using powerful optimization techniques such as the gradient descent algorithm or other convex optimization problem, and it leads to high computational complexity. To solve this problem, we firstly define this as a discrete combinatorial optimization problem, and propose a novel efficient incremental learning algorithm-based Bayesian optimization, which guarantees the global convergence in a non-convex and non-differential optimization. Additionally, our proposed algorithm can adaptively find the optimal number of sharing layers via adjusting the threshold accuracy parameter in the proposed loss function. With the proposed method, the global optimal sharing layer can be found in only six or eight iterations without searching for all possible layer cases. Hence, the proposed method can find the global optimal sharing layers by utilizing Bayesian optimization, which achieves both high combined accuracy and low computational complexity.https://www.mdpi.com/2076-3417/11/5/2171Bayesian optimizationincremental learninglayer sharing algorithm
spellingShingle Bomi Kim
Taehyeon Kim
Yoonsik Choe
Bayesian Optimization Based Efficient Layer Sharing for Incremental Learning
Applied Sciences
Bayesian optimization
incremental learning
layer sharing algorithm
title Bayesian Optimization Based Efficient Layer Sharing for Incremental Learning
title_full Bayesian Optimization Based Efficient Layer Sharing for Incremental Learning
title_fullStr Bayesian Optimization Based Efficient Layer Sharing for Incremental Learning
title_full_unstemmed Bayesian Optimization Based Efficient Layer Sharing for Incremental Learning
title_short Bayesian Optimization Based Efficient Layer Sharing for Incremental Learning
title_sort bayesian optimization based efficient layer sharing for incremental learning
topic Bayesian optimization
incremental learning
layer sharing algorithm
url https://www.mdpi.com/2076-3417/11/5/2171
work_keys_str_mv AT bomikim bayesianoptimizationbasedefficientlayersharingforincrementallearning
AT taehyeonkim bayesianoptimizationbasedefficientlayersharingforincrementallearning
AT yoonsikchoe bayesianoptimizationbasedefficientlayersharingforincrementallearning