Ensemble Method of Diverse Regularized Extreme Learning Machines

As a fast training algorithm of single hidden layer forward networks, extreme learning machine (ELM) randomly initializes the input layer weights and hidden layer biases, and gets the weights of output layer through the analysis method. It overcomes many shortcomings of gradient based learning algor...

Full description

Bibliographic Details
Main Author: CHEN Yang, WANG Shitong
Format: Article
Language:zho
Published: Journal of Computer Engineering and Applications Beijing Co., Ltd., Science Press 2022-08-01
Series:Jisuanji kexue yu tansuo
Subjects:
Online Access:http://fcst.ceaj.org/fileup/1673-9418/PDF/2101001.pdf
_version_ 1818000573363912704
author CHEN Yang, WANG Shitong
author_facet CHEN Yang, WANG Shitong
author_sort CHEN Yang, WANG Shitong
collection DOAJ
description As a fast training algorithm of single hidden layer forward networks, extreme learning machine (ELM) randomly initializes the input layer weights and hidden layer biases, and gets the weights of output layer through the analysis method. It overcomes many shortcomings of gradient based learning algorithm, such as local minimum, inappropriate learning rate, slow learning speed, etc. However, ELM still inevitably has overfitting and poorly stable phenomenon, especially on large-scale datasets. This paper proposes the ensemble method of diverse regularized extreme learning machines (DRELM) to solve the above problems. First, its own random distribution weigthts are used to assure the diversity between each ELM base learner, then leave-one-out (LOO) cross validation method and M S E P R E S Smethod are used to find the optimal hidden node number of each base learner, calculate the optimal hidden layer output weights to train better and different base learners. Then the new penalty term about diversity is explicitly added to the objective function and the output matrix of each learner is updated iteratively. Finally, the final output of the whole network model is obtained by averaging the output of all base learners. This method can effectively realize the ensemble of regularized extreme learning machines (RELM) with both accuracy and diversity. Experimental results on 10 UCI datasets indicate the effectiveness of DRELM.
first_indexed 2024-04-14T03:23:12Z
format Article
id doaj.art-fdb336931b064523ba1c06ef750b1311
institution Directory Open Access Journal
issn 1673-9418
language zho
last_indexed 2024-04-14T03:23:12Z
publishDate 2022-08-01
publisher Journal of Computer Engineering and Applications Beijing Co., Ltd., Science Press
record_format Article
series Jisuanji kexue yu tansuo
spelling doaj.art-fdb336931b064523ba1c06ef750b13112022-12-22T02:15:15ZzhoJournal of Computer Engineering and Applications Beijing Co., Ltd., Science PressJisuanji kexue yu tansuo1673-94182022-08-011681819182810.3778/j.issn.1673-9418.2101001Ensemble Method of Diverse Regularized Extreme Learning MachinesCHEN Yang, WANG Shitong01. School of Artificial Intelligence and Computer Science, Jiangnan University, Wuxi, Jiangsu 214122, China;2. Key Laboratory of Media Design and Software Technology of Jiangsu Province, Jiangnan University, Wuxi, Jiangsu 214122, ChinaAs a fast training algorithm of single hidden layer forward networks, extreme learning machine (ELM) randomly initializes the input layer weights and hidden layer biases, and gets the weights of output layer through the analysis method. It overcomes many shortcomings of gradient based learning algorithm, such as local minimum, inappropriate learning rate, slow learning speed, etc. However, ELM still inevitably has overfitting and poorly stable phenomenon, especially on large-scale datasets. This paper proposes the ensemble method of diverse regularized extreme learning machines (DRELM) to solve the above problems. First, its own random distribution weigthts are used to assure the diversity between each ELM base learner, then leave-one-out (LOO) cross validation method and M S E P R E S Smethod are used to find the optimal hidden node number of each base learner, calculate the optimal hidden layer output weights to train better and different base learners. Then the new penalty term about diversity is explicitly added to the objective function and the output matrix of each learner is updated iteratively. Finally, the final output of the whole network model is obtained by averaging the output of all base learners. This method can effectively realize the ensemble of regularized extreme learning machines (RELM) with both accuracy and diversity. Experimental results on 10 UCI datasets indicate the effectiveness of DRELM.http://fcst.ceaj.org/fileup/1673-9418/PDF/2101001.pdf|extreme learning machine (elm)|ensemble learning|diversity|regularized extreme learning machines (relm)
spellingShingle CHEN Yang, WANG Shitong
Ensemble Method of Diverse Regularized Extreme Learning Machines
Jisuanji kexue yu tansuo
|extreme learning machine (elm)|ensemble learning|diversity|regularized extreme learning machines (relm)
title Ensemble Method of Diverse Regularized Extreme Learning Machines
title_full Ensemble Method of Diverse Regularized Extreme Learning Machines
title_fullStr Ensemble Method of Diverse Regularized Extreme Learning Machines
title_full_unstemmed Ensemble Method of Diverse Regularized Extreme Learning Machines
title_short Ensemble Method of Diverse Regularized Extreme Learning Machines
title_sort ensemble method of diverse regularized extreme learning machines
topic |extreme learning machine (elm)|ensemble learning|diversity|regularized extreme learning machines (relm)
url http://fcst.ceaj.org/fileup/1673-9418/PDF/2101001.pdf
work_keys_str_mv AT chenyangwangshitong ensemblemethodofdiverseregularizedextremelearningmachines