An Efficient Algorithm for the Incremental Broad Learning System by Inverse Cholesky Factorization of a Partitioned Matrix

In this paper, we propose an efficient algorithm to accelerate the existing Broad Learning System (BLS) algorithm for new added nodes. The existing BLS algorithm computes the output weights from the pseudoinverse with the ridge regression approximation, and updates the pseudoinverse iteratively. As...

Full description

Bibliographic Details
Main Authors: Hufei Zhu, Zhulin Liu, C. L. Philip Chen, Yanyang Liang
Format: Article
Language:English
Published: IEEE 2021-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/9326429/
_version_ 1811210596317659136
author Hufei Zhu
Zhulin Liu
C. L. Philip Chen
Yanyang Liang
author_facet Hufei Zhu
Zhulin Liu
C. L. Philip Chen
Yanyang Liang
author_sort Hufei Zhu
collection DOAJ
description In this paper, we propose an efficient algorithm to accelerate the existing Broad Learning System (BLS) algorithm for new added nodes. The existing BLS algorithm computes the output weights from the pseudoinverse with the ridge regression approximation, and updates the pseudoinverse iteratively. As a comparison, the proposed BLS algorithm computes the output weights from the inverse Cholesky factor of the Hermitian matrix in the calculation of the pseudoinverse, and updates the inverse Cholesky factor efficiently. Since the Hermitian matrix in the definition of the pseudoinverse is smaller than the pseudoinverse, the proposed BLS algorithm can reduce the computational complexity, and usually requires less than <inline-formula> <tex-math notation="LaTeX">$\frac {2}{3}$ </tex-math></inline-formula> of complexities with respect to the existing BLS algorithm. Our experiments on the Modified National Institute of Standards and Technology (MNIST) dataset show that the speedups in accumulative training time and each additional training time of the proposed BLS over the existing BLS are 24.81&#x0025;&#x007E; 37.99&#x0025; and 36.45&#x0025;&#x007E; 58.96&#x0025;, respectively, and the speedup in total training time is 37.99&#x0025;. In our experiments, the proposed BLS and the existing BLS both achieve the same testing accuracy when the tiny differences (&#x2264; 0.05&#x0025;) caused by the numerical errors are neglected, and the above-mentioned tiny differences and numerical errors become zeroes and ignorable, respectively, when the ridge parameter is not too small.
first_indexed 2024-04-12T04:58:36Z
format Article
id doaj.art-ddb72c27b20d4877895f359987d641bd
institution Directory Open Access Journal
issn 2169-3536
language English
last_indexed 2024-04-12T04:58:36Z
publishDate 2021-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj.art-ddb72c27b20d4877895f359987d641bd2022-12-22T03:47:03ZengIEEEIEEE Access2169-35362021-01-019192941930310.1109/ACCESS.2021.30521029326429An Efficient Algorithm for the Incremental Broad Learning System by Inverse Cholesky Factorization of a Partitioned MatrixHufei Zhu0https://orcid.org/0000-0002-1629-6227Zhulin Liu1https://orcid.org/0000-0003-4145-823XC. L. Philip Chen2Yanyang Liang3Faculty of Intelligent Manufacturing, Wuyi University, Jiangmen, ChinaSchool of Computer Science and Engineering, South China University of Technology, Guangzhou, ChinaSchool of Computer Science and Engineering, South China University of Technology, Guangzhou, ChinaFaculty of Intelligent Manufacturing, Wuyi University, Jiangmen, ChinaIn this paper, we propose an efficient algorithm to accelerate the existing Broad Learning System (BLS) algorithm for new added nodes. The existing BLS algorithm computes the output weights from the pseudoinverse with the ridge regression approximation, and updates the pseudoinverse iteratively. As a comparison, the proposed BLS algorithm computes the output weights from the inverse Cholesky factor of the Hermitian matrix in the calculation of the pseudoinverse, and updates the inverse Cholesky factor efficiently. Since the Hermitian matrix in the definition of the pseudoinverse is smaller than the pseudoinverse, the proposed BLS algorithm can reduce the computational complexity, and usually requires less than <inline-formula> <tex-math notation="LaTeX">$\frac {2}{3}$ </tex-math></inline-formula> of complexities with respect to the existing BLS algorithm. Our experiments on the Modified National Institute of Standards and Technology (MNIST) dataset show that the speedups in accumulative training time and each additional training time of the proposed BLS over the existing BLS are 24.81&#x0025;&#x007E; 37.99&#x0025; and 36.45&#x0025;&#x007E; 58.96&#x0025;, respectively, and the speedup in total training time is 37.99&#x0025;. In our experiments, the proposed BLS and the existing BLS both achieve the same testing accuracy when the tiny differences (&#x2264; 0.05&#x0025;) caused by the numerical errors are neglected, and the above-mentioned tiny differences and numerical errors become zeroes and ignorable, respectively, when the ridge parameter is not too small.https://ieeexplore.ieee.org/document/9326429/Broad learning system (BLS)incremental learningadded nodesrandom vector functional-link neural networks (RVFLNN)single layer feedforward neural networks (SLFN)efficient algorithms
spellingShingle Hufei Zhu
Zhulin Liu
C. L. Philip Chen
Yanyang Liang
An Efficient Algorithm for the Incremental Broad Learning System by Inverse Cholesky Factorization of a Partitioned Matrix
IEEE Access
Broad learning system (BLS)
incremental learning
added nodes
random vector functional-link neural networks (RVFLNN)
single layer feedforward neural networks (SLFN)
efficient algorithms
title An Efficient Algorithm for the Incremental Broad Learning System by Inverse Cholesky Factorization of a Partitioned Matrix
title_full An Efficient Algorithm for the Incremental Broad Learning System by Inverse Cholesky Factorization of a Partitioned Matrix
title_fullStr An Efficient Algorithm for the Incremental Broad Learning System by Inverse Cholesky Factorization of a Partitioned Matrix
title_full_unstemmed An Efficient Algorithm for the Incremental Broad Learning System by Inverse Cholesky Factorization of a Partitioned Matrix
title_short An Efficient Algorithm for the Incremental Broad Learning System by Inverse Cholesky Factorization of a Partitioned Matrix
title_sort efficient algorithm for the incremental broad learning system by inverse cholesky factorization of a partitioned matrix
topic Broad learning system (BLS)
incremental learning
added nodes
random vector functional-link neural networks (RVFLNN)
single layer feedforward neural networks (SLFN)
efficient algorithms
url https://ieeexplore.ieee.org/document/9326429/
work_keys_str_mv AT hufeizhu anefficientalgorithmfortheincrementalbroadlearningsystembyinversecholeskyfactorizationofapartitionedmatrix
AT zhulinliu anefficientalgorithmfortheincrementalbroadlearningsystembyinversecholeskyfactorizationofapartitionedmatrix
AT clphilipchen anefficientalgorithmfortheincrementalbroadlearningsystembyinversecholeskyfactorizationofapartitionedmatrix
AT yanyangliang anefficientalgorithmfortheincrementalbroadlearningsystembyinversecholeskyfactorizationofapartitionedmatrix
AT hufeizhu efficientalgorithmfortheincrementalbroadlearningsystembyinversecholeskyfactorizationofapartitionedmatrix
AT zhulinliu efficientalgorithmfortheincrementalbroadlearningsystembyinversecholeskyfactorizationofapartitionedmatrix
AT clphilipchen efficientalgorithmfortheincrementalbroadlearningsystembyinversecholeskyfactorizationofapartitionedmatrix
AT yanyangliang efficientalgorithmfortheincrementalbroadlearningsystembyinversecholeskyfactorizationofapartitionedmatrix