Ensemble Feature Selection Method with Fast Transfer Model

Compared with the traditional ensemble feature selection methods, the recently-developed ensemble feature selection with block-regularized [m×2] cross-validation (EFSBCV) not only has a variance of the estimator smaller than that of random [m×2] cross-validation, but also enhances the selection prob...

Full description

Bibliographic Details
Main Author: NING Baobin, WANG Shitong
Format: Article
Language:zho
Published: Journal of Computer Engineering and Applications Beijing Co., Ltd., Science Press 2024-02-01
Series:Jisuanji kexue yu tansuo
Subjects:
Online Access:http://fcst.ceaj.org/fileup/1673-9418/PDF/2211073.pdf
_version_ 1797334195547144192
author NING Baobin, WANG Shitong
author_facet NING Baobin, WANG Shitong
author_sort NING Baobin, WANG Shitong
collection DOAJ
description Compared with the traditional ensemble feature selection methods, the recently-developed ensemble feature selection with block-regularized [m×2] cross-validation (EFSBCV) not only has a variance of the estimator smaller than that of random [m×2] cross-validation, but also enhances the selection probability of important features and reduces the selection probability of noise features. However, the adopted linear regression model without the use of the bias term in EFSBCV may easily lead to underfitting. Moreover, EFSBCV does not consider the importance of each feature subset. Aiming at these two problems, an ensemble feature selection method called EFSFT (ensemble feature selection method using fast transfer model) is proposed in this paper. The basic idea is that the base feature selector in EFSBCV adopts the fast transfer model in this paper, so as to introduce the bias term. EFSFT transfers 2m subsets of features as the source knowledge, and then recalculates the weight of each feature subset, and the linear model fitting ability with the addition of bias terms is better. The results on real datasets show that compared with EFSBCV, the average FP value by EFSFT reduces up to 58%, proving that EFSFT has more advantages in removing noise features. In contrast to least-squares support vector machine (LSSVM), the average TP value by EFSFT increases up to 5%, which clearly indicates the superiority of EFSFT over LSSVM in choosing important features.
first_indexed 2024-03-08T08:17:15Z
format Article
id doaj.art-4d212dcc5130468d92e3e8ec7f63cae3
institution Directory Open Access Journal
issn 1673-9418
language zho
last_indexed 2024-03-08T08:17:15Z
publishDate 2024-02-01
publisher Journal of Computer Engineering and Applications Beijing Co., Ltd., Science Press
record_format Article
series Jisuanji kexue yu tansuo
spelling doaj.art-4d212dcc5130468d92e3e8ec7f63cae32024-02-02T07:10:08ZzhoJournal of Computer Engineering and Applications Beijing Co., Ltd., Science PressJisuanji kexue yu tansuo1673-94182024-02-0118249650510.3778/j.issn.1673-9418.2211073Ensemble Feature Selection Method with Fast Transfer ModelNING Baobin, WANG Shitong0School of Artificial Intelligence and Computer Science, Jiangnan University, Wuxi, Jiangsu 214122, ChinaCompared with the traditional ensemble feature selection methods, the recently-developed ensemble feature selection with block-regularized [m×2] cross-validation (EFSBCV) not only has a variance of the estimator smaller than that of random [m×2] cross-validation, but also enhances the selection probability of important features and reduces the selection probability of noise features. However, the adopted linear regression model without the use of the bias term in EFSBCV may easily lead to underfitting. Moreover, EFSBCV does not consider the importance of each feature subset. Aiming at these two problems, an ensemble feature selection method called EFSFT (ensemble feature selection method using fast transfer model) is proposed in this paper. The basic idea is that the base feature selector in EFSBCV adopts the fast transfer model in this paper, so as to introduce the bias term. EFSFT transfers 2m subsets of features as the source knowledge, and then recalculates the weight of each feature subset, and the linear model fitting ability with the addition of bias terms is better. The results on real datasets show that compared with EFSBCV, the average FP value by EFSFT reduces up to 58%, proving that EFSFT has more advantages in removing noise features. In contrast to least-squares support vector machine (LSSVM), the average TP value by EFSFT increases up to 5%, which clearly indicates the superiority of EFSFT over LSSVM in choosing important features.http://fcst.ceaj.org/fileup/1673-9418/PDF/2211073.pdfensemble feature selection; cross-validation; transfer learning; regression
spellingShingle NING Baobin, WANG Shitong
Ensemble Feature Selection Method with Fast Transfer Model
Jisuanji kexue yu tansuo
ensemble feature selection; cross-validation; transfer learning; regression
title Ensemble Feature Selection Method with Fast Transfer Model
title_full Ensemble Feature Selection Method with Fast Transfer Model
title_fullStr Ensemble Feature Selection Method with Fast Transfer Model
title_full_unstemmed Ensemble Feature Selection Method with Fast Transfer Model
title_short Ensemble Feature Selection Method with Fast Transfer Model
title_sort ensemble feature selection method with fast transfer model
topic ensemble feature selection; cross-validation; transfer learning; regression
url http://fcst.ceaj.org/fileup/1673-9418/PDF/2211073.pdf
work_keys_str_mv AT ningbaobinwangshitong ensemblefeatureselectionmethodwithfasttransfermodel