Efficient Feature Selection Using Weighted Superposition Attraction Optimization Algorithm

As the volume of data generated by information systems continues to increase, machine learning (ML) techniques have become essential for the extraction of meaningful insights. However, the sheer volume of data often causes these techniques to become sluggish. To overcome this, feature selection is a...

Full description

Bibliographic Details
Main Authors: Narayanan Ganesh, Rajendran Shankar, Robert Čep, Shankar Chakraborty, Kanak Kalita
Format: Article
Language:English
Published: MDPI AG 2023-03-01
Series:Applied Sciences
Subjects:
Online Access:https://www.mdpi.com/2076-3417/13/5/3223
Description
Summary:As the volume of data generated by information systems continues to increase, machine learning (ML) techniques have become essential for the extraction of meaningful insights. However, the sheer volume of data often causes these techniques to become sluggish. To overcome this, feature selection is a vital step in the pre-processing of data. In this paper, we introduce a novel K-nearest neighborhood (KNN)-based wrapper system for feature selection that leverages the iterative improvement ability of the weighted superposition attraction (WSA). We evaluate the performance of WSA against seven well-known metaheuristic algorithms, i.e., differential evolution (DE), genetic algorithm (GA), particle swarm optimization (PSO), flower pollination algorithm (FPA), symbiotic organisms search (SOS), marine predators’ algorithm (MPA) and manta ray foraging optimization (MRFO). Our extensive numerical experiments demonstrate that WSA is highly effective for feature selection, achieving a decrease of up to 99% in the number of features for large datasets without sacrificing classification accuracy. In fact, WSA-KNN outperforms traditional ML methods by about 18% and ensemble ML algorithms by 9%. Moreover, WSA-KNN achieves comparable or slightly better solutions when compared with neural networks hybridized with metaheuristics. These findings highlight the importance and potential of WSA for feature selection in modern-day data processing systems.
ISSN:2076-3417