Multiclass classification of imagined speech EEG using noise-assisted multivariate empirical mode decomposition and multireceptive field convolutional neural network

IntroductionIn this study, we classified electroencephalography (EEG) data of imagined speech using signal decomposition and multireceptive convolutional neural network. The imagined speech EEG with five vowels /a/, /e/, /i/, /o/, and /u/, and mute (rest) sounds were obtained from ten study particip...

Full description

Bibliographic Details
Main Authors: Hyeong-jun Park, Boreom Lee
Format: Article
Language:English
Published: Frontiers Media S.A. 2023-08-01
Series:Frontiers in Human Neuroscience
Subjects:
Online Access:https://www.frontiersin.org/articles/10.3389/fnhum.2023.1186594/full
_version_ 1797743158166028288
author Hyeong-jun Park
Boreom Lee
Boreom Lee
author_facet Hyeong-jun Park
Boreom Lee
Boreom Lee
author_sort Hyeong-jun Park
collection DOAJ
description IntroductionIn this study, we classified electroencephalography (EEG) data of imagined speech using signal decomposition and multireceptive convolutional neural network. The imagined speech EEG with five vowels /a/, /e/, /i/, /o/, and /u/, and mute (rest) sounds were obtained from ten study participants.Materials and methodsFirst, two different signal decomposition methods were applied for comparison: noise-assisted multivariate empirical mode decomposition and wavelet packet decomposition. Six statistical features were calculated from the decomposed eight sub-frequency bands EEG. Next, all features obtained from each channel of the trial were vectorized and used as the input vector of classifiers. Lastly, EEG was classified using multireceptive field convolutional neural network and several other classifiers for comparison.ResultsWe achieved an average classification rate of 73.09 and up to 80.41% in a multiclass (six classes) setup (Chance: 16.67%). In comparison with various other classifiers, significant improvements for other classifiers were achieved (p-value < 0.05). From the frequency sub-band analysis, high-frequency band regions and the lowest-frequency band region contain more information about imagined vowel EEG data. The misclassification and classification rate of each vowel imaginary EEG was analyzed through a confusion matrix.DiscussionImagined speech EEG can be classified successfully using the proposed signal decomposition method and a convolutional neural network. The proposed classification method for imagined speech EEG can contribute to developing a practical imagined speech-based brain-computer interfaces system.
first_indexed 2024-03-12T14:51:29Z
format Article
id doaj.art-3388d149d3004efe918f099756eb4d97
institution Directory Open Access Journal
issn 1662-5161
language English
last_indexed 2024-03-12T14:51:29Z
publishDate 2023-08-01
publisher Frontiers Media S.A.
record_format Article
series Frontiers in Human Neuroscience
spelling doaj.art-3388d149d3004efe918f099756eb4d972023-08-15T09:04:55ZengFrontiers Media S.A.Frontiers in Human Neuroscience1662-51612023-08-011710.3389/fnhum.2023.11865941186594Multiclass classification of imagined speech EEG using noise-assisted multivariate empirical mode decomposition and multireceptive field convolutional neural networkHyeong-jun Park0Boreom Lee1Boreom Lee2Department of Biomedical Science and Engineering, Gwangju Institute of Science and Technology, Gwangju, Republic of KoreaDepartment of Biomedical Science and Engineering, Gwangju Institute of Science and Technology, Gwangju, Republic of KoreaAI Graduate School, Gwangju Institute of Science and Technology, Gwangju, Republic of KoreaIntroductionIn this study, we classified electroencephalography (EEG) data of imagined speech using signal decomposition and multireceptive convolutional neural network. The imagined speech EEG with five vowels /a/, /e/, /i/, /o/, and /u/, and mute (rest) sounds were obtained from ten study participants.Materials and methodsFirst, two different signal decomposition methods were applied for comparison: noise-assisted multivariate empirical mode decomposition and wavelet packet decomposition. Six statistical features were calculated from the decomposed eight sub-frequency bands EEG. Next, all features obtained from each channel of the trial were vectorized and used as the input vector of classifiers. Lastly, EEG was classified using multireceptive field convolutional neural network and several other classifiers for comparison.ResultsWe achieved an average classification rate of 73.09 and up to 80.41% in a multiclass (six classes) setup (Chance: 16.67%). In comparison with various other classifiers, significant improvements for other classifiers were achieved (p-value < 0.05). From the frequency sub-band analysis, high-frequency band regions and the lowest-frequency band region contain more information about imagined vowel EEG data. The misclassification and classification rate of each vowel imaginary EEG was analyzed through a confusion matrix.DiscussionImagined speech EEG can be classified successfully using the proposed signal decomposition method and a convolutional neural network. The proposed classification method for imagined speech EEG can contribute to developing a practical imagined speech-based brain-computer interfaces system.https://www.frontiersin.org/articles/10.3389/fnhum.2023.1186594/fullbrain-computer interfacesimagined speech EEGmulticlass classificationmultireceptive field convolutional neural networknoise-assisted empirical mode decomposition
spellingShingle Hyeong-jun Park
Boreom Lee
Boreom Lee
Multiclass classification of imagined speech EEG using noise-assisted multivariate empirical mode decomposition and multireceptive field convolutional neural network
Frontiers in Human Neuroscience
brain-computer interfaces
imagined speech EEG
multiclass classification
multireceptive field convolutional neural network
noise-assisted empirical mode decomposition
title Multiclass classification of imagined speech EEG using noise-assisted multivariate empirical mode decomposition and multireceptive field convolutional neural network
title_full Multiclass classification of imagined speech EEG using noise-assisted multivariate empirical mode decomposition and multireceptive field convolutional neural network
title_fullStr Multiclass classification of imagined speech EEG using noise-assisted multivariate empirical mode decomposition and multireceptive field convolutional neural network
title_full_unstemmed Multiclass classification of imagined speech EEG using noise-assisted multivariate empirical mode decomposition and multireceptive field convolutional neural network
title_short Multiclass classification of imagined speech EEG using noise-assisted multivariate empirical mode decomposition and multireceptive field convolutional neural network
title_sort multiclass classification of imagined speech eeg using noise assisted multivariate empirical mode decomposition and multireceptive field convolutional neural network
topic brain-computer interfaces
imagined speech EEG
multiclass classification
multireceptive field convolutional neural network
noise-assisted empirical mode decomposition
url https://www.frontiersin.org/articles/10.3389/fnhum.2023.1186594/full
work_keys_str_mv AT hyeongjunpark multiclassclassificationofimaginedspeecheegusingnoiseassistedmultivariateempiricalmodedecompositionandmultireceptivefieldconvolutionalneuralnetwork
AT boreomlee multiclassclassificationofimaginedspeecheegusingnoiseassistedmultivariateempiricalmodedecompositionandmultireceptivefieldconvolutionalneuralnetwork
AT boreomlee multiclassclassificationofimaginedspeecheegusingnoiseassistedmultivariateempiricalmodedecompositionandmultireceptivefieldconvolutionalneuralnetwork