An improved multi-input deep convolutional neural network for automatic emotion recognition

Current decoding algorithms based on a one-dimensional (1D) convolutional neural network (CNN) have shown effectiveness in the automatic recognition of emotional tasks using physiological signals. However, these recognition models usually take a single modal of physiological signal as input, and the...

Full description

Bibliographic Details
Main Authors: Peiji Chen, Bochao Zou, Abdelkader Nasreddine Belkacem, Xiangwen Lyu, Xixi Zhao, Weibo Yi, Zhaoyang Huang, Jun Liang, Chao Chen
Format: Article
Language:English
Published: Frontiers Media S.A. 2022-10-01
Series:Frontiers in Neuroscience
Subjects:
Online Access:https://www.frontiersin.org/articles/10.3389/fnins.2022.965871/full
_version_ 1811181377336377344
author Peiji Chen
Bochao Zou
Abdelkader Nasreddine Belkacem
Xiangwen Lyu
Xixi Zhao
Weibo Yi
Zhaoyang Huang
Jun Liang
Chao Chen
Chao Chen
author_facet Peiji Chen
Bochao Zou
Abdelkader Nasreddine Belkacem
Xiangwen Lyu
Xixi Zhao
Weibo Yi
Zhaoyang Huang
Jun Liang
Chao Chen
Chao Chen
author_sort Peiji Chen
collection DOAJ
description Current decoding algorithms based on a one-dimensional (1D) convolutional neural network (CNN) have shown effectiveness in the automatic recognition of emotional tasks using physiological signals. However, these recognition models usually take a single modal of physiological signal as input, and the inter-correlates between different modalities of physiological signals are completely ignored, which could be an important source of information for emotion recognition. Therefore, a complete end-to-end multi-input deep convolutional neural network (MI-DCNN) structure was designed in this study. The newly designed 1D-CNN structure can take full advantage of multi-modal physiological signals and automatically complete the process from feature extraction to emotion classification simultaneously. To evaluate the effectiveness of the proposed model, we designed an emotion elicitation experiment and collected a total of 52 participants' physiological signals including electrocardiography (ECG), electrodermal activity (EDA), and respiratory activity (RSP) while watching emotion elicitation videos. Subsequently, traditional machine learning methods were applied as baseline comparisons; for arousal, the baseline accuracy and f1-score of our dataset were 62.9 ± 0.9% and 0.628 ± 0.01, respectively; for valence, the baseline accuracy and f1-score of our dataset were 60.3 ± 0.8% and 0.600 ± 0.01, respectively. Differences between the MI-DCNN and single-input DCNN were also compared, and the proposed method was verified on two public datasets (DEAP and DREAMER) as well as our dataset. The computing results in our dataset showed a significant improvement in both tasks compared to traditional machine learning methods (t-test, arousal: p = 9.7E-03 < 0.01, valence: 6.5E-03 < 0.01), which demonstrated the strength of introducing a multi-input convolutional neural network for emotion recognition based on multi-modal physiological signals.
first_indexed 2024-04-11T09:15:40Z
format Article
id doaj.art-1d1e1d85277c47178512f4844acd2f79
institution Directory Open Access Journal
issn 1662-453X
language English
last_indexed 2024-04-11T09:15:40Z
publishDate 2022-10-01
publisher Frontiers Media S.A.
record_format Article
series Frontiers in Neuroscience
spelling doaj.art-1d1e1d85277c47178512f4844acd2f792022-12-22T04:32:20ZengFrontiers Media S.A.Frontiers in Neuroscience1662-453X2022-10-011610.3389/fnins.2022.965871965871An improved multi-input deep convolutional neural network for automatic emotion recognitionPeiji Chen0Bochao Zou1Abdelkader Nasreddine Belkacem2Xiangwen Lyu3Xixi Zhao4Weibo Yi5Zhaoyang Huang6Jun Liang7Chao Chen8Chao Chen9Key Laboratory of Complex System Control Theory and Application, Tianjin University of Technology, Tianjin, ChinaSchool of Computer and Communication Engineering, University of Science and Technology, Beijing, ChinaDepartment of Computer and Network Engineering, College of Information Technology, United Arab Emirates University (UAEU), Al Ain, United Arab EmiratesNational Engineering Laboratory for Risk Perception and Prevention, Beijing, ChinaBeijing Key Laboratory of Mental Disorders, Beijing Anding Hospital, Capital Medical University, Beijing, ChinaBeijing Machine and Equipment Institute, Beijing, ChinaDepartment of Neurology, Xuanwu Hospital, Capital Medical University, Beijing, ChinaDepartment of Rehabilitation, Tianjin Medical University General Hospital, Tianjin, ChinaKey Laboratory of Complex System Control Theory and Application, Tianjin University of Technology, Tianjin, ChinaAcademy of Medical Engineering and Translational Medicine, Tianjin University, Tianjin, ChinaCurrent decoding algorithms based on a one-dimensional (1D) convolutional neural network (CNN) have shown effectiveness in the automatic recognition of emotional tasks using physiological signals. However, these recognition models usually take a single modal of physiological signal as input, and the inter-correlates between different modalities of physiological signals are completely ignored, which could be an important source of information for emotion recognition. Therefore, a complete end-to-end multi-input deep convolutional neural network (MI-DCNN) structure was designed in this study. The newly designed 1D-CNN structure can take full advantage of multi-modal physiological signals and automatically complete the process from feature extraction to emotion classification simultaneously. To evaluate the effectiveness of the proposed model, we designed an emotion elicitation experiment and collected a total of 52 participants' physiological signals including electrocardiography (ECG), electrodermal activity (EDA), and respiratory activity (RSP) while watching emotion elicitation videos. Subsequently, traditional machine learning methods were applied as baseline comparisons; for arousal, the baseline accuracy and f1-score of our dataset were 62.9 ± 0.9% and 0.628 ± 0.01, respectively; for valence, the baseline accuracy and f1-score of our dataset were 60.3 ± 0.8% and 0.600 ± 0.01, respectively. Differences between the MI-DCNN and single-input DCNN were also compared, and the proposed method was verified on two public datasets (DEAP and DREAMER) as well as our dataset. The computing results in our dataset showed a significant improvement in both tasks compared to traditional machine learning methods (t-test, arousal: p = 9.7E-03 < 0.01, valence: 6.5E-03 < 0.01), which demonstrated the strength of introducing a multi-input convolutional neural network for emotion recognition based on multi-modal physiological signals.https://www.frontiersin.org/articles/10.3389/fnins.2022.965871/fullbiological signalsmulti-modalityemotion recognitionconvolutional neural networkmachine learning
spellingShingle Peiji Chen
Bochao Zou
Abdelkader Nasreddine Belkacem
Xiangwen Lyu
Xixi Zhao
Weibo Yi
Zhaoyang Huang
Jun Liang
Chao Chen
Chao Chen
An improved multi-input deep convolutional neural network for automatic emotion recognition
Frontiers in Neuroscience
biological signals
multi-modality
emotion recognition
convolutional neural network
machine learning
title An improved multi-input deep convolutional neural network for automatic emotion recognition
title_full An improved multi-input deep convolutional neural network for automatic emotion recognition
title_fullStr An improved multi-input deep convolutional neural network for automatic emotion recognition
title_full_unstemmed An improved multi-input deep convolutional neural network for automatic emotion recognition
title_short An improved multi-input deep convolutional neural network for automatic emotion recognition
title_sort improved multi input deep convolutional neural network for automatic emotion recognition
topic biological signals
multi-modality
emotion recognition
convolutional neural network
machine learning
url https://www.frontiersin.org/articles/10.3389/fnins.2022.965871/full
work_keys_str_mv AT peijichen animprovedmultiinputdeepconvolutionalneuralnetworkforautomaticemotionrecognition
AT bochaozou animprovedmultiinputdeepconvolutionalneuralnetworkforautomaticemotionrecognition
AT abdelkadernasreddinebelkacem animprovedmultiinputdeepconvolutionalneuralnetworkforautomaticemotionrecognition
AT xiangwenlyu animprovedmultiinputdeepconvolutionalneuralnetworkforautomaticemotionrecognition
AT xixizhao animprovedmultiinputdeepconvolutionalneuralnetworkforautomaticemotionrecognition
AT weiboyi animprovedmultiinputdeepconvolutionalneuralnetworkforautomaticemotionrecognition
AT zhaoyanghuang animprovedmultiinputdeepconvolutionalneuralnetworkforautomaticemotionrecognition
AT junliang animprovedmultiinputdeepconvolutionalneuralnetworkforautomaticemotionrecognition
AT chaochen animprovedmultiinputdeepconvolutionalneuralnetworkforautomaticemotionrecognition
AT chaochen animprovedmultiinputdeepconvolutionalneuralnetworkforautomaticemotionrecognition
AT peijichen improvedmultiinputdeepconvolutionalneuralnetworkforautomaticemotionrecognition
AT bochaozou improvedmultiinputdeepconvolutionalneuralnetworkforautomaticemotionrecognition
AT abdelkadernasreddinebelkacem improvedmultiinputdeepconvolutionalneuralnetworkforautomaticemotionrecognition
AT xiangwenlyu improvedmultiinputdeepconvolutionalneuralnetworkforautomaticemotionrecognition
AT xixizhao improvedmultiinputdeepconvolutionalneuralnetworkforautomaticemotionrecognition
AT weiboyi improvedmultiinputdeepconvolutionalneuralnetworkforautomaticemotionrecognition
AT zhaoyanghuang improvedmultiinputdeepconvolutionalneuralnetworkforautomaticemotionrecognition
AT junliang improvedmultiinputdeepconvolutionalneuralnetworkforautomaticemotionrecognition
AT chaochen improvedmultiinputdeepconvolutionalneuralnetworkforautomaticemotionrecognition
AT chaochen improvedmultiinputdeepconvolutionalneuralnetworkforautomaticemotionrecognition