Speech Emotion Recognition Based on Two-Stream Deep Learning Model Using Korean Audio Information
Identifying a person’s emotions is an important element in communication. In particular, voice is a means of communication for easily and naturally expressing emotions. Speech emotion recognition technology is a crucial component of human–computer interaction (HCI), in which accurately identifying e...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2023-02-01
|
Series: | Applied Sciences |
Subjects: | |
Online Access: | https://www.mdpi.com/2076-3417/13/4/2167 |
_version_ | 1827758881953021952 |
---|---|
author | A-Hyeon Jo Keun-Chang Kwak |
author_facet | A-Hyeon Jo Keun-Chang Kwak |
author_sort | A-Hyeon Jo |
collection | DOAJ |
description | Identifying a person’s emotions is an important element in communication. In particular, voice is a means of communication for easily and naturally expressing emotions. Speech emotion recognition technology is a crucial component of human–computer interaction (HCI), in which accurately identifying emotions is key. Therefore, this study presents a two-stream-based emotion recognition model based on bidirectional long short-term memory (Bi-LSTM) and convolutional neural networks (CNNs) using a Korean speech emotion database, and the performance is comparatively analyzed. The data used in the experiment were obtained from the Korean speech emotion recognition database built by Chosun University. Two deep learning models, Bi-LSTM and YAMNet, which is a CNN-based transfer learning model, were connected in a two-stream architecture to design an emotion recognition model. Various speech feature extraction methods and deep learning models were compared in terms of performance. Consequently, the speech emotion recognition performance of Bi-LSTM and YAMNet was 90.38% and 94.91%, respectively. However, the performance of the two-stream model was 96%, which was a minimum of 1.09% and up to 5.62% improved compared with a single model. |
first_indexed | 2024-03-11T09:12:48Z |
format | Article |
id | doaj.art-33fd04da5ca94008b234853881718abd |
institution | Directory Open Access Journal |
issn | 2076-3417 |
language | English |
last_indexed | 2024-03-11T09:12:48Z |
publishDate | 2023-02-01 |
publisher | MDPI AG |
record_format | Article |
series | Applied Sciences |
spelling | doaj.art-33fd04da5ca94008b234853881718abd2023-11-16T18:51:52ZengMDPI AGApplied Sciences2076-34172023-02-01134216710.3390/app13042167Speech Emotion Recognition Based on Two-Stream Deep Learning Model Using Korean Audio InformationA-Hyeon Jo0Keun-Chang Kwak1Electronic Engineering IT-Bio Convergence System Major, Chosun University, Gwangju 61452, Republic of KoreaElectronic Engineering, Chosun University, Gwangju 61452, Republic of KoreaIdentifying a person’s emotions is an important element in communication. In particular, voice is a means of communication for easily and naturally expressing emotions. Speech emotion recognition technology is a crucial component of human–computer interaction (HCI), in which accurately identifying emotions is key. Therefore, this study presents a two-stream-based emotion recognition model based on bidirectional long short-term memory (Bi-LSTM) and convolutional neural networks (CNNs) using a Korean speech emotion database, and the performance is comparatively analyzed. The data used in the experiment were obtained from the Korean speech emotion recognition database built by Chosun University. Two deep learning models, Bi-LSTM and YAMNet, which is a CNN-based transfer learning model, were connected in a two-stream architecture to design an emotion recognition model. Various speech feature extraction methods and deep learning models were compared in terms of performance. Consequently, the speech emotion recognition performance of Bi-LSTM and YAMNet was 90.38% and 94.91%, respectively. However, the performance of the two-stream model was 96%, which was a minimum of 1.09% and up to 5.62% improved compared with a single model.https://www.mdpi.com/2076-3417/13/4/2167speech emotion recognitionhuman–computer interactiontwo-streambidirectional long-short term memoryconvolutional neural network |
spellingShingle | A-Hyeon Jo Keun-Chang Kwak Speech Emotion Recognition Based on Two-Stream Deep Learning Model Using Korean Audio Information Applied Sciences speech emotion recognition human–computer interaction two-stream bidirectional long-short term memory convolutional neural network |
title | Speech Emotion Recognition Based on Two-Stream Deep Learning Model Using Korean Audio Information |
title_full | Speech Emotion Recognition Based on Two-Stream Deep Learning Model Using Korean Audio Information |
title_fullStr | Speech Emotion Recognition Based on Two-Stream Deep Learning Model Using Korean Audio Information |
title_full_unstemmed | Speech Emotion Recognition Based on Two-Stream Deep Learning Model Using Korean Audio Information |
title_short | Speech Emotion Recognition Based on Two-Stream Deep Learning Model Using Korean Audio Information |
title_sort | speech emotion recognition based on two stream deep learning model using korean audio information |
topic | speech emotion recognition human–computer interaction two-stream bidirectional long-short term memory convolutional neural network |
url | https://www.mdpi.com/2076-3417/13/4/2167 |
work_keys_str_mv | AT ahyeonjo speechemotionrecognitionbasedontwostreamdeeplearningmodelusingkoreanaudioinformation AT keunchangkwak speechemotionrecognitionbasedontwostreamdeeplearningmodelusingkoreanaudioinformation |