Deep emotion recognition based on audio–visual correlation
Human emotion recognition is studied by means of unimodal channels over the last decade. However, efforts continue to answer tempting questions about how variant modalities can complement each other. This study proposes a multimodal approach using three‐dimensional (3D) convolutional neural networks...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
Wiley
2020-10-01
|
Series: | IET Computer Vision |
Subjects: | |
Online Access: | https://doi.org/10.1049/iet-cvi.2020.0013 |
_version_ | 1797684639771394048 |
---|---|
author | Noushin Hajarolasvadi Hasan Demirel |
author_facet | Noushin Hajarolasvadi Hasan Demirel |
author_sort | Noushin Hajarolasvadi |
collection | DOAJ |
description | Human emotion recognition is studied by means of unimodal channels over the last decade. However, efforts continue to answer tempting questions about how variant modalities can complement each other. This study proposes a multimodal approach using three‐dimensional (3D) convolutional neural networks (CNNs) to model human emotion through a modality‐referenced system while investigating the solution to such questions. The proposed modality‐referenced system selects the input data based on one of the modalities regarded as reference or master. The other modality which is referred to as a slave simply adjusts or attunes itself with the master in the temporal domain. In this context, the authors developed three multimodal emotion recognition system, namely, video‐referenced system, audio‐referenced system, and the audio–visual‐referenced system to explore the congruence impact of audio and video modalities on each other. Two pipelines of 3D CNN architectures are employed where k‐means clustering is used in the master pipeline and the slave pipeline adapts itself in a temporal sense. The outputs of the two pipelines are fused to improve recognition performance. In addition, canonical correlation analysis and t‐distributed stochastic neighbour embedding is used validating the experiments. Results show that temporal alignment of the data between two modalities improves the recognition performance significantly. |
first_indexed | 2024-03-12T00:32:38Z |
format | Article |
id | doaj.art-ded5ce8b1c97472d93b103c80476b275 |
institution | Directory Open Access Journal |
issn | 1751-9632 1751-9640 |
language | English |
last_indexed | 2024-03-12T00:32:38Z |
publishDate | 2020-10-01 |
publisher | Wiley |
record_format | Article |
series | IET Computer Vision |
spelling | doaj.art-ded5ce8b1c97472d93b103c80476b2752023-09-15T10:11:27ZengWileyIET Computer Vision1751-96321751-96402020-10-0114751752710.1049/iet-cvi.2020.0013Deep emotion recognition based on audio–visual correlationNoushin Hajarolasvadi0Hasan Demirel1Department of Electrical and Electronic EngineeringEastern Mediterranean UniversityTurkey, 10 via MersinNicosia99628CyprusDepartment of Electrical and Electronic EngineeringEastern Mediterranean UniversityTurkey, 10 via MersinNicosia99628CyprusHuman emotion recognition is studied by means of unimodal channels over the last decade. However, efforts continue to answer tempting questions about how variant modalities can complement each other. This study proposes a multimodal approach using three‐dimensional (3D) convolutional neural networks (CNNs) to model human emotion through a modality‐referenced system while investigating the solution to such questions. The proposed modality‐referenced system selects the input data based on one of the modalities regarded as reference or master. The other modality which is referred to as a slave simply adjusts or attunes itself with the master in the temporal domain. In this context, the authors developed three multimodal emotion recognition system, namely, video‐referenced system, audio‐referenced system, and the audio–visual‐referenced system to explore the congruence impact of audio and video modalities on each other. Two pipelines of 3D CNN architectures are employed where k‐means clustering is used in the master pipeline and the slave pipeline adapts itself in a temporal sense. The outputs of the two pipelines are fused to improve recognition performance. In addition, canonical correlation analysis and t‐distributed stochastic neighbour embedding is used validating the experiments. Results show that temporal alignment of the data between two modalities improves the recognition performance significantly.https://doi.org/10.1049/iet-cvi.2020.0013temporal data alignmentslave pipelinek‐means clustering3D CNN architecturestemporal domainunimodal channels |
spellingShingle | Noushin Hajarolasvadi Hasan Demirel Deep emotion recognition based on audio–visual correlation IET Computer Vision temporal data alignment slave pipeline k‐means clustering 3D CNN architectures temporal domain unimodal channels |
title | Deep emotion recognition based on audio–visual correlation |
title_full | Deep emotion recognition based on audio–visual correlation |
title_fullStr | Deep emotion recognition based on audio–visual correlation |
title_full_unstemmed | Deep emotion recognition based on audio–visual correlation |
title_short | Deep emotion recognition based on audio–visual correlation |
title_sort | deep emotion recognition based on audio visual correlation |
topic | temporal data alignment slave pipeline k‐means clustering 3D CNN architectures temporal domain unimodal channels |
url | https://doi.org/10.1049/iet-cvi.2020.0013 |
work_keys_str_mv | AT noushinhajarolasvadi deepemotionrecognitionbasedonaudiovisualcorrelation AT hasandemirel deepemotionrecognitionbasedonaudiovisualcorrelation |