Multimodal sentiment analysis using hierarchical fusion with context modeling

Multimodal sentiment analysis is a very actively growing field of research. A promising area of opportunity in this field is to improve the multimodal fusion mechanism. We present a novel feature fusion strategy that proceeds in a hierarchical fashion, first fusing the modalities two in two and only...

Full description

Bibliographic Details
Main Authors: Majumder, Navonil, Hazarika, Devamanyu, Gelbukh, Alexander, Cambria, Erik, Poria, Soujanya
Other Authors: School of Computer Science and Engineering
Format: Journal Article
Language:English
Published: 2020
Subjects:
Online Access:https://hdl.handle.net/10356/139583
_version_ 1811696665950683136
author Majumder, Navonil
Hazarika, Devamanyu
Gelbukh, Alexander
Cambria, Erik
Poria, Soujanya
author2 School of Computer Science and Engineering
author_facet School of Computer Science and Engineering
Majumder, Navonil
Hazarika, Devamanyu
Gelbukh, Alexander
Cambria, Erik
Poria, Soujanya
author_sort Majumder, Navonil
collection NTU
description Multimodal sentiment analysis is a very actively growing field of research. A promising area of opportunity in this field is to improve the multimodal fusion mechanism. We present a novel feature fusion strategy that proceeds in a hierarchical fashion, first fusing the modalities two in two and only then fusing all three modalities. On multimodal sentiment analysis of individual utterances, our strategy outperforms conventional concatenation of features by 1%, which amounts to 5% reduction in error rate. On utterance-level multimodal sentiment analysis of multi-utterance video clips, for which current state-of-the-art techniques incorporate contextual information from other utterances of the same clip, our hierarchical fusion gives up to 2.4% (almost 10% error rate reduction) over currently used concatenation. The implementation of our method is publicly available in the form of open-source code.
first_indexed 2024-10-01T07:42:59Z
format Journal Article
id ntu-10356/139583
institution Nanyang Technological University
language English
last_indexed 2024-10-01T07:42:59Z
publishDate 2020
record_format dspace
spelling ntu-10356/1395832020-05-20T07:01:54Z Multimodal sentiment analysis using hierarchical fusion with context modeling Majumder, Navonil Hazarika, Devamanyu Gelbukh, Alexander Cambria, Erik Poria, Soujanya School of Computer Science and Engineering Engineering::Computer science and engineering Multimodal Fusion Sentiment Analysis Multimodal sentiment analysis is a very actively growing field of research. A promising area of opportunity in this field is to improve the multimodal fusion mechanism. We present a novel feature fusion strategy that proceeds in a hierarchical fashion, first fusing the modalities two in two and only then fusing all three modalities. On multimodal sentiment analysis of individual utterances, our strategy outperforms conventional concatenation of features by 1%, which amounts to 5% reduction in error rate. On utterance-level multimodal sentiment analysis of multi-utterance video clips, for which current state-of-the-art techniques incorporate contextual information from other utterances of the same clip, our hierarchical fusion gives up to 2.4% (almost 10% error rate reduction) over currently used concatenation. The implementation of our method is publicly available in the form of open-source code. 2020-05-20T07:01:54Z 2020-05-20T07:01:54Z 2018 Journal Article Majumder, N., Hazarika, D., Gelbukh, A., Cambria, E., & Poria, S. (2018). Multimodal sentiment analysis using hierarchical fusion with context modeling. Knowledge-Based Systems, 161, 124-133. doi:10.1016/j.knosys.2018.07.041 0950-7051 https://hdl.handle.net/10356/139583 10.1016/j.knosys.2018.07.041 2-s2.0-85050999093 161 124 133 en Knowledge-Based Systems © 2018 Elsevier B.V. All rights reserved.
spellingShingle Engineering::Computer science and engineering
Multimodal Fusion
Sentiment Analysis
Majumder, Navonil
Hazarika, Devamanyu
Gelbukh, Alexander
Cambria, Erik
Poria, Soujanya
Multimodal sentiment analysis using hierarchical fusion with context modeling
title Multimodal sentiment analysis using hierarchical fusion with context modeling
title_full Multimodal sentiment analysis using hierarchical fusion with context modeling
title_fullStr Multimodal sentiment analysis using hierarchical fusion with context modeling
title_full_unstemmed Multimodal sentiment analysis using hierarchical fusion with context modeling
title_short Multimodal sentiment analysis using hierarchical fusion with context modeling
title_sort multimodal sentiment analysis using hierarchical fusion with context modeling
topic Engineering::Computer science and engineering
Multimodal Fusion
Sentiment Analysis
url https://hdl.handle.net/10356/139583
work_keys_str_mv AT majumdernavonil multimodalsentimentanalysisusinghierarchicalfusionwithcontextmodeling
AT hazarikadevamanyu multimodalsentimentanalysisusinghierarchicalfusionwithcontextmodeling
AT gelbukhalexander multimodalsentimentanalysisusinghierarchicalfusionwithcontextmodeling
AT cambriaerik multimodalsentimentanalysisusinghierarchicalfusionwithcontextmodeling
AT poriasoujanya multimodalsentimentanalysisusinghierarchicalfusionwithcontextmodeling