Multimodal robotic music performance art based on GRU-GoogLeNet model fusing audiovisual perception

The field of multimodal robotic musical performing arts has garnered significant interest due to its innovative potential. Conventional robots face limitations in understanding emotions and artistic expression in musical performances. Therefore, this paper explores the application of multimodal robo...

Full description

Bibliographic Details
Main Author: Lu Wang
Format: Article
Language:English
Published: Frontiers Media S.A. 2024-01-01
Series:Frontiers in Neurorobotics
Subjects:
Online Access:https://www.frontiersin.org/articles/10.3389/fnbot.2023.1324831/full
_version_ 1827369202073206784
author Lu Wang
author_facet Lu Wang
author_sort Lu Wang
collection DOAJ
description The field of multimodal robotic musical performing arts has garnered significant interest due to its innovative potential. Conventional robots face limitations in understanding emotions and artistic expression in musical performances. Therefore, this paper explores the application of multimodal robots that integrate visual and auditory perception to enhance the quality and artistic expression in music performance. Our approach involves integrating GRU (Gated Recurrent Unit) and GoogLeNet models for sentiment analysis. The GRU model processes audio data and captures the temporal dynamics of musical elements, including long-term dependencies, to extract emotional information. The GoogLeNet model excels in image processing, extracting complex visual details and aesthetic features. This synergy deepens the understanding of musical and visual elements, aiming to produce more emotionally resonant and interactive robot performances. Experimental results demonstrate the effectiveness of our approach, showing significant improvements in music performance by multimodal robots. These robots, equipped with our method, deliver high-quality, artistic performances that effectively evoke emotional engagement from the audience. Multimodal robots that merge audio-visual perception in music performance enrich the art form and offer diverse human-machine interactions. This research demonstrates the potential of multimodal robots in music performance, promoting the integration of technology and art. It opens new realms in performing arts and human-robot interactions, offering a unique and innovative experience. Our findings provide valuable insights for the development of multimodal robots in the performing arts sector.
first_indexed 2024-03-08T09:41:49Z
format Article
id doaj.art-78c7401b567a4c62874ee80dd031bae9
institution Directory Open Access Journal
issn 1662-5218
language English
last_indexed 2024-03-08T09:41:49Z
publishDate 2024-01-01
publisher Frontiers Media S.A.
record_format Article
series Frontiers in Neurorobotics
spelling doaj.art-78c7401b567a4c62874ee80dd031bae92024-01-30T04:15:24ZengFrontiers Media S.A.Frontiers in Neurorobotics1662-52182024-01-011710.3389/fnbot.2023.13248311324831Multimodal robotic music performance art based on GRU-GoogLeNet model fusing audiovisual perceptionLu WangThe field of multimodal robotic musical performing arts has garnered significant interest due to its innovative potential. Conventional robots face limitations in understanding emotions and artistic expression in musical performances. Therefore, this paper explores the application of multimodal robots that integrate visual and auditory perception to enhance the quality and artistic expression in music performance. Our approach involves integrating GRU (Gated Recurrent Unit) and GoogLeNet models for sentiment analysis. The GRU model processes audio data and captures the temporal dynamics of musical elements, including long-term dependencies, to extract emotional information. The GoogLeNet model excels in image processing, extracting complex visual details and aesthetic features. This synergy deepens the understanding of musical and visual elements, aiming to produce more emotionally resonant and interactive robot performances. Experimental results demonstrate the effectiveness of our approach, showing significant improvements in music performance by multimodal robots. These robots, equipped with our method, deliver high-quality, artistic performances that effectively evoke emotional engagement from the audience. Multimodal robots that merge audio-visual perception in music performance enrich the art form and offer diverse human-machine interactions. This research demonstrates the potential of multimodal robots in music performance, promoting the integration of technology and art. It opens new realms in performing arts and human-robot interactions, offering a unique and innovative experience. Our findings provide valuable insights for the development of multimodal robots in the performing arts sector.https://www.frontiersin.org/articles/10.3389/fnbot.2023.1324831/fullemotionmusicmultimodal robotGRUGoogLeNet
spellingShingle Lu Wang
Multimodal robotic music performance art based on GRU-GoogLeNet model fusing audiovisual perception
Frontiers in Neurorobotics
emotion
music
multimodal robot
GRU
GoogLeNet
title Multimodal robotic music performance art based on GRU-GoogLeNet model fusing audiovisual perception
title_full Multimodal robotic music performance art based on GRU-GoogLeNet model fusing audiovisual perception
title_fullStr Multimodal robotic music performance art based on GRU-GoogLeNet model fusing audiovisual perception
title_full_unstemmed Multimodal robotic music performance art based on GRU-GoogLeNet model fusing audiovisual perception
title_short Multimodal robotic music performance art based on GRU-GoogLeNet model fusing audiovisual perception
title_sort multimodal robotic music performance art based on gru googlenet model fusing audiovisual perception
topic emotion
music
multimodal robot
GRU
GoogLeNet
url https://www.frontiersin.org/articles/10.3389/fnbot.2023.1324831/full
work_keys_str_mv AT luwang multimodalroboticmusicperformanceartbasedongrugooglenetmodelfusingaudiovisualperception