Multimodal fusion of EMG and vision for human grasp intent inference in prosthetic hand control

Objective: For transradial amputees, robotic prosthetic hands promise to regain the capability to perform daily living activities. Current control methods based on physiological signals such as electromyography (EMG) are prone to yielding poor inference outcomes due to motion artifacts, muscle fatig...

Full description

Bibliographic Details
Main Authors: Mehrshad Zandigohar, Mo Han, Mohammadreza Sharif, Sezen Yağmur Günay, Mariusz P. Furmanek, Mathew Yarossi, Paolo Bonato, Cagdas Onal, Taşkın Padır, Deniz Erdoğmuş, Gunar Schirner
Format: Article
Language:English
Published: Frontiers Media S.A. 2024-02-01
Series:Frontiers in Robotics and AI
Subjects:
Online Access:https://www.frontiersin.org/articles/10.3389/frobt.2024.1312554/full
_version_ 1797294171044708352
author Mehrshad Zandigohar
Mo Han
Mohammadreza Sharif
Sezen Yağmur Günay
Mariusz P. Furmanek
Mariusz P. Furmanek
Mathew Yarossi
Paolo Bonato
Cagdas Onal
Taşkın Padır
Deniz Erdoğmuş
Gunar Schirner
author_facet Mehrshad Zandigohar
Mo Han
Mohammadreza Sharif
Sezen Yağmur Günay
Mariusz P. Furmanek
Mariusz P. Furmanek
Mathew Yarossi
Paolo Bonato
Cagdas Onal
Taşkın Padır
Deniz Erdoğmuş
Gunar Schirner
author_sort Mehrshad Zandigohar
collection DOAJ
description Objective: For transradial amputees, robotic prosthetic hands promise to regain the capability to perform daily living activities. Current control methods based on physiological signals such as electromyography (EMG) are prone to yielding poor inference outcomes due to motion artifacts, muscle fatigue, and many more. Vision sensors are a major source of information about the environment state and can play a vital role in inferring feasible and intended gestures. However, visual evidence is also susceptible to its own artifacts, most often due to object occlusion, lighting changes, etc. Multimodal evidence fusion using physiological and vision sensor measurements is a natural approach due to the complementary strengths of these modalities.Methods: In this paper, we present a Bayesian evidence fusion framework for grasp intent inference using eye-view video, eye-gaze, and EMG from the forearm processed by neural network models. We analyze individual and fused performance as a function of time as the hand approaches the object to grasp it. For this purpose, we have also developed novel data processing and augmentation techniques to train neural network components.Results: Our results indicate that, on average, fusion improves the instantaneous upcoming grasp type classification accuracy while in the reaching phase by 13.66% and 14.8%, relative to EMG (81.64% non-fused) and visual evidence (80.5% non-fused) individually, resulting in an overall fusion accuracy of 95.3%.Conclusion: Our experimental data analyses demonstrate that EMG and visual evidence show complementary strengths, and as a consequence, fusion of multimodal evidence can outperform each individual evidence modality at any given time.
first_indexed 2024-03-07T21:26:22Z
format Article
id doaj.art-f1c3baaf51ee46a7a73971adf623eff3
institution Directory Open Access Journal
issn 2296-9144
language English
last_indexed 2024-03-07T21:26:22Z
publishDate 2024-02-01
publisher Frontiers Media S.A.
record_format Article
series Frontiers in Robotics and AI
spelling doaj.art-f1c3baaf51ee46a7a73971adf623eff32024-02-27T04:39:12ZengFrontiers Media S.A.Frontiers in Robotics and AI2296-91442024-02-011110.3389/frobt.2024.13125541312554Multimodal fusion of EMG and vision for human grasp intent inference in prosthetic hand controlMehrshad Zandigohar0Mo Han1Mohammadreza Sharif2Sezen Yağmur Günay3Mariusz P. Furmanek4Mariusz P. Furmanek5Mathew Yarossi6Paolo Bonato7Cagdas Onal8Taşkın Padır9Deniz Erdoğmuş10Gunar Schirner11Department of Electrical and Computer Engineering, Northeastern University, Boston, MA, United StatesDepartment of Electrical and Computer Engineering, Northeastern University, Boston, MA, United StatesDepartment of Electrical and Computer Engineering, Northeastern University, Boston, MA, United StatesDepartment of Electrical and Computer Engineering, Northeastern University, Boston, MA, United StatesDepartment of Physical Therapy, Movement and Rehabilitation Sciences, Northeastern University, Boston, MA, United StatesInstitute of Sport Sciences, Academy of Physical Education in Katowice, Katowice, PolandDepartment of Physical Therapy, Movement and Rehabilitation Sciences, Northeastern University, Boston, MA, United StatesMotion Analysis Lab, Spaulding Rehabilitation Hospital, Charlestown, MA, United StatesSoft Robotics Lab, Worcester Polytechnic Institute, Worcester, MA, United StatesDepartment of Electrical and Computer Engineering, Northeastern University, Boston, MA, United StatesDepartment of Electrical and Computer Engineering, Northeastern University, Boston, MA, United StatesDepartment of Electrical and Computer Engineering, Northeastern University, Boston, MA, United StatesObjective: For transradial amputees, robotic prosthetic hands promise to regain the capability to perform daily living activities. Current control methods based on physiological signals such as electromyography (EMG) are prone to yielding poor inference outcomes due to motion artifacts, muscle fatigue, and many more. Vision sensors are a major source of information about the environment state and can play a vital role in inferring feasible and intended gestures. However, visual evidence is also susceptible to its own artifacts, most often due to object occlusion, lighting changes, etc. Multimodal evidence fusion using physiological and vision sensor measurements is a natural approach due to the complementary strengths of these modalities.Methods: In this paper, we present a Bayesian evidence fusion framework for grasp intent inference using eye-view video, eye-gaze, and EMG from the forearm processed by neural network models. We analyze individual and fused performance as a function of time as the hand approaches the object to grasp it. For this purpose, we have also developed novel data processing and augmentation techniques to train neural network components.Results: Our results indicate that, on average, fusion improves the instantaneous upcoming grasp type classification accuracy while in the reaching phase by 13.66% and 14.8%, relative to EMG (81.64% non-fused) and visual evidence (80.5% non-fused) individually, resulting in an overall fusion accuracy of 95.3%.Conclusion: Our experimental data analyses demonstrate that EMG and visual evidence show complementary strengths, and as a consequence, fusion of multimodal evidence can outperform each individual evidence modality at any given time.https://www.frontiersin.org/articles/10.3389/frobt.2024.1312554/fulldatasetEMGgrasp detectionneural networksrobotic prosthetic hand
spellingShingle Mehrshad Zandigohar
Mo Han
Mohammadreza Sharif
Sezen Yağmur Günay
Mariusz P. Furmanek
Mariusz P. Furmanek
Mathew Yarossi
Paolo Bonato
Cagdas Onal
Taşkın Padır
Deniz Erdoğmuş
Gunar Schirner
Multimodal fusion of EMG and vision for human grasp intent inference in prosthetic hand control
Frontiers in Robotics and AI
dataset
EMG
grasp detection
neural networks
robotic prosthetic hand
title Multimodal fusion of EMG and vision for human grasp intent inference in prosthetic hand control
title_full Multimodal fusion of EMG and vision for human grasp intent inference in prosthetic hand control
title_fullStr Multimodal fusion of EMG and vision for human grasp intent inference in prosthetic hand control
title_full_unstemmed Multimodal fusion of EMG and vision for human grasp intent inference in prosthetic hand control
title_short Multimodal fusion of EMG and vision for human grasp intent inference in prosthetic hand control
title_sort multimodal fusion of emg and vision for human grasp intent inference in prosthetic hand control
topic dataset
EMG
grasp detection
neural networks
robotic prosthetic hand
url https://www.frontiersin.org/articles/10.3389/frobt.2024.1312554/full
work_keys_str_mv AT mehrshadzandigohar multimodalfusionofemgandvisionforhumangraspintentinferenceinprosthetichandcontrol
AT mohan multimodalfusionofemgandvisionforhumangraspintentinferenceinprosthetichandcontrol
AT mohammadrezasharif multimodalfusionofemgandvisionforhumangraspintentinferenceinprosthetichandcontrol
AT sezenyagmurgunay multimodalfusionofemgandvisionforhumangraspintentinferenceinprosthetichandcontrol
AT mariuszpfurmanek multimodalfusionofemgandvisionforhumangraspintentinferenceinprosthetichandcontrol
AT mariuszpfurmanek multimodalfusionofemgandvisionforhumangraspintentinferenceinprosthetichandcontrol
AT mathewyarossi multimodalfusionofemgandvisionforhumangraspintentinferenceinprosthetichandcontrol
AT paolobonato multimodalfusionofemgandvisionforhumangraspintentinferenceinprosthetichandcontrol
AT cagdasonal multimodalfusionofemgandvisionforhumangraspintentinferenceinprosthetichandcontrol
AT taskınpadır multimodalfusionofemgandvisionforhumangraspintentinferenceinprosthetichandcontrol
AT denizerdogmus multimodalfusionofemgandvisionforhumangraspintentinferenceinprosthetichandcontrol
AT gunarschirner multimodalfusionofemgandvisionforhumangraspintentinferenceinprosthetichandcontrol