Modelling the Trust Value for Human Agents Based on Real-Time Human States in Human-Autonomous Teaming Systems

The modelling of trust values on agents is broadly considered fundamental for decision-making in human-autonomous teaming (HAT) systems. Compared to the evaluation of trust values for robotic agents, estimating human trust is more challenging due to trust miscalibration issues, including undertrust...

Full description

Bibliographic Details
Main Authors: Chin-Teng Lin, Hsiu-Yu Fan, Yu-Cheng Chang, Liang Ou, Jia Liu, Yu-Kai Wang, Tzyy-Ping Jung
Format: Article
Language:English
Published: MDPI AG 2022-11-01
Series:Technologies
Subjects:
Online Access:https://www.mdpi.com/2227-7080/10/6/115
_version_ 1827636789781725184
author Chin-Teng Lin
Hsiu-Yu Fan
Yu-Cheng Chang
Liang Ou
Jia Liu
Yu-Kai Wang
Tzyy-Ping Jung
author_facet Chin-Teng Lin
Hsiu-Yu Fan
Yu-Cheng Chang
Liang Ou
Jia Liu
Yu-Kai Wang
Tzyy-Ping Jung
author_sort Chin-Teng Lin
collection DOAJ
description The modelling of trust values on agents is broadly considered fundamental for decision-making in human-autonomous teaming (HAT) systems. Compared to the evaluation of trust values for robotic agents, estimating human trust is more challenging due to trust miscalibration issues, including undertrust and overtrust problems. From a subjective perception, human trust could be altered along with dynamic human cognitive states, which makes trust values hard to calibrate properly. Thus, in an attempt to capture the dynamics of human trust, the present study evaluated the dynamic nature of trust for human agents through real-time multievidence measures, including human states of attention, stress and perception abilities. The proposed multievidence human trust model applied an adaptive fusion method based on fuzzy reinforcement learning to fuse multievidence from eye trackers, heart rate monitors and human awareness. In addition, fuzzy reinforcement learning was applied to generate rewards via a fuzzy logic inference process that has tolerance for uncertainty in human physiological signals. The results of robot simulation suggest that the proposed trust model can generate reliable human trust values based on real-time cognitive states in the process of ongoing tasks. Moreover, the human-autonomous team with the proposed trust model improved the system efficiency by over <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mn>50</mn><mo>%</mo></mrow></semantics></math></inline-formula> compared to the team with only autonomous agents. These results may demonstrate that the proposed model could provide insight into the real-time adaptation of HAT systems based on human states and, thus, might help develop new ways to enhance future HAT systems better.
first_indexed 2024-03-09T15:47:36Z
format Article
id doaj.art-4e8f693d00c0489083c5df85d6fd1f75
institution Directory Open Access Journal
issn 2227-7080
language English
last_indexed 2024-03-09T15:47:36Z
publishDate 2022-11-01
publisher MDPI AG
record_format Article
series Technologies
spelling doaj.art-4e8f693d00c0489083c5df85d6fd1f752023-11-24T18:22:59ZengMDPI AGTechnologies2227-70802022-11-0110611510.3390/technologies10060115Modelling the Trust Value for Human Agents Based on Real-Time Human States in Human-Autonomous Teaming SystemsChin-Teng Lin0Hsiu-Yu Fan1Yu-Cheng Chang2Liang Ou3Jia Liu4Yu-Kai Wang5Tzyy-Ping Jung6CIBCI Lab, Australian Artificial Intelligence Institute, University of Technology Sydney, Ultimo, NSW 2007, AustraliaInstitute of Imaging and Biomedical Photonics, National Yang Ming Chiao Tung University, Hsinchu City 30010, TaiwanCIBCI Lab, Australian Artificial Intelligence Institute, University of Technology Sydney, Ultimo, NSW 2007, AustraliaCIBCI Lab, Australian Artificial Intelligence Institute, University of Technology Sydney, Ultimo, NSW 2007, AustraliaCIBCI Lab, Australian Artificial Intelligence Institute, University of Technology Sydney, Ultimo, NSW 2007, AustraliaCIBCI Lab, Australian Artificial Intelligence Institute, University of Technology Sydney, Ultimo, NSW 2007, AustraliaInstitute of Engineering in Medicine and Institute for Neural Computation, University of California San Diego, La Jolla, CA 92093, USAThe modelling of trust values on agents is broadly considered fundamental for decision-making in human-autonomous teaming (HAT) systems. Compared to the evaluation of trust values for robotic agents, estimating human trust is more challenging due to trust miscalibration issues, including undertrust and overtrust problems. From a subjective perception, human trust could be altered along with dynamic human cognitive states, which makes trust values hard to calibrate properly. Thus, in an attempt to capture the dynamics of human trust, the present study evaluated the dynamic nature of trust for human agents through real-time multievidence measures, including human states of attention, stress and perception abilities. The proposed multievidence human trust model applied an adaptive fusion method based on fuzzy reinforcement learning to fuse multievidence from eye trackers, heart rate monitors and human awareness. In addition, fuzzy reinforcement learning was applied to generate rewards via a fuzzy logic inference process that has tolerance for uncertainty in human physiological signals. The results of robot simulation suggest that the proposed trust model can generate reliable human trust values based on real-time cognitive states in the process of ongoing tasks. Moreover, the human-autonomous team with the proposed trust model improved the system efficiency by over <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mn>50</mn><mo>%</mo></mrow></semantics></math></inline-formula> compared to the team with only autonomous agents. These results may demonstrate that the proposed model could provide insight into the real-time adaptation of HAT systems based on human states and, thus, might help develop new ways to enhance future HAT systems better.https://www.mdpi.com/2227-7080/10/6/115trust modellinginformation fusionhuman-autonomous teaming
spellingShingle Chin-Teng Lin
Hsiu-Yu Fan
Yu-Cheng Chang
Liang Ou
Jia Liu
Yu-Kai Wang
Tzyy-Ping Jung
Modelling the Trust Value for Human Agents Based on Real-Time Human States in Human-Autonomous Teaming Systems
Technologies
trust modelling
information fusion
human-autonomous teaming
title Modelling the Trust Value for Human Agents Based on Real-Time Human States in Human-Autonomous Teaming Systems
title_full Modelling the Trust Value for Human Agents Based on Real-Time Human States in Human-Autonomous Teaming Systems
title_fullStr Modelling the Trust Value for Human Agents Based on Real-Time Human States in Human-Autonomous Teaming Systems
title_full_unstemmed Modelling the Trust Value for Human Agents Based on Real-Time Human States in Human-Autonomous Teaming Systems
title_short Modelling the Trust Value for Human Agents Based on Real-Time Human States in Human-Autonomous Teaming Systems
title_sort modelling the trust value for human agents based on real time human states in human autonomous teaming systems
topic trust modelling
information fusion
human-autonomous teaming
url https://www.mdpi.com/2227-7080/10/6/115
work_keys_str_mv AT chintenglin modellingthetrustvalueforhumanagentsbasedonrealtimehumanstatesinhumanautonomousteamingsystems
AT hsiuyufan modellingthetrustvalueforhumanagentsbasedonrealtimehumanstatesinhumanautonomousteamingsystems
AT yuchengchang modellingthetrustvalueforhumanagentsbasedonrealtimehumanstatesinhumanautonomousteamingsystems
AT liangou modellingthetrustvalueforhumanagentsbasedonrealtimehumanstatesinhumanautonomousteamingsystems
AT jialiu modellingthetrustvalueforhumanagentsbasedonrealtimehumanstatesinhumanautonomousteamingsystems
AT yukaiwang modellingthetrustvalueforhumanagentsbasedonrealtimehumanstatesinhumanautonomousteamingsystems
AT tzyypingjung modellingthetrustvalueforhumanagentsbasedonrealtimehumanstatesinhumanautonomousteamingsystems