Modelling the Trust Value for Human Agents Based on Real-Time Human States in Human-Autonomous Teaming Systems
The modelling of trust values on agents is broadly considered fundamental for decision-making in human-autonomous teaming (HAT) systems. Compared to the evaluation of trust values for robotic agents, estimating human trust is more challenging due to trust miscalibration issues, including undertrust...
Main Authors: | , , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2022-11-01
|
Series: | Technologies |
Subjects: | |
Online Access: | https://www.mdpi.com/2227-7080/10/6/115 |
_version_ | 1827636789781725184 |
---|---|
author | Chin-Teng Lin Hsiu-Yu Fan Yu-Cheng Chang Liang Ou Jia Liu Yu-Kai Wang Tzyy-Ping Jung |
author_facet | Chin-Teng Lin Hsiu-Yu Fan Yu-Cheng Chang Liang Ou Jia Liu Yu-Kai Wang Tzyy-Ping Jung |
author_sort | Chin-Teng Lin |
collection | DOAJ |
description | The modelling of trust values on agents is broadly considered fundamental for decision-making in human-autonomous teaming (HAT) systems. Compared to the evaluation of trust values for robotic agents, estimating human trust is more challenging due to trust miscalibration issues, including undertrust and overtrust problems. From a subjective perception, human trust could be altered along with dynamic human cognitive states, which makes trust values hard to calibrate properly. Thus, in an attempt to capture the dynamics of human trust, the present study evaluated the dynamic nature of trust for human agents through real-time multievidence measures, including human states of attention, stress and perception abilities. The proposed multievidence human trust model applied an adaptive fusion method based on fuzzy reinforcement learning to fuse multievidence from eye trackers, heart rate monitors and human awareness. In addition, fuzzy reinforcement learning was applied to generate rewards via a fuzzy logic inference process that has tolerance for uncertainty in human physiological signals. The results of robot simulation suggest that the proposed trust model can generate reliable human trust values based on real-time cognitive states in the process of ongoing tasks. Moreover, the human-autonomous team with the proposed trust model improved the system efficiency by over <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mn>50</mn><mo>%</mo></mrow></semantics></math></inline-formula> compared to the team with only autonomous agents. These results may demonstrate that the proposed model could provide insight into the real-time adaptation of HAT systems based on human states and, thus, might help develop new ways to enhance future HAT systems better. |
first_indexed | 2024-03-09T15:47:36Z |
format | Article |
id | doaj.art-4e8f693d00c0489083c5df85d6fd1f75 |
institution | Directory Open Access Journal |
issn | 2227-7080 |
language | English |
last_indexed | 2024-03-09T15:47:36Z |
publishDate | 2022-11-01 |
publisher | MDPI AG |
record_format | Article |
series | Technologies |
spelling | doaj.art-4e8f693d00c0489083c5df85d6fd1f752023-11-24T18:22:59ZengMDPI AGTechnologies2227-70802022-11-0110611510.3390/technologies10060115Modelling the Trust Value for Human Agents Based on Real-Time Human States in Human-Autonomous Teaming SystemsChin-Teng Lin0Hsiu-Yu Fan1Yu-Cheng Chang2Liang Ou3Jia Liu4Yu-Kai Wang5Tzyy-Ping Jung6CIBCI Lab, Australian Artificial Intelligence Institute, University of Technology Sydney, Ultimo, NSW 2007, AustraliaInstitute of Imaging and Biomedical Photonics, National Yang Ming Chiao Tung University, Hsinchu City 30010, TaiwanCIBCI Lab, Australian Artificial Intelligence Institute, University of Technology Sydney, Ultimo, NSW 2007, AustraliaCIBCI Lab, Australian Artificial Intelligence Institute, University of Technology Sydney, Ultimo, NSW 2007, AustraliaCIBCI Lab, Australian Artificial Intelligence Institute, University of Technology Sydney, Ultimo, NSW 2007, AustraliaCIBCI Lab, Australian Artificial Intelligence Institute, University of Technology Sydney, Ultimo, NSW 2007, AustraliaInstitute of Engineering in Medicine and Institute for Neural Computation, University of California San Diego, La Jolla, CA 92093, USAThe modelling of trust values on agents is broadly considered fundamental for decision-making in human-autonomous teaming (HAT) systems. Compared to the evaluation of trust values for robotic agents, estimating human trust is more challenging due to trust miscalibration issues, including undertrust and overtrust problems. From a subjective perception, human trust could be altered along with dynamic human cognitive states, which makes trust values hard to calibrate properly. Thus, in an attempt to capture the dynamics of human trust, the present study evaluated the dynamic nature of trust for human agents through real-time multievidence measures, including human states of attention, stress and perception abilities. The proposed multievidence human trust model applied an adaptive fusion method based on fuzzy reinforcement learning to fuse multievidence from eye trackers, heart rate monitors and human awareness. In addition, fuzzy reinforcement learning was applied to generate rewards via a fuzzy logic inference process that has tolerance for uncertainty in human physiological signals. The results of robot simulation suggest that the proposed trust model can generate reliable human trust values based on real-time cognitive states in the process of ongoing tasks. Moreover, the human-autonomous team with the proposed trust model improved the system efficiency by over <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mn>50</mn><mo>%</mo></mrow></semantics></math></inline-formula> compared to the team with only autonomous agents. These results may demonstrate that the proposed model could provide insight into the real-time adaptation of HAT systems based on human states and, thus, might help develop new ways to enhance future HAT systems better.https://www.mdpi.com/2227-7080/10/6/115trust modellinginformation fusionhuman-autonomous teaming |
spellingShingle | Chin-Teng Lin Hsiu-Yu Fan Yu-Cheng Chang Liang Ou Jia Liu Yu-Kai Wang Tzyy-Ping Jung Modelling the Trust Value for Human Agents Based on Real-Time Human States in Human-Autonomous Teaming Systems Technologies trust modelling information fusion human-autonomous teaming |
title | Modelling the Trust Value for Human Agents Based on Real-Time Human States in Human-Autonomous Teaming Systems |
title_full | Modelling the Trust Value for Human Agents Based on Real-Time Human States in Human-Autonomous Teaming Systems |
title_fullStr | Modelling the Trust Value for Human Agents Based on Real-Time Human States in Human-Autonomous Teaming Systems |
title_full_unstemmed | Modelling the Trust Value for Human Agents Based on Real-Time Human States in Human-Autonomous Teaming Systems |
title_short | Modelling the Trust Value for Human Agents Based on Real-Time Human States in Human-Autonomous Teaming Systems |
title_sort | modelling the trust value for human agents based on real time human states in human autonomous teaming systems |
topic | trust modelling information fusion human-autonomous teaming |
url | https://www.mdpi.com/2227-7080/10/6/115 |
work_keys_str_mv | AT chintenglin modellingthetrustvalueforhumanagentsbasedonrealtimehumanstatesinhumanautonomousteamingsystems AT hsiuyufan modellingthetrustvalueforhumanagentsbasedonrealtimehumanstatesinhumanautonomousteamingsystems AT yuchengchang modellingthetrustvalueforhumanagentsbasedonrealtimehumanstatesinhumanautonomousteamingsystems AT liangou modellingthetrustvalueforhumanagentsbasedonrealtimehumanstatesinhumanautonomousteamingsystems AT jialiu modellingthetrustvalueforhumanagentsbasedonrealtimehumanstatesinhumanautonomousteamingsystems AT yukaiwang modellingthetrustvalueforhumanagentsbasedonrealtimehumanstatesinhumanautonomousteamingsystems AT tzyypingjung modellingthetrustvalueforhumanagentsbasedonrealtimehumanstatesinhumanautonomousteamingsystems |