Adaptive Human–Machine Evaluation Framework Using Stochastic Gradient Descent-Based Reinforcement Learning for Dynamic Competing Network

Complex problems require considerable work, extensive computation, and the development of effective solution methods. Recently, physical hardware- and software-based technologies have been utilized to support problem solving with computers. However, problem solving often involves human expertise and...

Full description

Bibliographic Details
Main Authors: Jinbae Kim, Hyunsoo Lee
Format: Article
Language:English
Published: MDPI AG 2020-04-01
Series:Applied Sciences
Subjects:
Online Access:https://www.mdpi.com/2076-3417/10/7/2558
_version_ 1797571166508941312
author Jinbae Kim
Hyunsoo Lee
author_facet Jinbae Kim
Hyunsoo Lee
author_sort Jinbae Kim
collection DOAJ
description Complex problems require considerable work, extensive computation, and the development of effective solution methods. Recently, physical hardware- and software-based technologies have been utilized to support problem solving with computers. However, problem solving often involves human expertise and guidance. In these cases, accurate human evaluations and diagnoses must be communicated to the system, which should be done using a series of real numbers. In previous studies, only binary numbers have been used for this purpose. Hence, to achieve this objective, this paper proposes a new method of learning complex network topologies that coexist and compete in the same environment and interfere with the learning objectives of the others. Considering the special problem of reinforcement learning in an environment in which multiple network topologies coexist, we propose a policy that properly computes and updates the rewards derived from quantitative human evaluation and computes together with the rewards of the system. The rewards derived from the quantitative human evaluation are designed to be updated quickly and easily in an adaptive manner. Our new framework was applied to a basketball game for validation and demonstrated greater effectiveness than the existing methods.
first_indexed 2024-03-10T20:35:27Z
format Article
id doaj.art-05a18986b9c84d47a29e800dd40d5ade
institution Directory Open Access Journal
issn 2076-3417
language English
last_indexed 2024-03-10T20:35:27Z
publishDate 2020-04-01
publisher MDPI AG
record_format Article
series Applied Sciences
spelling doaj.art-05a18986b9c84d47a29e800dd40d5ade2023-11-19T21:00:42ZengMDPI AGApplied Sciences2076-34172020-04-01107255810.3390/app10072558Adaptive Human–Machine Evaluation Framework Using Stochastic Gradient Descent-Based Reinforcement Learning for Dynamic Competing NetworkJinbae Kim0Hyunsoo Lee1School of Industrial Engineering, Kumoh National Institute of Technology, Gumi 39177, KoreaSchool of Industrial Engineering, Kumoh National Institute of Technology, Gumi 39177, KoreaComplex problems require considerable work, extensive computation, and the development of effective solution methods. Recently, physical hardware- and software-based technologies have been utilized to support problem solving with computers. However, problem solving often involves human expertise and guidance. In these cases, accurate human evaluations and diagnoses must be communicated to the system, which should be done using a series of real numbers. In previous studies, only binary numbers have been used for this purpose. Hence, to achieve this objective, this paper proposes a new method of learning complex network topologies that coexist and compete in the same environment and interfere with the learning objectives of the others. Considering the special problem of reinforcement learning in an environment in which multiple network topologies coexist, we propose a policy that properly computes and updates the rewards derived from quantitative human evaluation and computes together with the rewards of the system. The rewards derived from the quantitative human evaluation are designed to be updated quickly and easily in an adaptive manner. Our new framework was applied to a basketball game for validation and demonstrated greater effectiveness than the existing methods.https://www.mdpi.com/2076-3417/10/7/2558adaptive human evaluationdynamic competing networkreinforcement learningstochastic gradient descent
spellingShingle Jinbae Kim
Hyunsoo Lee
Adaptive Human–Machine Evaluation Framework Using Stochastic Gradient Descent-Based Reinforcement Learning for Dynamic Competing Network
Applied Sciences
adaptive human evaluation
dynamic competing network
reinforcement learning
stochastic gradient descent
title Adaptive Human–Machine Evaluation Framework Using Stochastic Gradient Descent-Based Reinforcement Learning for Dynamic Competing Network
title_full Adaptive Human–Machine Evaluation Framework Using Stochastic Gradient Descent-Based Reinforcement Learning for Dynamic Competing Network
title_fullStr Adaptive Human–Machine Evaluation Framework Using Stochastic Gradient Descent-Based Reinforcement Learning for Dynamic Competing Network
title_full_unstemmed Adaptive Human–Machine Evaluation Framework Using Stochastic Gradient Descent-Based Reinforcement Learning for Dynamic Competing Network
title_short Adaptive Human–Machine Evaluation Framework Using Stochastic Gradient Descent-Based Reinforcement Learning for Dynamic Competing Network
title_sort adaptive human machine evaluation framework using stochastic gradient descent based reinforcement learning for dynamic competing network
topic adaptive human evaluation
dynamic competing network
reinforcement learning
stochastic gradient descent
url https://www.mdpi.com/2076-3417/10/7/2558
work_keys_str_mv AT jinbaekim adaptivehumanmachineevaluationframeworkusingstochasticgradientdescentbasedreinforcementlearningfordynamiccompetingnetwork
AT hyunsoolee adaptivehumanmachineevaluationframeworkusingstochasticgradientdescentbasedreinforcementlearningfordynamiccompetingnetwork