Latent Structure Matching for Knowledge Transfer in Reinforcement Learning
Reinforcement learning algorithms usually require a large number of empirical samples and give rise to a slow convergence in practical applications. One solution is to introduce transfer learning: Knowledge from well-learned source tasks can be reused to reduce sample request and accelerate the lear...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2020-02-01
|
Series: | Future Internet |
Subjects: | |
Online Access: | https://www.mdpi.com/1999-5903/12/2/36 |
_version_ | 1828287904390053888 |
---|---|
author | Yi Zhou Fenglei Yang |
author_facet | Yi Zhou Fenglei Yang |
author_sort | Yi Zhou |
collection | DOAJ |
description | Reinforcement learning algorithms usually require a large number of empirical samples and give rise to a slow convergence in practical applications. One solution is to introduce transfer learning: Knowledge from well-learned source tasks can be reused to reduce sample request and accelerate the learning of target tasks. However, if an unmatched source task is selected, it will slow down or even disrupt the learning procedure. Therefore, it is very important for knowledge transfer to select appropriate source tasks that have a high degree of matching with target tasks. In this paper, a novel task matching algorithm is proposed to derive the latent structures of value functions of tasks, and align the structures for similarity estimation. Through the latent structure matching, the highly-matched source tasks are selected effectively, from which knowledge is then transferred to give action advice, and improve exploration strategies of the target tasks. Experiments are conducted on the simulated navigation environment and the mountain car environment. The results illustrate the significant performance gain of the improved exploration strategy, compared with traditional <inline-formula> <math display="inline"> <semantics> <mi>ϵ</mi> </semantics> </math> </inline-formula>-greedy exploration strategy. A theoretical proof is also given to verify the improvement of the exploration strategy based on latent structure matching. |
first_indexed | 2024-04-13T09:55:56Z |
format | Article |
id | doaj.art-be5c51b910654e0fa3603f81e237810f |
institution | Directory Open Access Journal |
issn | 1999-5903 |
language | English |
last_indexed | 2024-04-13T09:55:56Z |
publishDate | 2020-02-01 |
publisher | MDPI AG |
record_format | Article |
series | Future Internet |
spelling | doaj.art-be5c51b910654e0fa3603f81e237810f2022-12-22T02:51:25ZengMDPI AGFuture Internet1999-59032020-02-011223610.3390/fi12020036fi12020036Latent Structure Matching for Knowledge Transfer in Reinforcement LearningYi Zhou0Fenglei Yang1School of Computer Engineering and Science, Shanghai University, Shanghai 200444, ChinaSchool of Computer Engineering and Science, Shanghai University, Shanghai 200444, ChinaReinforcement learning algorithms usually require a large number of empirical samples and give rise to a slow convergence in practical applications. One solution is to introduce transfer learning: Knowledge from well-learned source tasks can be reused to reduce sample request and accelerate the learning of target tasks. However, if an unmatched source task is selected, it will slow down or even disrupt the learning procedure. Therefore, it is very important for knowledge transfer to select appropriate source tasks that have a high degree of matching with target tasks. In this paper, a novel task matching algorithm is proposed to derive the latent structures of value functions of tasks, and align the structures for similarity estimation. Through the latent structure matching, the highly-matched source tasks are selected effectively, from which knowledge is then transferred to give action advice, and improve exploration strategies of the target tasks. Experiments are conducted on the simulated navigation environment and the mountain car environment. The results illustrate the significant performance gain of the improved exploration strategy, compared with traditional <inline-formula> <math display="inline"> <semantics> <mi>ϵ</mi> </semantics> </math> </inline-formula>-greedy exploration strategy. A theoretical proof is also given to verify the improvement of the exploration strategy based on latent structure matching.https://www.mdpi.com/1999-5903/12/2/36latent structure matchingreinforcement learningtransfer learningaction advicepolicy improvementmountain car |
spellingShingle | Yi Zhou Fenglei Yang Latent Structure Matching for Knowledge Transfer in Reinforcement Learning Future Internet latent structure matching reinforcement learning transfer learning action advice policy improvement mountain car |
title | Latent Structure Matching for Knowledge Transfer in Reinforcement Learning |
title_full | Latent Structure Matching for Knowledge Transfer in Reinforcement Learning |
title_fullStr | Latent Structure Matching for Knowledge Transfer in Reinforcement Learning |
title_full_unstemmed | Latent Structure Matching for Knowledge Transfer in Reinforcement Learning |
title_short | Latent Structure Matching for Knowledge Transfer in Reinforcement Learning |
title_sort | latent structure matching for knowledge transfer in reinforcement learning |
topic | latent structure matching reinforcement learning transfer learning action advice policy improvement mountain car |
url | https://www.mdpi.com/1999-5903/12/2/36 |
work_keys_str_mv | AT yizhou latentstructurematchingforknowledgetransferinreinforcementlearning AT fengleiyang latentstructurematchingforknowledgetransferinreinforcementlearning |