Effective Learning in Non-Stationary Multiagent Environments

Multiagent reinforcement learning (MARL) provides a principled framework for a group of artificial intelligence agents to learn collaborative and/or competitive behaviors at the level of human experts. Multiagent learning settings inherently solve much more complex problems than single-agent learnin...

Full description

Bibliographic Details
Main Author: Kim, Dong Ki
Other Authors: How, Jonathan P.
Format: Thesis
Published: Massachusetts Institute of Technology 2023
Online Access:https://hdl.handle.net/1721.1/150177
_version_ 1826215896653234176
author Kim, Dong Ki
author2 How, Jonathan P.
author_facet How, Jonathan P.
Kim, Dong Ki
author_sort Kim, Dong Ki
collection MIT
description Multiagent reinforcement learning (MARL) provides a principled framework for a group of artificial intelligence agents to learn collaborative and/or competitive behaviors at the level of human experts. Multiagent learning settings inherently solve much more complex problems than single-agent learning because an agent interacts both with the environment and other agents. In particular, multiple agents simultaneously learn in MARL, leading to natural non-stationarity in the experiences encountered and thus requiring each agent to its behavior with respect to potentially large changes in other agents' policies. This thesis aims to address the non-stationarity challenge in multiagent learning from three important topics: 1) adaptation, 2) convergence, and 3) state space. The first topic answers how an agent can learn effective adaptation strategies concerning other agents' changing policies by developing a new meta-learning framework. The second topic answers how agents can adapt and influence the joint learning process such that policies converge to more desirable limiting behaviors by the end of learning based on a new game-theoretical solution concept. Lastly, the last topic answers how state space size can be reduced based on knowledge sharing and context-specific abstraction such that the learning complexity is less affected by non-stationarity. In summary, this thesis develops theoretical and algorithmic contributions to provide principled answers to the aforementioned topics on non-stationarity. The developed algorithms in this thesis demonstrate their effectiveness in a diverse suite of multiagent benchmark domains, including the full spectrum of mixed incentive, competitive, and cooperative environments.
first_indexed 2024-09-23T16:38:46Z
format Thesis
id mit-1721.1/150177
institution Massachusetts Institute of Technology
last_indexed 2024-09-23T16:38:46Z
publishDate 2023
publisher Massachusetts Institute of Technology
record_format dspace
spelling mit-1721.1/1501772023-04-01T03:03:34Z Effective Learning in Non-Stationary Multiagent Environments Kim, Dong Ki How, Jonathan P. Agrawal, Pulkit Foerster, Jakob N. Massachusetts Institute of Technology. Department of Aeronautics and Astronautics Multiagent reinforcement learning (MARL) provides a principled framework for a group of artificial intelligence agents to learn collaborative and/or competitive behaviors at the level of human experts. Multiagent learning settings inherently solve much more complex problems than single-agent learning because an agent interacts both with the environment and other agents. In particular, multiple agents simultaneously learn in MARL, leading to natural non-stationarity in the experiences encountered and thus requiring each agent to its behavior with respect to potentially large changes in other agents' policies. This thesis aims to address the non-stationarity challenge in multiagent learning from three important topics: 1) adaptation, 2) convergence, and 3) state space. The first topic answers how an agent can learn effective adaptation strategies concerning other agents' changing policies by developing a new meta-learning framework. The second topic answers how agents can adapt and influence the joint learning process such that policies converge to more desirable limiting behaviors by the end of learning based on a new game-theoretical solution concept. Lastly, the last topic answers how state space size can be reduced based on knowledge sharing and context-specific abstraction such that the learning complexity is less affected by non-stationarity. In summary, this thesis develops theoretical and algorithmic contributions to provide principled answers to the aforementioned topics on non-stationarity. The developed algorithms in this thesis demonstrate their effectiveness in a diverse suite of multiagent benchmark domains, including the full spectrum of mixed incentive, competitive, and cooperative environments. Ph.D. 2023-03-31T14:37:46Z 2023-03-31T14:37:46Z 2023-02 2023-02-15T14:05:24.284Z Thesis https://hdl.handle.net/1721.1/150177 In Copyright - Educational Use Permitted Copyright MIT http://rightsstatements.org/page/InC-EDU/1.0/ application/pdf Massachusetts Institute of Technology
spellingShingle Kim, Dong Ki
Effective Learning in Non-Stationary Multiagent Environments
title Effective Learning in Non-Stationary Multiagent Environments
title_full Effective Learning in Non-Stationary Multiagent Environments
title_fullStr Effective Learning in Non-Stationary Multiagent Environments
title_full_unstemmed Effective Learning in Non-Stationary Multiagent Environments
title_short Effective Learning in Non-Stationary Multiagent Environments
title_sort effective learning in non stationary multiagent environments
url https://hdl.handle.net/1721.1/150177
work_keys_str_mv AT kimdongki effectivelearninginnonstationarymultiagentenvironments