Decentralized control of Partially Observable Markov Decision Processes using belief space macro-actions

The focus of this paper is on solving multi-robot planning problems in continuous spaces with partial observability. Decentralized Partially Observable Markov Decision Processes (Dec-POMDPs) are general models for multi-robot coordination problems, but representing and solving Dec-POMDPs is often in...

Full description

Bibliographic Details
Main Authors: Omidshafiei, Shayegan, Aghamohammadi, Aliakbar, Amato, Christopher, How, Jonathan P
Other Authors: Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory
Format: Article
Published: Institute of Electrical and Electronics Engineers (IEEE) 2018
Online Access:http://hdl.handle.net/1721.1/116391
https://orcid.org/0000-0003-0903-0137
https://orcid.org/0000-0002-6786-7384
https://orcid.org/0000-0001-8576-1930
_version_ 1826216774900645888
author Omidshafiei, Shayegan
Aghamohammadi, Aliakbar
Amato, Christopher
How, Jonathan P
author2 Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory
author_facet Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory
Omidshafiei, Shayegan
Aghamohammadi, Aliakbar
Amato, Christopher
How, Jonathan P
author_sort Omidshafiei, Shayegan
collection MIT
description The focus of this paper is on solving multi-robot planning problems in continuous spaces with partial observability. Decentralized Partially Observable Markov Decision Processes (Dec-POMDPs) are general models for multi-robot coordination problems, but representing and solving Dec-POMDPs is often intractable for large problems. To allow for a high-level representation that is natural for multi-robot problems and scalable to large discrete and continuous problems, this paper extends the Dec-POMDP model to the Decentralized Partially Observable Semi-Markov Decision Process (Dec-POSMDP). The Dec-POSMDP formulation allows asynchronous decision-making by the robots, which is crucial in multi-robot domains. We also present an algorithm for solving this Dec-POSMDP which is much more scalable than previous methods since it can incorporate closed-loop belief space macro-actions in planning. These macro-actions are automatically constructed to produce robust solutions. The proposed method's performance is evaluated on a complex multi-robot package delivery problem under uncertainty, showing that our approach can naturally represent multi-robot problems and provide high-quality solutions for large-scale problems.
first_indexed 2024-09-23T16:52:54Z
format Article
id mit-1721.1/116391
institution Massachusetts Institute of Technology
last_indexed 2024-09-23T16:52:54Z
publishDate 2018
publisher Institute of Electrical and Electronics Engineers (IEEE)
record_format dspace
spelling mit-1721.1/1163912022-09-29T22:10:03Z Decentralized control of Partially Observable Markov Decision Processes using belief space macro-actions Omidshafiei, Shayegan Aghamohammadi, Aliakbar Amato, Christopher How, Jonathan P Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory Massachusetts Institute of Technology. Department of Aeronautics and Astronautics Omidshafiei, Shayegan Aghamohammadi, Aliakbar Amato, Christopher How, Jonathan P The focus of this paper is on solving multi-robot planning problems in continuous spaces with partial observability. Decentralized Partially Observable Markov Decision Processes (Dec-POMDPs) are general models for multi-robot coordination problems, but representing and solving Dec-POMDPs is often intractable for large problems. To allow for a high-level representation that is natural for multi-robot problems and scalable to large discrete and continuous problems, this paper extends the Dec-POMDP model to the Decentralized Partially Observable Semi-Markov Decision Process (Dec-POSMDP). The Dec-POSMDP formulation allows asynchronous decision-making by the robots, which is crucial in multi-robot domains. We also present an algorithm for solving this Dec-POSMDP which is much more scalable than previous methods since it can incorporate closed-loop belief space macro-actions in planning. These macro-actions are automatically constructed to produce robust solutions. The proposed method's performance is evaluated on a complex multi-robot package delivery problem under uncertainty, showing that our approach can naturally represent multi-robot problems and provide high-quality solutions for large-scale problems. 2018-06-19T12:32:50Z 2018-06-19T12:32:50Z 2015-07 2018-03-22T12:31:07Z Article http://purl.org/eprint/type/ConferencePaper 978-1-4799-6923-4 http://hdl.handle.net/1721.1/116391 Omidshafiei, Shayegan, Ali-akbar Agha-mohammadi, Christopher Amato, and Jonathan P. How. “Decentralized Control of Partially Observable Markov Decision Processes Using Belief Space Macro-Actions.” 2015 IEEE International Conference on Robotics and Automation (ICRA) (May 2015). https://orcid.org/0000-0003-0903-0137 https://orcid.org/0000-0002-6786-7384 https://orcid.org/0000-0001-8576-1930 http://dx.doi.org/10.1109/ICRA.2015.7140035 2015 IEEE International Conference on Robotics and Automation (ICRA) Creative Commons Attribution-Noncommercial-Share Alike http://creativecommons.org/licenses/by-nc-sa/4.0/ application/pdf Institute of Electrical and Electronics Engineers (IEEE) Other repository
spellingShingle Omidshafiei, Shayegan
Aghamohammadi, Aliakbar
Amato, Christopher
How, Jonathan P
Decentralized control of Partially Observable Markov Decision Processes using belief space macro-actions
title Decentralized control of Partially Observable Markov Decision Processes using belief space macro-actions
title_full Decentralized control of Partially Observable Markov Decision Processes using belief space macro-actions
title_fullStr Decentralized control of Partially Observable Markov Decision Processes using belief space macro-actions
title_full_unstemmed Decentralized control of Partially Observable Markov Decision Processes using belief space macro-actions
title_short Decentralized control of Partially Observable Markov Decision Processes using belief space macro-actions
title_sort decentralized control of partially observable markov decision processes using belief space macro actions
url http://hdl.handle.net/1721.1/116391
https://orcid.org/0000-0003-0903-0137
https://orcid.org/0000-0002-6786-7384
https://orcid.org/0000-0001-8576-1930
work_keys_str_mv AT omidshafieishayegan decentralizedcontrolofpartiallyobservablemarkovdecisionprocessesusingbeliefspacemacroactions
AT aghamohammadialiakbar decentralizedcontrolofpartiallyobservablemarkovdecisionprocessesusingbeliefspacemacroactions
AT amatochristopher decentralizedcontrolofpartiallyobservablemarkovdecisionprocessesusingbeliefspacemacroactions
AT howjonathanp decentralizedcontrolofpartiallyobservablemarkovdecisionprocessesusingbeliefspacemacroactions