TY - GEN
T1 - Decentralized control of partially observable Markov decision processes
AU - Amato, Christopher
AU - Chowdhary, Girish
AU - Geramifard, Alborz
AU - Üre, N. Kemal
AU - Kochenderfer, Mykel J.
PY - 2013
Y1 - 2013
N2 - Markov decision processes (MDPs) are often used to model sequential decision problems involving uncertainty under the assumption of centralized control. However, many large, distributed systems do not permit centralized control due to communication limitations (such as cost, latency or corruption). This paper surveys recent work on decentralized control of MDPs in which control of each agent depends on a partial view of the world.We focus on a general framework where there may be uncertainty about the state of the environment, represented as a decentralized partially observable MDP (Dec-POMDP), but consider a number of subclasses with different assumptions about uncertainty and agent independence. In these models, a shared objective function is used, but plans of action must be based on a partial view of the environment. We describe the frameworks, along with the complexity of optimal control and important properties. We also provide an overview of exact and approximate solution methods as well as relevant applications. This survey provides an introduction to what has become an active area of research on these models and their solutions.
AB - Markov decision processes (MDPs) are often used to model sequential decision problems involving uncertainty under the assumption of centralized control. However, many large, distributed systems do not permit centralized control due to communication limitations (such as cost, latency or corruption). This paper surveys recent work on decentralized control of MDPs in which control of each agent depends on a partial view of the world.We focus on a general framework where there may be uncertainty about the state of the environment, represented as a decentralized partially observable MDP (Dec-POMDP), but consider a number of subclasses with different assumptions about uncertainty and agent independence. In these models, a shared objective function is used, but plans of action must be based on a partial view of the environment. We describe the frameworks, along with the complexity of optimal control and important properties. We also provide an overview of exact and approximate solution methods as well as relevant applications. This survey provides an introduction to what has become an active area of research on these models and their solutions.
UR - http://www.scopus.com/inward/record.url?scp=84902354924&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84902354924&partnerID=8YFLogxK
U2 - 10.1109/CDC.2013.6760239
DO - 10.1109/CDC.2013.6760239
M3 - Conference contribution
AN - SCOPUS:84902354924
SN - 9781467357173
T3 - Proceedings of the IEEE Conference on Decision and Control
SP - 2398
EP - 2405
BT - 2013 IEEE 52nd Annual Conference on Decision and Control, CDC 2013
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 52nd IEEE Conference on Decision and Control, CDC 2013
Y2 - 10 December 2013 through 13 December 2013
ER -