A Cross-Entropy Approach to Solving Dec-POMDPs

Authors
Publication date 2008
Host editors
  • C. Badica
  • M. Paprzycki
Book title Advances in Intelligent and Distributed Computing
Book subtitle proceedings of the 1st International Symposium on Intelligent and Distributed Computing IDC'2007, Craiova, Romania, October 2007
ISBN
  • 9783540749295
ISBN (electronic)
  • 9783540749301
Series Studies in Computational Intelligence
Event 1st International Symposium on Intelligent and Distributed Computing
Pages (from-to) 145-154
Publisher Berlin: Springer
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
In this paper we focus on distributed multiagent planning under uncertainty. For single-agent planning under uncertainty, the partially observable Markov decision process (POMDP) is the dominant model (see [Spaan and Vlassis, 2005] and references therein). Recently, several generalizations of the POMDP to multiagent settings have been proposed. Here we focus on the decentralized POMDP (Dec-POMDP) model for multiagent planning under uncertainty [Bernstein et al., 2002, Goldman and Zilberstein, 2004]. Solving a Dec-POMDP amounts to finding a set of optimal policies for the agents that maximize the expected shared reward. However, solving a Dec-POMDP has proven to be hard (NEXP-complete): The number of possible deterministic policies for a single agent grows doubly exponentially with the planning horizon, and exponentially with the number of actions and observations available. As a result, the focus has shifted to approximate solution techniques [Nair et al., 2003, Emery-Montemerlo et al., 2005, Oliehoek and Vlassis, 2007].
Document type Conference contribution
Language English
Published at https://doi.org/10.1007/978-3-540-74930-1_15
Permalink to this page
Back