A Cross-Entropy Approach to Solving Dec-POMDPs
| Authors |
|
|---|---|
| Publication date | 2008 |
| Host editors |
|
| Book title | Advances in Intelligent and Distributed Computing |
| Book subtitle | proceedings of the 1st International Symposium on Intelligent and Distributed Computing IDC'2007, Craiova, Romania, October 2007 |
| ISBN |
|
| ISBN (electronic) |
|
| Series | Studies in Computational Intelligence |
| Event | 1st International Symposium on Intelligent and Distributed Computing |
| Pages (from-to) | 145-154 |
| Publisher | Berlin: Springer |
| Organisations |
|
| Abstract |
In this paper we focus on distributed multiagent planning under uncertainty. For single-agent planning under uncertainty, the partially observable Markov decision process (POMDP) is the dominant model (see [Spaan and Vlassis, 2005] and references therein). Recently, several generalizations of the POMDP to multiagent settings have been proposed. Here we focus on the decentralized POMDP (Dec-POMDP) model for multiagent planning under uncertainty [Bernstein et al., 2002, Goldman and Zilberstein, 2004]. Solving a Dec-POMDP amounts to finding a set of optimal policies for the agents that maximize the expected shared reward. However, solving a Dec-POMDP has proven to be hard (NEXP-complete): The number of possible deterministic policies for a single agent grows doubly exponentially with the planning horizon, and exponentially with the number of actions and observations available. As a result, the focus has shifted to approximate solution techniques [Nair et al., 2003, Emery-Montemerlo et al., 2005, Oliehoek and Vlassis, 2007].
|
| Document type | Conference contribution |
| Language | English |
| Published at | https://doi.org/10.1007/978-3-540-74930-1_15 |
| Permalink to this page | |