- An Analysis of Piecewise-Linear and Convex Value Functions for Active Perception POMDPs
- Number of pages
- Amsterdam: Intelligent Autonomous Systems, University of Amsterdam
- IAS technical report
- Volume | Edition (Serie)
- Document type
- Faculty of Science (FNWI)
- Informatics Institute (IVI)
In active perception tasks, an agent aims to select actions that reduce its uncertainty about a hidden state. While partially observable Markov decision processes (POMDPs) are a natural model for such problems, reward functions that directly penalize uncertainty in the agent's belief can remove the piecewise-linear and convex (PWLC) property of the value function required by most POMDP planners.
This paper analyses POMDP and POMDP-IR, two frameworks that restore the PWLC property in active perception tasks. We establish the mathematical equivalence of the two frameworks and show that both admit a decomposition of the maximization performed in the Bellman backup, yielding substantial computational savings. We also present an empirical analysis on data from real multi-camera tracking systems that illustrates these savings and analyzes the critical factors in the performance of POMDP planners in such tasks.
If you believe that digital publication of certain material infringes any of your rights or (privacy) interests, please let the Library know, stating your reasons. In case of a legitimate complaint, the Library will make the material inaccessible and/or remove it from the website. Please Ask the Library, or send a letter to: Library of the University of Amsterdam, Secretariat, Singel 425, 1012 WP Amsterdam, The Netherlands. You will be contacted as soon as possible.