Automatic feature selection for model-based reinforcement learning in factored MDPs

Authors
Publication date 2009
Host editors
  • M.A. Wani
  • M. Kantardzic
  • V. Palade
  • L. Kurgan
  • A. Qi
Book title The Eighth International Conference on Machine Learning and Applications
Book subtitle proceedings : Miami Beach, Florida : 13-15 December 2009
ISBN
  • 9780769539263
Event Eighth International Conference on Machine Learning and Applications (ICMLA 2009), Miami Beach, FL, USA
Pages (from-to) 324-330
Publisher Los Alamitos, CA: IEEE Computer Society
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Feature selection is an important challenge in machine learning. Unfortunately, most methods for automating feature selection are designed for supervised learning tasks and are thus either inapplicable or impractical for reinforcement learning. This paper presents a new approach to feature selection specifically designed for the challenges of reinforcement learning. In our method, the agent learns a model, represented as a dynamic Bayesian network, of a factored Markov decision process, deduces a minimal feature set from this network, and efficiently computes a policy on this feature set using dynamic programming methods. Experiments in a stock-trading benchmark task demonstrate that this approach can reliably deduce minimal feature sets and that doing so can substantially improve performance and reduce the computational expense of planning.
Document type Conference contribution
Language English
Published at https://doi.org/10.1109/ICMLA.2009.71
Permalink to this page
Back