Automatic feature selection for model-based reinforcement learning in factored MDPs
| Authors |
|
|---|---|
| Publication date | 2009 |
| Host editors |
|
| Book title | The Eighth International Conference on Machine Learning and Applications |
| Book subtitle | proceedings : Miami Beach, Florida : 13-15 December 2009 |
| ISBN |
|
| Event | Eighth International Conference on Machine Learning and Applications (ICMLA 2009), Miami Beach, FL, USA |
| Pages (from-to) | 324-330 |
| Publisher | Los Alamitos, CA: IEEE Computer Society |
| Organisations |
|
| Abstract |
Feature selection is an important challenge in machine learning. Unfortunately, most methods for automating feature selection are designed for supervised learning tasks and are thus either inapplicable or impractical for reinforcement learning. This paper presents a new approach to feature selection specifically designed for the challenges of reinforcement learning. In our method, the agent learns a model, represented as a dynamic Bayesian network, of a factored Markov decision process, deduces a minimal feature set from this network, and efficiently computes a policy on this feature set using dynamic programming methods. Experiments in a stock-trading benchmark task demonstrate that this approach can reliably deduce minimal feature sets and that doing so can substantially improve performance and reduce the computational expense of planning.
|
| Document type | Conference contribution |
| Language | English |
| Published at | https://doi.org/10.1109/ICMLA.2009.71 |
| Permalink to this page | |