Switching between different state representations in reinforcement learning
| Authors | |
|---|---|
| Publication date | 2008 |
| Host editors |
|
| Book title | Proceedings of the IASTED International Conference on Artificial Intelligence and Applications: As part of the 26th IASTED International Multi-Conference on Applied Informatics, February 11-13, 2008, Innsbruck, Austria |
| ISBN |
|
| Event | IASTED International Conference on Artificial Intelligence and Applications (AIA 2008), Innsbruck, Austria |
| Pages (from-to) | 595-163 |
| Publisher | Anaheim, CA: ACTA |
| Organisations |
|
| Abstract |
This paper proposes a reinforcement learning architecture con taining multiple "experts", each of which is a specialist in a dif ferent region in the overall state space. The central idea is that the different experts use qualitatively different (but sufficiently Markov) state representations, each of which captures different information regarding the true underlying world state, and which for that reason is suitable for a different part of the state space. The experts themselves learn to switch to another state represen tation (other expert) by having switching actions. Value functions can be learned using standard reinforcement learning algorithms. This architecture has important advantages in RL problems that have large state spaces or where a sensor system must inherently choose between mutually exclusive state representations. Experi ments in a small, proof-of-principle experiment as well as a larger, more realistic experiment illustrate the validity of this approach.
|
| Document type | Conference contribution |
| Published at | http://www.actapress.com/Abstract.aspx?paperId=32307 |
| Permalink to this page | |