Evolutionary computation for reinforcement learning
| Authors | |
|---|---|
| Publication date | 2012 |
| Host editors |
|
| Book title | Reinforcement learning: state-of-the-art |
| ISBN |
|
| Series | Adaptation, learning, and optimization, 12 |
| Pages (from-to) | 325-358 |
| Publisher | Heidelberg: Springer |
| Organisations |
|
| Abstract |
Algorithms for evolutionary computation, which simulate the process of natural selection to solve optimization problems, are an effective tool for discovering high-performing reinforcement-learning policies. Because they can automatically find good representations, handle continuous action spaces, and cope with partial observability, evolutionary reinforcement-learning approaches have a strong empirical track record, sometimes significantly outperforming temporal-difference methods. This chapter surveys research on the application of evolutionary computation to reinforcement learning, overviewing methods for evolving neural-network topologies and weights, hybrid methods that also use temporal-difference methods, coevolutionary methods for multi-agent settings, generative and developmental systems, and methods for on-line evolutionary reinforcement learning.
|
| Document type | Chapter |
| Language | English |
| Published at | https://doi.org/10.1007/978-3-642-27645-3_10 |
| Permalink to this page | |