A Large-Scale Study of Agents Learning from Human Reward
| Authors |
|
|---|---|
| Publication date | 2015 |
| Book title | AAMAS '15 |
| Book subtitle | proceedings of the 2015 International Conference on Autonomous Agents & Multiagent Systems : May, 4-8, 2015, Istanbul, Turkey |
| ISBN |
|
| ISBN (electronic) |
|
| Event | 14th International Joint Conference on Autonomous Agents and Multi-Agent Systems |
| Volume | Issue number | 3 |
| Pages (from-to) | 1771-1772 |
| Publisher | Richland, SC: International Foundation for Autonomous Agents and Multiagent Systems |
| Organisations |
|
| Abstract |
The TAMER framework, which provides a way for agents to learn to
solve tasks using human-generated rewards, has been examined in several
small-scale studies, each with a few dozen subjects. In this paper, we
present the results of the first large-scale study of TAMER, which was
performed at the NEMO science museum in Amsterdam and involved 561
subjects. Our results show for the first time that an agent using TAMER
can successfully learn to play Infinite Mario, a challenging
reinforcement-learning benchmark problem based on the popular video
game, given feedback from both adult (N=209) and child (N=352)
trainers. In addition, our study supports prior studies demonstrating
the importance of bidirectional feedback and competitive elements in the
training interface. Finally, our results also shed light on the
potential for using trainers' facial expressions as a reward signal, as
well as the role of age and gender in trainer behavior and agent
performance.
|
| Document type | Conference contribution |
| Note | Extended abstract. |
| Language | English |
| Published at | http://www.aamas-conference.org/Proceedings/aamas2015/aamas/p1771.pdf https://dl.acm.org/citation.cfm?id=2773428 |
| Downloads |
p1771-li
(Final published version)
|
| Permalink to this page | |