Accounting for Task-Difficulty in Active Multi-Task Robot Control Learning
Alexander Fabisch, Jan Hendrik Metzen, Mario Michael Krell, Frank Kirchner
Editors: Hermann Engesser
In KI - Künstliche Intelligenz, German Journal on Artificial Intelligence - Organ des Fachbereiches "Künstliche Intelligenz" der Gesellschaft für Informatik e.V., Springer Verlag GmbH Heidelberg, the final publication is available at http://link.springer.com, volume 29, pages 369-377, 2015.
Contextual policy search is a reinforcement learning approach for multi-task learning in the context of robot control learning. It can be used to learn versatilely applicable skills that generalize over a range of tasks specified by a context vector. In this work, we combine contextual policy search with ideas from active learning for selecting the task in which the next trial will be performed. Moreover, we use active training set selection for reducing detrimental effects of exploration in the sampling policy. A core challenge in this approach is that the distribution of the obtained rewards may not be directly comparable between different tasks. We propose the novel approach PUBSVE for estimating a reward baseline and investigate empirically on benchmark problems and simulated robotic tasks to which extent this method can remedy the issue of non-comparable reward.
Contextual policy search, Multi-task learning, Active learning