TY - JOUR
T1 - Robots Learn Increasingly Complex Tasks with Intrinsic Motivation and Automatic Curriculum Learning
T2 - Domain Knowledge by Emergence of Affordances, Hierarchical Reinforcement and Active Imitation Learning
AU - Nguyen, Sao Mai
AU - Duminy, Nicolas
AU - Manoury, Alexandre
AU - Duhaut, Dominique
AU - Buche, Cedric
N1 - Publisher Copyright:
© 2021, Gesellschaft für Informatik e.V. and Springer-Verlag GmbH Germany, part of Springer Nature.
PY - 2021/3/1
Y1 - 2021/3/1
N2 - Multi-task learning by robots poses the challenge of the domain knowledge: complexity of tasks, complexity of the actions required, relationship between tasks for transfer learning. We demonstrate that this domain knowledge can be learned to address the challenges in life-long learning. Specifically, the hierarchy between tasks of various complexities is key to infer a curriculum from simple to composite tasks. We propose a framework for robots to learn sequences of actions of unbounded complexity in order to achieve multiple control tasks of various complexity. Our hierarchical reinforcement learning framework, named SGIM-SAHT, offers a new direction of research, and tries to unify partial implementations on robot arms and mobile robots. We outline our contributions to enable robots to map multiple control tasks to sequences of actions: representations of task dependencies, an intrinsically motivated exploration to learn task hierarchies, and active imitation learning. While learning the hierarchy of tasks, it infers its curriculum by deciding which tasks to explore first, how to transfer knowledge, and when, how and whom to imitate.
AB - Multi-task learning by robots poses the challenge of the domain knowledge: complexity of tasks, complexity of the actions required, relationship between tasks for transfer learning. We demonstrate that this domain knowledge can be learned to address the challenges in life-long learning. Specifically, the hierarchy between tasks of various complexities is key to infer a curriculum from simple to composite tasks. We propose a framework for robots to learn sequences of actions of unbounded complexity in order to achieve multiple control tasks of various complexity. Our hierarchical reinforcement learning framework, named SGIM-SAHT, offers a new direction of research, and tries to unify partial implementations on robot arms and mobile robots. We outline our contributions to enable robots to map multiple control tasks to sequences of actions: representations of task dependencies, an intrinsically motivated exploration to learn task hierarchies, and active imitation learning. While learning the hierarchy of tasks, it infers its curriculum by deciding which tasks to explore first, how to transfer knowledge, and when, how and whom to imitate.
KW - Continual learning
KW - Curriculum learning
KW - Hierarchical reinforcement learning
KW - Intrinsic motivation
KW - Multi-task learning
KW - Transfer learning
U2 - 10.1007/s13218-021-00708-8
DO - 10.1007/s13218-021-00708-8
M3 - Article
AN - SCOPUS:85100834197
SN - 0933-1875
VL - 35
SP - 81
EP - 90
JO - KI - Kunstliche Intelligenz
JF - KI - Kunstliche Intelligenz
IS - 1
ER -