Probabilistic Movement Primitives for Coordination of Multiple Human-Robot Collaborative Tasks
Abstract
This paper proposes an interaction learning method for collaborative and assistive robots based on movement primitives. The method allows for both action recognition and human-robot movement coordination. It uses imitation learning to construct a mixture model of human-robot interaction primitives. This probabilistic model allows the assistive trajectory of the robot to be inferred from human observations. The method is scalable in relation to the number of tasks and can learn nonlinear correlations between the trajectories that describe the human-robot interaction. We evaluated the method experimentally with a lightweight robot arm in a variety of assistive scenarios, including the coordinated handover of a bottle to a human, and the collaborative assembly of a toolbox. Potential applications of the method are personal caregiver robots, control of intelligent prosthetic devices, and robot coworkers in factories.
BibTeX
@article{Maeda-2017-112215,author = {Guilherme Maeda and Gerhard Neumann and Marco Ewerton and Rudolf Lioutikov and Oliver Kroemer and Jan Peters},
title = {Probabilistic Movement Primitives for Coordination of Multiple Human-Robot Collaborative Tasks},
journal = {Autonomous Robots},
year = {2017},
month = {March},
volume = {41},
number = {3},
pages = {593 - 612},
}