Billard, A., Calinon, S., Dillmann R. and Schaal, S. (2008)
Robot Programming by Demonstration
Siciliano, B. and Khatib, O. (eds.). Handbook of Robotics, pp. 1371-1394. Springer.
Abstract
Robot Programming by demonstration (PbD) has become a central
topic of robotics that spans across general research areas such as
human-robot interaction, machine learning, machine vision and
motor control.
Robot PbD started about 30 years ago, and has grown importantly during
the past decade. The rationale for moving from purely
preprogrammed robots to very flexible user-based interfaces for
training robots to perform a task is three-fold.
First and foremost, PbD, also referred to as imitation
learning, is a powerful mechanism for reducing the complexity of
search spaces for learning. When observing either good or bad
examples, one can reduce the search for a possible solution, by
either starting the search from the observed good solution (local
optima), or conversely, by eliminating from the search space what
is known as a bad solution. Imitation learning is, thus, a
powerful tool for enhancing and accelerating learning in both
animals and artifacts.
Second, imitation learning offers an implicit means of training a
machine, such that explicit and tedious programming of a task by a
human user can be minimized or eliminated. Imitation learning is thus a natural
means of interacting with a machine that would be accessible to
lay people.
Third, studying and modeling the coupling of perception and
action, which is at the core of imitation learning, helps us to
understand the mechanisms by which the self-organization of
perception and action could arise during development. The
reciprocal interaction of perception and action could explain how
competence in motor control can be grounded in rich structure of
perceptual variables, and vice versa, how the processes of
perception can develop as means to create successful actions.
PbD promises were thus multiple. On the one hand, one hoped that
it would make learning faster, in contrast to tedious
reinforcement learning methods or trials-and-error learning. On
the other hand, one expected that the methods, being
user-friendly, would enhance the application of robots in human
daily environments. Recent progresses in the field, which we
review in this chapter, show that the field has made a leap
forward during the past decade toward these goals. In addition, we
anticipate that these promises may be fulfilled very soon.
Bibtex reference
@incollection{Billard08chapter, author="A. Billard and S. Calinon and R. Dillmann and S. Schaal", title="Robot Programming by Demonstration", booktitle="Handbook of Robotics", editor="B. Siciliano and O. Khatib", publisher="Springer", address="Secaucus, NJ, USA", year="2008", pages="1371--1394" }