Learning to Control a Low-Cost Manipulator using Data-Efficient Reinforcement Learning
File(s)rss2011_revision.pdf (2.67 MB)
Accepted version
Author(s)
Deisenroth, Marc P
Rasmussen, Carl E
Fox, Dieter
Type
Conference Paper
Abstract
Over the last years, there has been substantial progress in robust manipulation in unstructured environments. The long-term goal of our work is to get away from precise, but very expensive robotic systems and to develop affordable, potentially imprecise, self-adaptive manipulator systems that can interactively perform tasks such as playing with children. In this paper, we demonstrate how a low-cost off-the-shelf robotic system can learn closed-loop policies for a stacking task in only a handful of trials-from scratch. Our manipulator is inaccurate and provides no pose feedback. For learning a controller in the work space of a Kinect-style depth camera, we use a model-based reinforcement learning technique. Our learning method is data efficient, reduces model bias, and deals with several noise sources in a principled way during long-term planning. We present a way of incorporating state-space constraints into the learning process and analyze the learning gain by exploiting the sequential structure of the stacking task.
Date Issued
2011-06
Citation
Proceedings of the International Conference on Robotics: Science and Systems (RSS 2011), 2011
ISBN
9780262519687
0262517795
Publisher
MIT Press
Journal / Book Title
Proceedings of the International Conference on Robotics: Science and Systems (RSS 2011)
Copyright Statement
© 2011 MIT Press
Description
02.07.13 KB. OK to add accepted version to Sprial, embargo has elapsed. MIT Press
Source
2011 Robotics: Science and Systems Conference
Source Place
Los Angeles, California
Notes
timestamp: 2011.01.20
Start Date
2011-06-27
Finish Date
2011-07-01
Coverage Spatial
Los Angeles, California.