Learning to Control a Low-Cost Manipulator using Data-Efficient Reinforcement Learning
Citations Over TimeTop 1% of 2011 papers
Abstract
Over the last years, there has been substantial progress in robust manipulation in unstructured environments. The long-term goal of our work is to get away from precise, but very expensive robotic systems and to develop affordable, potentially imprecise, self-adaptive manipulator systems that can interactively perform tasks such as playing with children. In this paper, we demonstrate how a low-cost off-the-shelf robotic system can learn closed-loop policies for a stacking task in only a handful of trials-from scratch. Our manipulator is inaccurate and provides no pose feedback. For learning a controller in the work space of a Kinect-style depth camera, we use a model-based reinforcement learning technique. Our learning method is data efficient, reduces model bias, and deals with several noise sources in a principled way during long-term planning. We present a way of incorporating state-space constraints into the learning process and analyze the learning gain by exploiting the sequential structure of the stacking task.
Related Papers
- → Optimization of lightweight task offloading strategy for mobile edge computing based on deep reinforcement learning(2019)207 cited
- → Task offloading method of edge computing in internet of vehicles based on deep reinforcement learning(2022)112 cited
- → Physician-Friendly Machine Learning: A Case Study with Cardiovascular Disease Risk Prediction(2019)71 cited
- → Resource allocation for network slicing in dynamic multi-tenant networks: A deep reinforcement learning approach(2022)17 cited
- → Performance Improvement of MapReduce Framework in Heterogeneous Context using Reinforcement Learning(2015)19 cited