Emergent Real-World Robotic Skills via Unsupervised Off-Policy Reinforcement Learning
Citations Over TimeTop 12% of 2020 papers
Abstract
Reinforcement learning provides a general framework for learning robotic skills while minimizing engineering effort.However, most reinforcement learning algorithms assume that a well-designed reward function is provided, and learn a single behavior for that single reward function.Such reward functions can be difficult to design in practice.Can we instead develop efficient reinforcement learning methods that acquire diverse skills without any reward function, and then re-purpose these skills for downstream tasks?In this paper, we demonstrate that a recently proposed unsupervised skill discovery algorithm can be extended into an efficient off-policy method, making it suitable for performing unsupervised reinforcement learning in the real world.Firstly, we show that our proposed algorithm provides substantial improvement in learning efficiency, making rewardfree real-world training feasible.Secondly, we move beyond the simulation environments and evaluate the algorithm on real physical hardware.On quadrupeds, we observe that locomotion skills with diverse gaits and different orientations emerge without any rewards or demonstrations.We also demonstrate that the learned skills can be composed using model predictive control for goal-oriented navigation, without any additional training.
Related Papers
- → A Deep Reinforcement Learning Strategy for UAV Autonomous Landing on a Platform(2022)15 cited
- → Cost-Efficient Reinforcement Learning for Optimal Trade Execution on Dynamic Market Environment(2022)2 cited
- Using DataGrid Control to Realize DataBase of Querying in VB6.0(2000)
- Susquehanna Chorale Spring Concert "Roots and Wings"(2017)
- → DETERMINING QUALITY REQUIREMENTS AT THE UNIVERSITIES TO IMPROVE THE QUALITY OF EDUCATION(2018)