Variable Impedance Control - A Reinforcement Learning Approach

One of the hallmarks of the performance, versatility, and robustness of biological motor control is the ability to adapt the impedance of the overall biomechanical system to different task requirements and stochastic disturbances. A transfer of this principle to robotics is desirable, for instance to enable robots to work robustly and safely in everyday human environments. It is, however, not trivial to derive variable impedance controllers for practical high DOF robotic tasks. In this contribution, we ac- complish such gain scheduling with a reinforcement learning ap- proach algorithm, PI 2 (Policy Improvement with Path Integrals). PI 2 is a model-free, sampling based learning method derived from first principles of optimal control. The PI 2 algorithm requires no tuning of algorithmic parameters besides the exploration noise. The designer can thus fully focus on cost function design to specify the task. From the viewpoint of robotics, a particular useful property of PI 2 is that it can scale to problems of many DOFs, so that RL on real robotic systems becomes feasible. We sketch the PI 2 algorithm and its theoretical properties, and how it is applied to gain scheduling. We evaluate our approach by presenting results on two different simulated robotic systems, a 3-DOF Phantom Premium Robot and a 6-DOF Kuka Lightweight Robot. We investigate tasks where the optimal strategy requires both tuning of the impedance of the end-effector, and tuning of a reference trajectory. The results show that we can use path integral based RL not only for planning but also to derive variable gain feedback controllers in realistic scenarios. Thus, the power of variable impedance control is made available to a wide variety of robotic systems and practical applications. I. INTRODUCTION

[1]  Johan E. Harris Vertebrate Locomotion , 1961, Nature.

[2]  David Q. Mayne,et al.  Differential dynamic programming , 1972, The Mathematical Gazette.

[3]  Richard A. Tapia,et al.  Practical Methods of Optimization, Volume 2: Constrained Optimization (R. Fletcher) , 1984 .

[4]  Neville Hogan,et al.  Impedance Control: An Approach to Manipulation , 1984, 1984 American Control Conference.

[5]  Neville Hogan,et al.  Impedance control - An approach to manipulation. I - Theory. II - Implementation. III - Applications , 1985 .

[6]  S. Yakowitz The stagewise Kuhn-Tucker condition and differential dynamic programming , 1986 .

[7]  B. Øksendal Stochastic differential equations : an introduction with applications , 1987 .

[8]  R. Fletcher Practical Methods of Optimization , 1988 .

[9]  Robert F. Stengel,et al.  Optimal Control and Estimation , 1994 .

[10]  Tamer Başar,et al.  H1-Optimal Control and Related Minimax Design Problems , 1995 .

[11]  T. Basar,et al.  H∞-0ptimal Control and Related Minimax Design Problems: A Dynamic Game Approach , 1996, IEEE Trans. Autom. Control..

[12]  J. Yong Relations among ODEs, PDEs, FSDEs, BSDEs, and FBSDEs , 1997, Proceedings of the 36th IEEE Conference on Decision and Control.

[13]  T. Vincent,et al.  Nonlinear and Optimal Control Systems , 1997 .

[14]  Vijay Kumar,et al.  On the generation of smooth three-dimensional rigid body motions , 1998, IEEE Trans. Robotics Autom..

[15]  Bruno Siciliano,et al.  Modelling and Control of Robot Manipulators , 1997, Advanced Textbooks in Control and Signal Processing.

[16]  Jun Morimoto,et al.  Minimax Differential Dynamic Programming: An Application to Robust Biped Walking , 2002, NIPS.

[17]  Emanuel Todorov,et al.  Iterative Linear Quadratic Regulator Design for Nonlinear Biological Movement Systems , 2004, ICINCO.

[18]  H. Kappen Linear theory for control of nonlinear stochastic systems. , 2004, Physical review letters.

[19]  H. Kappen Path integrals and symmetry breaking for optimal control theory , 2005, physics/0505066.

[20]  Emanuel Todorov,et al.  Linearly-solvable Markov decision problems , 2006, NIPS.

[21]  William D. Smart,et al.  Receding Horizon Differential Dynamic Programming , 2007, NIPS.

[22]  G. Lantoine,et al.  A Hybrid Differential Dynamic Programming Algorithm for Robust Low-Thrust Optimization , 2008 .

[23]  Emanuel Todorov,et al.  General duality between optimal control and estimation , 2008, 2008 47th IEEE Conference on Decision and Control.

[24]  Hilbert J. Kappen,et al.  Graphical Model Inference in Optimal Control of Stochastic Multi-Agent Systems , 2008, J. Artif. Intell. Res..

[25]  Emanuel Todorov,et al.  Efficient computation of optimal actions , 2009, Proceedings of the National Academy of Sciences.

[26]  Stefan Schaal,et al.  Compliant quadruped locomotion over rough terrain , 2009, 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[27]  Stefan Schaal,et al.  Reinforcement learning of motor skills in high dimensions: A path integral approach , 2010, 2010 IEEE International Conference on Robotics and Automation.

[28]  Stefan Schaal,et al.  A Generalized Path Integral Control Approach to Reinforcement Learning , 2010, J. Mach. Learn. Res..