Robot learning by nonparametric regression

We present an approach to robot learning based on a nonparametric regression technique, locally weighted regression. The model of the task to be performed is represented by infinitely many local linear models, i.e., the (hyper-) tangent planes at every query point. Such a model, however, is only generated when a query performed and is not retained. The architectural parameters of our approach, such as distance metrics, are also a function of the current query point instead of being global. Statistical tests are presented for when a local model is good enough such that it can be reliably used to build a local controller. These statistical measures also direct the exploration of the robot. We explicitly deal with the case where prediction accuracy requirements exist during exploration. By gradually shifting a center of exploration and controlling the speed of the shift with local prediction accuracy, a goal-directed exploration of state space takes place along the fringes of the current data support until the task goal is achieved. We illustrate this approach by describing how it has been used to enable a robot to learn a challenging juggling task.<<ETX>>

[1]  S. Schaal,et al.  Robot juggling: implementation of memory-based learning , 1994, IEEE Control Systems.

[2]  Stefan Schaal,et al.  Open loop stable control strategies for robot juggling , 1993, [1993] Proceedings IEEE International Conference on Robotics and Automation.

[3]  Stefan Schaal,et al.  Learning passive motor control strategies with genetic algorithms , 1993 .

[4]  Michael I. Jordan,et al.  Forward Models: Supervised Learning with a Distal Teacher , 1992, Cogn. Sci..

[5]  Christopher G. Atkeson,et al.  What should be learned , 1992 .

[6]  Anders Krogh,et al.  Introduction to the theory of neural computation , 1994, The advanced book program.

[7]  Andrew W. Moore,et al.  Fast, Robust Adaptive Control by Learning only Forward Models , 1991, NIPS.

[8]  Richard S. Sutton,et al.  Integrated Architectures for Learning, Planning, and Reacting Based on Approximating Dynamic Programming , 1990, ML.

[9]  Christopher G. Atkeson,et al.  Task-level robot learning: juggling a tennis ball more accurately , 1989, Proceedings, 1989 International Conference on Robotics and Automation.

[10]  David J. Reinkensmeyer,et al.  Task-level robot learning , 1988, Proceedings. 1988 IEEE International Conference on Robotics and Automation.

[11]  Christopher G. Atkeson,et al.  Model-Based Control of a Robot Manipulator , 1988 .

[12]  Farmer,et al.  Predicting chaotic time series. , 1987, Physical review letters.

[13]  William H. Press,et al.  Numerical Recipes in FORTRAN - The Art of Scientific Computing, 2nd Edition , 1987 .

[14]  William H. Press,et al.  Numerical Recipes in C The Art of Scientific Computing , 1995 .

[15]  R. H. Myers Classical and modern regression with applications , 1986 .

[16]  Leo Breiman,et al.  Classification and Regression Trees , 1984 .

[17]  W. Cleveland Robust Locally Weighted Regression and Smoothing Scatterplots , 1979 .

[18]  Jon Louis Bentley,et al.  An Algorithm for Finding Best Matches in Logarithmic Expected Time , 1976, TOMS.

[19]  G. Wahba,et al.  A completely automatic french curve: fitting spline functions by cross validation , 1975 .

[20]  M. Ciletti,et al.  The computation and theory of optimal control , 1972 .

[21]  Stephen R. McReynolds,et al.  The computation and theory of optimal control , 1970 .