Using Neural Reinforcement Controllers in Robotics

Reinforcement Learning is a promising paradigm for the training of intelligent controllers when only a minimum of a priori knowledge is available. This is especially true for situations, where uncertainties and nonlinearities in sensors and actuators make the application of conventional control design techniques impossible or at least a very hard task. The article describes a neural control architecture based on the asynchronous dynamic programming approach (Barto et al., 1993). Its capabilities are shown on two applications from the domain of robotics.

[1]  R. Riedmiller,et al.  Aspects of learning neural control , 1994, Proceedings of IEEE International Conference on Systems, Man and Cybernetics.

[2]  Chris Watkins,et al.  Learning from delayed rewards , 1989 .

[3]  Sebastian Thrun,et al.  Planning with an Adaptive World Model , 1990, NIPS.

[4]  Andrew G. Barto,et al.  Learning to Act Using Real-Time Dynamic Programming , 1995, Artif. Intell..

[5]  Martin A. Riedmiller,et al.  Application of sequential reinforcement learning to control dynamic systems , 1996, Proceedings of International Conference on Neural Networks (ICNN'96).