Near optimal closed-loop control Application to electric power systems

[1]  Louis Wehenkel,et al.  A reinforcement learning based discrete supplementary control for power system transient stability enhancement , 2005 .

[2]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[3]  Rémi Munos,et al.  A Study of Reinforcement Learning in the Continuous Case by the Means of Viscosity Solutions , 2000, Machine Learning.

[4]  Paul Bourgine,et al.  Exploration of Multi-State Environments: Local Measures and Back-Propagation of Uncertainty , 1999, Machine Learning.

[5]  Richard S. Sutton,et al.  Reinforcement learning with replacing eligibility traces , 2004, Machine Learning.

[6]  T. Başar,et al.  A New Approach to Linear Filtering and Prediction Problems , 2001 .

[7]  Damien Ernst,et al.  Transient Stability of Power Systems: A Unified Approach to Assessment and Control , 2000 .

[8]  Tariq Samad,et al.  SEPIA. A simulator for electric power industry agents , 2000 .

[9]  Leslie Pack Kaelbling,et al.  Practical Reinforcement Learning in Continuous Spaces , 2000, ICML.

[10]  H. Kushner Numerical Methods for Stochastic Control Problems in Continuous Time , 2000 .

[11]  Mehrdad Ghandhari,et al.  Control Lyapunov Functions : A Control Strategy for Damping of Power Oscillations in Large Power Systems , 2000 .

[12]  Laszlo Gyugyi,et al.  Understanding FACTS: Concepts and Technology of Flexible AC Transmission Systems , 1999 .

[13]  O. Hernández-Lerma,et al.  Discrete-time Markov control processes , 1999 .

[14]  O. Hernández-Lerma,et al.  Further topics on discrete-time Markov control processes , 1999 .

[15]  Leslie Pack Kaelbling,et al.  Planning and Acting in Partially Observable Stochastic Domains , 1998, Artif. Intell..

[16]  A. M. Wildberger,et al.  Complex adaptive systems: concepts and power industry applications , 1997 .

[17]  Rémi Munos Apprentissage par renforcement, étude du cas continu , 1997 .

[18]  Nicolas Meuleau Le dilemme entre exploration et exploitation dans l'apprentissage par renforcement : optimisation adaptative des modeles de decision multi-etats , 1996 .

[19]  M. Littman The Witness Algorithm: Solving Partially Observable Markov Decision Processes , 1994 .

[20]  Michael I. Jordan,et al.  MASSACHUSETTS INSTITUTE OF TECHNOLOGY ARTIFICIAL INTELLIGENCE LABORATORY and CENTER FOR BIOLOGICAL AND COMPUTATIONAL LEARNING DEPARTMENT OF BRAIN AND COGNITIVE SCIENCES , 1996 .

[21]  John N. Tsitsiklis,et al.  Asynchronous stochastic approximation and Q-learning , 1993, Proceedings of 32nd IEEE Conference on Decision and Control.

[22]  Arun G. Phadke,et al.  Synchronized phasor measurements in power system , 1993 .

[23]  C. Atkeson,et al.  Prioritized Sweeping: Reinforcement Learning with Less Data and Less Time , 1993, Machine Learning.

[24]  D. Moore Simplicial Mesh Generation with Applications , 1992 .

[25]  K. R. Padiyar,et al.  ENERGY FUNCTION ANALYSIS FOR POWER SYSTEM STABILITY , 1990 .

[26]  C. Watkins Learning from delayed rewards , 1989 .

[27]  Warren D. Smith Studies in computational geometry motivated by mesh generation , 1989 .

[28]  Katsuhiko Ogata,et al.  Discrete-time control systems , 1987 .

[29]  P. Schweitzer,et al.  Generalized polynomial approximations in Markovian decision processes , 1985 .

[30]  John F. Sallee The Middle-Cut Triangulations of the n-Cube , 1984 .

[31]  P. L'Ecuyer,et al.  Approximation and bounds in discrete event dynamic programming , 1983, The 23rd IEEE Conference on Decision and Control.

[32]  Martin L. Puterman,et al.  Action Elimination Procedures for Modified Policy Iteration Algorithms , 1982, Oper. Res..

[33]  J. Hammersley SIMULATION AND THE MONTE CARLO METHOD , 1982 .

[34]  John F. Sallee A triangulation of the n-cube , 1982, Discret. Math..

[35]  Ward Whitt,et al.  Approximations of Dynamic Programs, I , 1978, Math. Oper. Res..

[36]  M. Puterman,et al.  Modified Policy Iteration Algorithms for Discounted Markov Decision Problems , 1978 .

[37]  D. Luenberger Optimization by Vector Space Methods , 1968 .