Near optimal closed-loop control Application to electric power systems
暂无分享,去创建一个
[1] Louis Wehenkel,et al. A reinforcement learning based discrete supplementary control for power system transient stability enhancement , 2005 .
[2] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[3] Rémi Munos,et al. A Study of Reinforcement Learning in the Continuous Case by the Means of Viscosity Solutions , 2000, Machine Learning.
[4] Paul Bourgine,et al. Exploration of Multi-State Environments: Local Measures and Back-Propagation of Uncertainty , 1999, Machine Learning.
[5] Richard S. Sutton,et al. Reinforcement learning with replacing eligibility traces , 2004, Machine Learning.
[6] T. Başar,et al. A New Approach to Linear Filtering and Prediction Problems , 2001 .
[7] Damien Ernst,et al. Transient Stability of Power Systems: A Unified Approach to Assessment and Control , 2000 .
[8] Tariq Samad,et al. SEPIA. A simulator for electric power industry agents , 2000 .
[9] Leslie Pack Kaelbling,et al. Practical Reinforcement Learning in Continuous Spaces , 2000, ICML.
[10] H. Kushner. Numerical Methods for Stochastic Control Problems in Continuous Time , 2000 .
[11] Mehrdad Ghandhari,et al. Control Lyapunov Functions : A Control Strategy for Damping of Power Oscillations in Large Power Systems , 2000 .
[12] Laszlo Gyugyi,et al. Understanding FACTS: Concepts and Technology of Flexible AC Transmission Systems , 1999 .
[13] O. Hernández-Lerma,et al. Discrete-time Markov control processes , 1999 .
[14] O. Hernández-Lerma,et al. Further topics on discrete-time Markov control processes , 1999 .
[15] Leslie Pack Kaelbling,et al. Planning and Acting in Partially Observable Stochastic Domains , 1998, Artif. Intell..
[16] A. M. Wildberger,et al. Complex adaptive systems: concepts and power industry applications , 1997 .
[17] Rémi Munos. Apprentissage par renforcement, étude du cas continu , 1997 .
[18] Nicolas Meuleau. Le dilemme entre exploration et exploitation dans l'apprentissage par renforcement : optimisation adaptative des modeles de decision multi-etats , 1996 .
[19] M. Littman. The Witness Algorithm: Solving Partially Observable Markov Decision Processes , 1994 .
[20] Michael I. Jordan,et al. MASSACHUSETTS INSTITUTE OF TECHNOLOGY ARTIFICIAL INTELLIGENCE LABORATORY and CENTER FOR BIOLOGICAL AND COMPUTATIONAL LEARNING DEPARTMENT OF BRAIN AND COGNITIVE SCIENCES , 1996 .
[21] John N. Tsitsiklis,et al. Asynchronous stochastic approximation and Q-learning , 1993, Proceedings of 32nd IEEE Conference on Decision and Control.
[22] Arun G. Phadke,et al. Synchronized phasor measurements in power system , 1993 .
[23] C. Atkeson,et al. Prioritized Sweeping: Reinforcement Learning with Less Data and Less Time , 1993, Machine Learning.
[24] D. Moore. Simplicial Mesh Generation with Applications , 1992 .
[25] K. R. Padiyar,et al. ENERGY FUNCTION ANALYSIS FOR POWER SYSTEM STABILITY , 1990 .
[26] C. Watkins. Learning from delayed rewards , 1989 .
[27] Warren D. Smith. Studies in computational geometry motivated by mesh generation , 1989 .
[28] Katsuhiko Ogata,et al. Discrete-time control systems , 1987 .
[29] P. Schweitzer,et al. Generalized polynomial approximations in Markovian decision processes , 1985 .
[30] John F. Sallee. The Middle-Cut Triangulations of the n-Cube , 1984 .
[31] P. L'Ecuyer,et al. Approximation and bounds in discrete event dynamic programming , 1983, The 23rd IEEE Conference on Decision and Control.
[32] Martin L. Puterman,et al. Action Elimination Procedures for Modified Policy Iteration Algorithms , 1982, Oper. Res..
[33] J. Hammersley. SIMULATION AND THE MONTE CARLO METHOD , 1982 .
[34] John F. Sallee. A triangulation of the n-cube , 1982, Discret. Math..
[35] Ward Whitt,et al. Approximations of Dynamic Programs, I , 1978, Math. Oper. Res..
[36] M. Puterman,et al. Modified Policy Iteration Algorithms for Discounted Markov Decision Problems , 1978 .
[37] D. Luenberger. Optimization by Vector Space Methods , 1968 .