暂无分享,去创建一个
[1] Mohammad Shahidehpour,et al. Security-Constrained Generation and Transmission Outage Scheduling With Uncertainties , 2010, IEEE Transactions on Power Systems.
[2] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[3] V. Lesser,et al. Accelerating Multi-agent Reinforcement Learning with Dynamic Co-learning , 2015 .
[4] Shie Mannor,et al. A Deep Hierarchical Approach to Lifelong Learning in Minecraft , 2016, AAAI.
[5] Masood Parvania,et al. A Two-Stage Framework for Power Transformer Asset Maintenance Management—Part I: Models and Formulations , 2013, IEEE Transactions on Power Systems.
[6] Leslie Pack Kaelbling,et al. Effective reinforcement learning for mobile robots , 2002, Proceedings 2002 IEEE International Conference on Robotics and Automation (Cat. No.02CH37292).
[7] Timothy A. Mann,et al. Iterative Hierarchical Optimization for Misspecified Problems , 2016 .
[8] Masood Parvania,et al. A two-stage framework for power transformer asset maintenance management—Part I: Models and formulations , 2013, 2013 IEEE Power & Energy Society General Meeting.
[9] Alborz Geramifard,et al. A Tutorial on Linear Function Approximators for Dynamic Programming and Reinforcement Learning , 2013, Found. Trends Mach. Learn..
[10] Shie Mannor,et al. Scaling Up Approximate Value Iteration with Options: Better Policies with Fewer Iterations , 2014, ICML.
[11] Brett Browning,et al. A survey of robot learning from demonstration , 2009, Robotics Auton. Syst..
[12] George Konidaris,et al. Value Function Approximation in Reinforcement Learning Using the Fourier Basis , 2011, AAAI.
[13] Andrew G. Barto,et al. Automatic Discovery of Subgoals in Reinforcement Learning using Diverse Density , 2001, ICML.
[14] Nicholas Roy,et al. Efficient Planning under Uncertainty with Macro-actions , 2014, J. Artif. Intell. Res..
[15] Huiyu Zhou,et al. Object tracking using SIFT features and mean shift , 2009, Comput. Vis. Image Underst..
[16] Michael I. Jordan,et al. Reinforcement Learning with Soft State Aggregation , 1994, NIPS.
[17] Andrew G. Barto,et al. Skill Discovery in Continuous Reinforcement Learning Domains using Skill Chaining , 2009, NIPS.
[18] Milos Hauskrecht,et al. Hierarchical Solution of Markov Decision Processes using Macro-actions , 1998, UAI.
[19] Pieter Abbeel,et al. Exploration and apprenticeship learning in reinforcement learning , 2005, ICML.
[20] Doina Precup,et al. Between MDPs and Semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning , 1999, Artif. Intell..
[21] N. Roy,et al. A Bayesian Approach to Finding Compact Representations for Reinforcement Learning , 2012 .
[22] Wilco Moerman,et al. Hierarchical Reinforcement Learning: Assignment of Behaviours to Subpolicies by Self-Organization , 2009 .
[23] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[24] Shalabh Bhatnagar,et al. Natural actor-critic algorithms , 2009, Autom..
[25] Shie Mannor,et al. Time-regularized interrupting options , 2014, ICML 2014.
[26] Yair Weiss,et al. Learning object detection from a small number of examples: the importance of good features , 2004, Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2004. CVPR 2004..
[27] Satinder P. Singh,et al. Linear options , 2010, AAMAS.
[28] Lihong Li,et al. PAC-inspired Option Discovery in Lifelong Reinforcement Learning , 2014, ICML.
[29] Richard S. Sutton,et al. Generalization in Reinforcement Learning: Successful Examples Using Sparse Coarse Coding , 1995, NIPS.
[30] Justin A. Boyan,et al. Technical Update: Least-Squares Temporal Difference Learning , 2002, Machine Learning.