Approximate Value Iteration with Temporally Extended Actions (Extended Abstract)
暂无分享,去创建一个
[1] Andrew G. Barto,et al. Using relative novelty to identify useful temporal abstractions in reinforcement learning , 2004, ICML.
[2] Doina Precup,et al. Between MDPs and Semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning , 1999, Artif. Intell..
[3] H. Scarf. THE OPTIMALITY OF (S,S) POLICIES IN THE DYNAMIC INVENTORY PROBLEM , 1959 .
[4] John N. Tsitsiklis,et al. Neuro-Dynamic Programming , 1996, Encyclopedia of Machine Learning.
[5] Zoubin Ghahramani,et al. Proceedings of the 24th international conference on Machine learning , 2007, ICML 2007.
[6] Luc De Raedt,et al. Proceedings of the 22nd international conference on Machine learning , 2005 .
[7] Suresh P. Sethi,et al. Optimality of (s, S) Policies in Inventory Models with Markovian Demand , 1995, Oper. Res..
[8] Shie Mannor,et al. Scaling Up Approximate Value Iteration with Options: Better Policies with Fewer Iterations , 2014, ICML.
[9] David Silver,et al. Compositional Planning Using Optimal Option Models , 2012, ICML.
[10] Craig Boutilier,et al. Proceedings of the 16th Conference on Uncertainty in Artificial Intelligence , 2000 .
[11] J. A. Salvato. John wiley & sons. , 1994, Environmental science & technology.
[12] Doina Precup,et al. Learning Options in Reinforcement Learning , 2002, SARA.
[13] Robert Givan,et al. FF-Replan: A Baseline for Probabilistic Planning , 2007, ICAPS.
[14] Alicia P. Wolfe,et al. Identifying useful subgoals in reinforcement learning by local graph partitioning , 2005, ICML.
[15] Andrew G. Barto,et al. Automatic Discovery of Subgoals in Reinforcement Learning using Diverse Density , 2001, ICML.
[16] Shie Mannor,et al. Approximate Value Iteration with Temporally Extended Actions , 2015, J. Artif. Intell. Res..
[17] Leslie Pack Kaelbling,et al. On the Complexity of Solving Markov Decision Problems , 1995, UAI.