Compositional Planning Using Optimal Option Models
暂无分享,去创建一个
[1] Saul Amarel,et al. On representations of problems of reasoning about actions , 1968 .
[2] Earl David Sacerdoti,et al. A Structure for Plans and Behavior , 1977 .
[3] R. Korf. Learning to solve problems by searching for macro-operators , 1983 .
[4] Richard S. Sutton,et al. TD Models: Modeling the World at a Mixture of Time Scales , 1995, ICML.
[5] Stuart J. Russell,et al. Reinforcement Learning with Hierarchies of Machines , 1997, NIPS.
[6] Doina Precup,et al. Theoretical Results on Reinforcement Learning with Temporally Abstract Options , 1998, ECML.
[7] Richard S. Sutton,et al. Introduction to Reinforcement Learning , 1998 .
[8] Doina Precup,et al. Between MDPs and Semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning , 1999, Artif. Intell..
[9] Thomas G. Dietterich. Hierarchical Reinforcement Learning with the MAXQ Value Function Decomposition , 1999, J. Artif. Intell. Res..
[10] David Andre,et al. State abstraction for programmable reinforcement learning agents , 2002, AAAI/IAAI.
[11] Allen Newell,et al. Chunking in Soar: The anatomy of a general learning mechanism , 1985, Machine Learning.
[12] Sean R Eddy,et al. What is dynamic programming? , 2004, Nature Biotechnology.
[13] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[14] Andrew G. Barto,et al. Efficient skill learning using abstraction selection , 2009, IJCAI 2009.
[15] Anders Jonsson. The Role of Macros in Tractable Planning , 2009, J. Artif. Intell. Res..