Reinforcement learning in partially observable mobile robot domains using unsupervised event extraction
暂无分享,去创建一个
[1] Jürgen Schmidhuber,et al. Networks adjusting networks , 1990 .
[2] Rodney A. Brooks,et al. Intelligence Without Reason , 1991, IJCAI.
[3] Jürgen Schmidhuber,et al. Learning Complex, Extended Sequences Using the Principle of History Compression , 1992, Neural Computation.
[4] Tom M. Mitchell,et al. Reinforcement learning with hidden states , 1993 .
[5] Mance E. Harmon,et al. Multi-Agent Residual Advantage Learning with General Function Approximation. , 1996 .
[6] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[7] Jürgen Schmidhuber,et al. HQ-Learning , 1997, Adapt. Behav..
[8] Richard S. Sutton,et al. Introduction to Reinforcement Learning , 1998 .
[9] Doina Precup,et al. Between MDPs and Semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning , 1999, Artif. Intell..
[10] Fernando Fernández,et al. VQQL. Applying Vector Quantization to Reinforcement Learning , 1999, RoboCup.
[11] Sridhar Mahadevan,et al. Hierarchical Memory-Based Reinforcement Learning , 2000, NIPS.
[12] Bram Bakker,et al. Reinforcement Learning with Long Short-Term Memory , 2001, NIPS.
[13] Henrik Jacobsson,et al. Mobile Robot Learning of Delayed Response Tasks through Event Extraction: A Solution to the Road Sign Problem and Beyond , 2001, IJCAI.
[14] Chris A. Czarnecki,et al. Embedding Connectionist Autonomous Agents in Time: The ‘Road Sign Problem’ , 2000, Neural Processing Letters.