Hierarchical Deep Reinforcement Learning: Integrating Temporal Abstraction and Intrinsic Motivation
暂无分享,去创建一个
Joshua B. Tenenbaum | Ardavan Saeedi | Karthik Narasimhan | Tejas D. Kulkarni | J. Tenenbaum | Karthik Narasimhan | Ardavan Saeedi | A. Saeedi
[1] Geoffrey E. Hinton,et al. Feudal Reinforcement Learning , 1992, NIPS.
[2] Peter Dayan,et al. Improving Generalization for Temporal Difference Learning: The Successor Representation , 1993, Neural Computation.
[3] Doina Precup,et al. Between MDPs and Semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning , 1999, Artif. Intell..
[4] Thomas G. Dietterich. Hierarchical Reinforcement Learning with the MAXQ Value Function Decomposition , 1999, J. Artif. Intell. Res..
[5] Sridhar Mahadevan,et al. Hierarchical Memory-Based Reinforcement Learning , 2000, NIPS.
[6] Andrew G. Barto,et al. Automatic Discovery of Subgoals in Reinforcement Learning using Diverse Density , 2001, ICML.
[7] Shie Mannor,et al. Q-Cut - Dynamic Discovery of Sub-goals in Reinforcement Learning , 2002, ECML.
[8] Carlos Guestrin,et al. Generalizing plans to new environments in relational MDPs , 2003, IJCAI 2003.
[9] Sridhar Mahadevan,et al. Recent Advances in Hierarchical Reinforcement Learning , 2003, Discret. Event Dyn. Syst..
[10] Shobha Venkataraman,et al. Efficient Solution Algorithms for Factored MDPs , 2003, J. Artif. Intell. Res..
[11] Manfred Huber,et al. Subgoal Discovery for Hierarchical Reinforcement Learning Using Learned Policies , 2003, FLAIRS.
[12] Sridhar Mahadevan,et al. Recent Advances in Hierarchical Reinforcement Learning , 2003, Discret. Event Dyn. Syst..
[13] Nuttapong Chentanez,et al. Intrinsically Motivated Reinforcement Learning , 2004, NIPS.
[14] Shie Mannor,et al. Dynamic abstraction in reinforcement learning via clustering , 2004, ICML.
[15] Alicia P. Wolfe,et al. Identifying useful subgoals in reinforcement learning by local graph partitioning , 2005, ICML.
[16] Katherine D. Kinzler,et al. Core knowledge. , 2007, Developmental science.
[17] International Foundation for Autonomous Agents and MultiAgent Systems ( IFAAMAS ) , 2007 .
[18] Pierre-Yves Oudeyer,et al. What is Intrinsic Motivation? A Typology of Computational Approaches , 2007, Frontiers Neurorobotics.
[19] Andre Cohen,et al. An object-oriented representation for efficient reinforcement learning , 2008, ICML '08.
[20] M. Botvinick,et al. Hierarchically organized behavior and its neural foundations: A reinforcement learning perspective , 2009, Cognition.
[21] Richard L. Lewis,et al. Where Do Rewards Come From , 2009 .
[22] Satinder P. Singh,et al. Linear options , 2010, AAMAS.
[23] Jürgen Schmidhuber,et al. Formal Theory of Creativity, Fun, and Intrinsic Motivation (1990–2010) , 2010, IEEE Transactions on Autonomous Mental Development.
[24] Richard L. Lewis,et al. Intrinsically Motivated Reinforcement Learning: An Evolutionary Perspective , 2010, IEEE Transactions on Autonomous Mental Development.
[25] Patrick M. Pilarski,et al. Horde: a scalable real-time architecture for learning knowledge from unsupervised sensorimotor interaction , 2011, AAMAS.
[26] Eduardo F. Morales,et al. An Introduction to Reinforcement Learning , 2011 .
[27] P. Cochat,et al. Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.
[28] Per B. Sederberg,et al. The Successor Representation and Temporal Context , 2012, Neural Computation.
[29] Andrea Lockerd Thomaz,et al. Object focused q-learning for autonomous agents , 2013, AAMAS.
[30] Pierre-Yves Oudeyer,et al. Active learning of inverse models with intrinsically motivated goal exploration in robots , 2013, Robotics Auton. Syst..
[31] Shalabh Bhatnagar,et al. Universal Option Models , 2014, NIPS.
[32] Jürgen Schmidhuber,et al. Evolving deep unsupervised convolutional networks for vision-based reinforcement learning , 2014, GECCO.
[33] Samuel Gershman,et al. Design Principles of the Hippocampal Cognitive Map , 2014, NIPS.
[34] Jürgen Leitner,et al. Curiosity driven reinforcement learning for motion planning on humanoids , 2014, Front. Neurorobot..
[35] Sergey Levine,et al. Incentivizing Exploration In Reinforcement Learning With Deep Predictive Models , 2015, ArXiv.
[36] Jürgen Schmidhuber,et al. Binding via Reconstruction Clustering , 2015, ArXiv.
[37] Tom Schaul,et al. Universal Value Function Approximators , 2015, ICML.
[38] Joshua B. Tenenbaum,et al. Deep Convolutional Inverse Graphics Network , 2015, NIPS.
[39] Kevin Murphy,et al. Efficient inference in occlusion-aware generative models of images , 2015, ArXiv.
[40] Alex Graves,et al. DRAW: A Recurrent Neural Network For Image Generation , 2015, ICML.
[41] Jitendra Malik,et al. Learning to segment moving objects in videos , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[42] Shakir Mohamed,et al. Variational Information Maximisation for Intrinsically Motivated Reinforcement Learning , 2015, NIPS.
[43] Regina Barzilay,et al. Language Understanding for Text-based Games using Deep Reinforcement Learning , 2015, EMNLP.
[44] Shane Legg,et al. Massively Parallel Methods for Deep Reinforcement Learning , 2015, ArXiv.
[45] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[46] Peter Stone,et al. Deep Recurrent Q-Learning for Partially Observable MDPs , 2015, AAAI Fall Symposia.
[47] Marc G. Bellemare,et al. The Arcade Learning Environment: An Evaluation Platform for General Agents (Extended Abstract) , 2012, IJCAI.
[48] Daan Wierstra,et al. One-Shot Generalization in Deep Generative Models , 2016, ICML.
[49] Joshua B. Tenenbaum,et al. Understanding Visual Concepts with Continuation Learning , 2016, ArXiv.
[50] Geoffrey E. Hinton,et al. Attend, Infer, Repeat: Fast Scene Understanding with Generative Models , 2016, NIPS.
[51] Marlos C. Machado,et al. Learning Purposeful Behaviour in the Absence of Rewards , 2016, ArXiv.
[52] Benjamin Van Roy,et al. Deep Exploration via Bootstrapped DQN , 2016, NIPS.
[53] Alex Graves,et al. Asynchronous Methods for Deep Reinforcement Learning , 2016, ICML.
[54] Joshua B. Tenenbaum,et al. Building machines that learn and think like people , 2016, Behavioral and Brain Sciences.
[55] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[56] Tom Schaul,et al. Prioritized Experience Replay , 2015, ICLR.