暂无分享,去创建一个
Tom Schaul | Rémi Munos | David Silver | Matteo Hessel | André Barreto | Diana Borsa | Daniel J. Mankowitz | John Quan | Augustin Zídek | T. Schaul | D. Silver | R. Munos | Matteo Hessel | D. Mankowitz | Augustin Zídek | André Barreto | John Quan | Diana Borsa | David Silver
[1] Geoffrey E. Hinton,et al. Learning internal representations by error propagation , 1986 .
[2] C. Watkins. Learning from delayed rewards , 1989 .
[3] Peter Dayan,et al. Improving Generalization for Temporal Difference Learning: The Successor Representation , 1993, Neural Computation.
[4] Martin L. Puterman,et al. Markov Decision Processes: Discrete Stochastic Dynamic Programming , 1994 .
[5] Sebastian Thrun,et al. Is Learning The n-th Thing Any Easier Than Learning The First? , 1995, NIPS.
[6] John N. Tsitsiklis,et al. Neuro-Dynamic Programming , 1996, Encyclopedia of Machine Learning.
[7] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[8] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[9] Rich Caruana,et al. Multitask Learning , 1997, Machine-mediated learning.
[10] Jonathan Baxter,et al. A Model of Inductive Bias Learning , 2000, J. Artif. Intell. Res..
[11] Simon Haykin,et al. GradientBased Learning Applied to Document Recognition , 2001 .
[12] Peter Dayan,et al. Q-learning , 1992, Machine Learning.
[13] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[14] Peter Stone,et al. Transfer Learning for Reinforcement Learning Domains: A Survey , 2009, J. Mach. Learn. Res..
[15] Csaba Szepesvári,et al. Algorithms for Reinforcement Learning , 2010, Synthesis Lectures on Artificial Intelligence and Machine Learning.
[16] Alessandro Lazaric,et al. Transfer in Reinforcement Learning: A Framework and a Survey , 2012, Reinforcement Learning.
[17] Neil Burch,et al. Heads-up limit hold’em poker is solved , 2015, Science.
[18] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[19] Samuel Gershman,et al. Deep Successor Reinforcement Learning , 2016, ArXiv.
[20] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[21] Yuval Tassa,et al. Learning and Transfer of Modulated Locomotor Controllers , 2016, ArXiv.
[22] Tom Schaul,et al. FeUdal Networks for Hierarchical Reinforcement Learning , 2017, ICML.
[23] Doina Precup,et al. The Option-Critic Architecture , 2016, AAAI.
[24] Wolfram Burgard,et al. Deep reinforcement learning with successor features for navigation across similar environments , 2016, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[25] Honglak Lee,et al. Zero-Shot Task Generalization with Multi-Task Deep Reinforcement Learning , 2017, ICML.
[26] Razvan Pascanu,et al. Overcoming catastrophic forgetting in neural networks , 2016, Proceedings of the National Academy of Sciences.
[27] Sergey Levine,et al. Learning modular neural network policies for multi-task and multi-robot transfer , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).
[28] Yuval Tassa,et al. Emergence of Locomotion Behaviours in Rich Environments , 2017, ArXiv.
[29] Demis Hassabis,et al. Mastering the game of Go without human knowledge , 2017, Nature.
[30] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[31] Yee Whye Teh,et al. Distral: Robust multitask reinforcement learning , 2017, NIPS.
[32] Pieter Abbeel,et al. Policy transfer via modularity and reward guiding , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[33] Tom Schaul,et al. Successor Features for Transfer in Reinforcement Learning , 2016, NIPS.
[34] Pieter Abbeel,et al. Meta Learning Shared Hierarchies , 2017, ICLR.
[35] Shane Legg,et al. IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures , 2018, ICML.