-
爱吃猫的鱼0于 2022年4月22日 04:50
[1] Pablo Samuel Castro,et al. Scalable methods for computing state similarity in deterministic Markov Decision Processes , 2019, AAAI.
[2] Sergey Levine,et al. Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynamics Models , 2018, NeurIPS.
[3] Nan Jiang,et al. Provably efficient RL with Rich Observations via Latent State Decoding , 2019, ICML.
[4] Doina Precup,et al. Metrics for Finite Markov Decision Processes , 2004, AAAI.
[5] Doina Precup,et al. Bisimulation Metrics are Optimal Value Functions , 2014, UAI.
[6] Marc G. Bellemare,et al. DeepMDP: Learning Continuous Latent Space Models for Representation Learning , 2019, ICML.
[7] Robert Givan,et al. Equivalence notions and model minimization in Markov decision processes , 2003, Artif. Intell..
[8] Sergey Levine,et al. Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor , 2018, ICML.
[9] Ali Razavi,et al. Data-Efficient Image Recognition with Contrastive Predictive Coding , 2019, ICML.
[10] Martin A. Riedmiller,et al. Deep auto-encoder neural networks in reinforcement learning , 2010, The 2010 International Joint Conference on Neural Networks (IJCNN).
[11] Martin A. Riedmiller,et al. Autonomous reinforcement learning on raw visual input data in a real world application , 2012, The 2012 International Joint Conference on Neural Networks (IJCNN).
[12] Kim G. Larsen,et al. Bisimulation through probabilistic testing (preliminary report) , 1989, POPL '89.
[13] Thomas J. Walsh,et al. Towards a Unified Theory of State Abstraction for MDPs , 2006, AI&M.
[14] Sergey Levine,et al. Efficient Off-Policy Meta-Reinforcement Learning via Probabilistic Context Variables , 2019, ICML.
[15] Doina Precup,et al. Bounding Performance Loss in Approximate MDP Homomorphisms , 2008, NIPS.
[16] James Worrell,et al. Towards Quantitative Verification of Probabilistic Transition Systems , 2001, ICALP.
[17] Thomas B. Schön,et al. From Pixels to Torques: Policy Learning with Deep Dynamical Models , 2015, ICML 2015.
[18] Martin A. Riedmiller,et al. Embed to Control: A Locally Linear Latent Dynamics Model for Control from Raw Images , 2015, NIPS.
[19] Piero Macaluso,et al. Deep Reinforcement Learning for Autonomous Systems , 2020 .
[20] Oliver Brock,et al. Learning state representations with robotic priors , 2015, Auton. Robots.
[21] Fabio Viola,et al. The Kinetics Human Action Video Dataset , 2017, ArXiv.
[22] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[23] Germán Ros,et al. CARLA: An Open Urban Driving Simulator , 2017, CoRL.
[24] Ruben Villegas,et al. Learning Latent Dynamics for Planning from Pixels , 2018, ICML.
[25] JonssonAnders,et al. Causal Graph Based Decomposition of Factored MDPs , 2006 .
[26] Pieter Abbeel,et al. CURL: Contrastive Unsupervised Representations for Reinforcement Learning , 2020, ICML.
[27] Sergey Levine,et al. Stochastic Latent Actor-Critic: Deep Reinforcement Learning with a Latent Variable Model , 2019, NeurIPS.
[28] Oriol Vinyals,et al. Representation Learning with Contrastive Predictive Coding , 2018, ArXiv.
[29] Bernhard Scholkopf. Causality for Machine Learning , 2019 .
[30] Yuval Tassa,et al. DeepMind Control Suite , 2018, ArXiv.
[31] Oriol Vinyals,et al. Representation Learning with Contrastive Predictive Coding , 2018, ArXiv.
[32] Geoffrey E. Hinton,et al. A Simple Framework for Contrastive Learning of Visual Representations , 2020, ICML.
[33] Doina Precup,et al. Bisimulation Metrics for Continuous Markov Decision Processes , 2011, SIAM J. Comput..