Domain Randomization for Simulation-Based Policy Optimization with Transferability Assessment
暂无分享,去创建一个
Jan Peters | Michael Gienger | Fabio Muratore | Felix Treede | Jan Peters | M. Gienger | Fabio Muratore | Felix Treede
[1] Benjamin F. Hobbs,et al. Is optimization optimistically biased , 1989 .
[2] Inman Harvey,et al. Noise and the Reality Gap: The Use of Simulation in Evolutionary Robotics , 1995, ECAL.
[3] B. Efron,et al. Bootstrap confidence intervals , 1996 .
[4] Nick Jakobi,et al. Evolutionary Robotics and the Radical Envelope-of-Noise Hypothesis , 1997, Adapt. Behav..
[5] David P. Morton,et al. Monte Carlo bounding techniques for determining solution quality in stochastic programs , 1999, Oper. Res. Lett..
[6] Raghu Pasupathy,et al. Retrospective approximation algorithms for the multidimensional stochastic root-finding problem , 2004, Proceedings of the 2004 Winter Simulation Conference, 2004..
[7] David P. Morton,et al. Assessing solution quality in stochastic programs , 2006, Algorithms for Optimization with Incomplete Information.
[8] Philippe L. Toint,et al. Convergence theory for nonconvex stochastic programming with an application to mixed logit , 2006, Math. Program..
[9] Hod Lipson,et al. Resilient Machines Through Continuous Self-Modeling , 2006, Science.
[10] David Hinkley,et al. Bootstrap Methods: Another Look at the Jackknife , 2008 .
[11] Yasemin Altun,et al. Relative Entropy Policy Search , 2010 .
[12] David J. Fleet,et al. Optimizing walking controllers for uncertain inputs and environments , 2010, SIGGRAPH 2010.
[13] P. Cochat,et al. Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.
[14] Stéphane Doncieux,et al. The Transferability Approach: Crossing the Reality Gap in Evolutionary Robotics , 2013, IEEE Transactions on Evolutionary Computation.
[15] Emanuel Todorov,et al. Ensemble-CIO: Full-body dynamic motion planning that transfers to physical humanoids , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[16] R. Pasupathy,et al. A Guide to Sample Average Approximation , 2015 .
[17] Sergey Levine,et al. Trust Region Policy Optimization , 2015, ICML.
[18] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[19] Yuval Tassa,et al. Continuous control with deep reinforcement learning , 2015, ICLR.
[20] Stephen James,et al. 3D Simulation for Robot Arm Control with Deep Q-Learning , 2016, ArXiv.
[21] Silvio Savarese,et al. Adversarially Robust Policy Learning: Active construction of physically-plausible perturbations , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[22] Razvan Pascanu,et al. Sim-to-Real Robot Learning from Pixels with Progressive Nets , 2016, CoRL.
[23] Wojciech Zaremba,et al. Domain randomization for transferring deep neural networks from simulation to the real world , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[24] Balaraman Ravindran,et al. EPOpt: Learning Robust Neural Network Policies Using Model Ensembles , 2016, ICLR.
[25] Abhinav Gupta,et al. Robust Adversarial Reinforcement Learning , 2017, ICML.
[26] Alec Radford,et al. Proximal Policy Optimization Algorithms , 2017, ArXiv.
[27] Rika Antonova,et al. Unlocking the Potential of Simulators: Design with RL in Mind , 2017, ArXiv.
[28] Sergey Levine,et al. (CAD)$^2$RL: Real Single-Image Flight without a Single Real Image , 2016, Robotics: Science and Systems.
[29] Marcin Andrychowicz,et al. Sim-to-Real Transfer of Robotic Control with Dynamics Randomization , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[30] Pieter Abbeel,et al. Model-Ensemble Trust-Region Policy Optimization , 2018, ICLR.
[31] Marcin Andrychowicz,et al. Asymmetric Actor Critic for Image-Based Robot Learning , 2017, Robotics: Science and Systems.