03 62 0 v 1 [ cs . L G ] 8 O ct 2 01 9 Receding Horizon Curiosity
暂无分享,去创建一个
[1] P. Silvia. Curiosity and Motivation , 2012 .
[2] Shie Mannor,et al. Bayesian Reinforcement Learning: A Survey , 2015, Found. Trends Mach. Learn..
[3] P. Cochat,et al. Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.
[4] David A. Cohn,et al. Active Learning with Statistical Models , 1996, NIPS.
[5] Leslie Pack Kaelbling,et al. Planning and Acting in Partially Observable Stochastic Domains , 1998, Artif. Intell..
[6] J. Stoer,et al. Introduction to Numerical Analysis , 2002 .
[7] Xavier Bombois,et al. Input design: from open-loop to control-oriented design , 2006 .
[8] Alexei A. Efros,et al. Curiosity-Driven Exploration by Self-Supervised Prediction , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[9] Sergey Levine,et al. Incentivizing Exploration In Reinforcement Learning With Deep Predictive Models , 2015, ArXiv.
[10] Filip De Turck,et al. VIME: Variational Information Maximizing Exploration , 2016, NIPS.
[11] Leslie Pack Kaelbling,et al. Belief space planning assuming maximum likelihood observations , 2010, Robotics: Science and Systems.
[12] Radford M. Neal. Pattern Recognition and Machine Learning , 2007, Technometrics.
[13] Andrew Y. Ng,et al. Near-Bayesian exploration in polynomial time , 2009, ICML '09.
[14] Amos J. Storkey,et al. Exploration by Random Network Distillation , 2018, ICLR.
[15] Pieter Abbeel,et al. Scaling up Gaussian Belief Space Planning Through Covariance-Free Trajectory Optimization and Automatic Differentiation , 2014, WAFR.
[16] T. L. Lai Andherbertrobbins. Asymptotically Efficient Adaptive Allocation Rules , 2022 .
[17] Yi Sun,et al. Planning to Be Surprised: Optimal Bayesian Exploration in Dynamic Environments , 2011, AGI.
[18] Carl E. Rasmussen,et al. PILCO: A Model-Based and Data-Efficient Approach to Policy Search , 2011, ICML.
[19] Stewart W. Wilson,et al. A Possibility for Implementing Curiosity and Boredom in Model-Building Neural Controllers , 1991 .
[20] Wojciech Jaskowski,et al. Model-Based Active Exploration , 2018, ICML.
[21] Raman K. Mehra,et al. Optimal input signals for parameter estimation in dynamic systems--Survey and new results , 1974 .
[22] Tom Schaul,et al. Unifying Count-Based Exploration and Intrinsic Motivation , 2016, NIPS.
[23] Moritz Diehl,et al. CasADi: a software framework for nonlinear optimization and optimal control , 2018, Mathematical Programming Computation.
[24] Benjamin Recht,et al. Random Features for Large-Scale Kernel Machines , 2007, NIPS.
[25] Sergey Levine,et al. Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor , 2018, ICML.
[26] Burr Settles,et al. Active Learning Literature Survey , 2009 .
[27] Kian Hsiang Low,et al. Gaussian Process Planning with Lipschitz Continuous Reward Functions: Towards Unifying Bayesian Optimization, Active Learning, and Beyond , 2015, AAAI.
[28] K. Chaloner,et al. Bayesian Experimental Design: A Review , 1995 .
[29] Jürgen Schmidhuber,et al. Curious model-building control systems , 1991, [Proceedings] 1991 IEEE International Joint Conference on Neural Networks.
[30] E. Deci,et al. Intrinsic and Extrinsic Motivations: Classic Definitions and New Directions. , 2000, Contemporary educational psychology.
[31] A. A. Feldbaum,et al. DUAL CONTROL THEORY, IV , 1961 .
[32] Alexei A. Efros,et al. Large-Scale Study of Curiosity-Driven Learning , 2018, ICLR.
[33] Sham M. Kakade,et al. Towards Generalization and Simplicity in Continuous Control , 2017, NIPS.
[34] John N. Tsitsiklis,et al. The Complexity of Markov Decision Processes , 1987, Math. Oper. Res..
[35] Philipp Hennig,et al. Dual Control for Approximate Bayesian Reinforcement Learning , 2015, J. Mach. Learn. Res..
[36] Michael Kearns,et al. Near-Optimal Reinforcement Learning in Polynomial Time , 2002, Machine Learning.
[37] X. Huan,et al. Sequential Bayesian optimal experimental design via approximate dynamic programming , 2016, 1604.08320.
[38] Martin B. Zarrop,et al. Optimal experiment design for dynamic system identification , 1977 .
[39] P. Alam. ‘A’ , 2021, Composites Engineering: An A–Z Guide.
[40] D. Lindley. On a Measure of the Information Provided by an Experiment , 1956 .