暂无分享,去创建一个
Tom Schaul | Erich Elsen | Karen Simonyan | Karel Lenc | T. Schaul | K. Simonyan | Karel Lenc | Erich Elsen
[1] Sergio Gomez Colmenarejo,et al. TF-Replicator: Distributed Machine Learning for Researchers , 2019, ArXiv.
[2] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[3] R. J. Williams,et al. Simple Statistical Gradient-Following Algorithms for Connectionist Reinforcement Learning , 2004, Machine Learning.
[4] Kenneth O. Stanley,et al. On the Relationship Between the OpenAI Evolution Strategy and Stochastic Gradient Descent , 2017, ArXiv.
[5] Mark J. Harris,et al. Parallel Prefix Sum (Scan) with CUDA , 2011 .
[6] Benjamin Recht,et al. Simple random search provides a competitive approach to reinforcement learning , 2018, ArXiv.
[7] Kenneth O. Stanley,et al. ES is more than just a traditional finite-difference approximator , 2017, GECCO.
[8] Song Han,et al. EIE: Efficient Inference Engine on Compressed Deep Neural Network , 2016, 2016 ACM/IEEE 43rd Annual International Symposium on Computer Architecture (ISCA).
[9] Nikko Strom,et al. Sparse connection and pruning in large dynamic artificial neural networks. , 1997 .
[10] Xi Chen,et al. Evolution Strategies as a Scalable Alternative to Reinforcement Learning , 2017, ArXiv.
[11] Nathan Bell,et al. Thrust: A Productivity-Oriented Library for CUDA , 2012 .
[12] Suyog Gupta,et al. To prune, or not to prune: exploring the efficacy of pruning for model compression , 2017, ICLR.
[13] Erich Elsen,et al. Exploring Sparsity in Recurrent Neural Networks , 2017, ICLR.
[14] Nikolaus Hansen,et al. Completely Derandomized Self-Adaptation in Evolution Strategies , 2001, Evolutionary Computation.
[15] Tom Schaul,et al. High dimensions and heavy tails for natural evolution strategies , 2011, GECCO '11.
[16] Yee Whye Teh,et al. The Concrete Distribution: A Continuous Relaxation of Discrete Random Variables , 2016, ICLR.
[17] Xin Yao,et al. Evolving artificial neural networks , 1999, Proc. IEEE.
[18] Anne Auger,et al. A Comparative Study of Large-Scale Variants of CMA-ES , 2018, PPSN.
[19] Guy E. Blelloch,et al. Prefix sums and their applications , 1990 .
[20] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[21] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.
[22] Erich Elsen,et al. Efficient Neural Audio Synthesis , 2018, ICML.
[23] Tom Schaul,et al. Natural Evolution Strategies , 2008, 2008 IEEE Congress on Evolutionary Computation (IEEE World Congress on Computational Intelligence).
[24] Yann LeCun,et al. Optimal Brain Damage , 1989, NIPS.
[25] Jascha Sohl-Dickstein,et al. Guided evolutionary strategies: escaping the curse of dimensionality in random search , 2018, ArXiv.
[26] Max Welling,et al. Learning Sparse Neural Networks through L0 Regularization , 2017, ICLR.
[27] Raymond Ros,et al. A Simple Modification in CMA-ES Achieving Linear Time and Space Complexity , 2008, PPSN.
[28] Ben Poole,et al. Categorical Reparameterization with Gumbel-Softmax , 2016, ICLR.
[29] Christian Igel,et al. Neuroevolution for reinforcement learning using evolution strategies , 2003, The 2003 Congress on Evolutionary Computation, 2003. CEC '03..
[30] Ingo Rechenberg,et al. Evolutionsstrategie : Optimierung technischer Systeme nach Prinzipien der biologischen Evolution , 1973 .
[31] Dmitry P. Vetrov,et al. Variational Dropout Sparsifies Deep Neural Networks , 2017, ICML.
[32] David Kappel,et al. Deep Rewiring: Training very sparse deep networks , 2017, ICLR.
[33] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[34] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[35] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[36] Lucas Theis,et al. Faster gaze prediction with dense networks and Fisher pruning , 2018, ArXiv.
[37] François Chollet,et al. Xception: Deep Learning with Depthwise Separable Convolutions , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[38] Erich Elsen,et al. The State of Sparsity in Deep Neural Networks , 2019, ArXiv.
[39] Martin Mandischer. A comparison of evolution strategies and backpropagation for neural network training , 2002, Neurocomputing.
[40] Alok Aggarwal,et al. Regularized Evolution for Image Classifier Architecture Search , 2018, AAAI.
[41] Quoc V. Le,et al. Don't Decay the Learning Rate, Increase the Batch Size , 2017, ICLR.
[42] Yann LeCun,et al. Regularization of Neural Networks using DropConnect , 2013, ICML.
[43] Bernhard Sendhoff,et al. Genesis of Organic Computing Systems: Coupling Evolution and Learning , 2008, Organic Computing.