Optimizing Loss Functions Through Multivariate Taylor Polynomial Parameterization
暂无分享,去创建一个
[1] Yaochu Jin,et al. Surrogate-assisted evolutionary computation: Recent advances and future challenges , 2011, Swarm Evol. Comput..
[2] P. Graves-Morris. The numerical calculation of Padé approximants , 1979 .
[3] Nitish Srivastava,et al. Improving neural networks by preventing co-adaptation of feature detectors , 2012, ArXiv.
[4] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[5] Nikolaus Hansen,et al. Evaluating the CMA Evolution Strategy on Multimodal Test Functions , 2004, PPSN.
[6] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[7] Frank Hutter,et al. Neural Architecture Search: A Survey , 2018, J. Mach. Learn. Res..
[8] Yann Dauphin,et al. Empirical Analysis of the Hessian of Over-Parametrized Neural Networks , 2017, ICLR.
[9] D. E. Roberts,et al. Calculation of Canterbury approximants , 1984 .
[10] Jorge Nocedal,et al. On Large-Batch Training for Deep Learning: Generalization Gap and Sharp Minima , 2016, ICLR.
[11] Hao Li,et al. Visualizing the Loss Landscape of Neural Nets , 2017, NeurIPS.
[12] P. J. Huber. Robust Estimation of a Location Parameter , 1964 .
[13] Risto Miikkulainen,et al. Improved Training Speed, Accuracy, and Data Utilization Through Loss Function Optimization , 2019, 2020 IEEE Congress on Evolutionary Computation (CEC).
[14] Nikolaus Hansen,et al. Adapting arbitrary normal mutation distributions in evolution strategies: the covariance matrix adaptation , 1996, Proceedings of IEEE International Conference on Evolutionary Computation.
[15] Bogdan Gabrys,et al. Metalearning: a survey of trends and technologies , 2013, Artificial Intelligence Review.
[16] J. Chisholm. Rational approximants defined from double power series , 1973 .
[17] Elliot Meyerson,et al. Evolving Deep Neural Networks , 2017, Artificial Intelligence in the Age of Neural Networks and Brain Computing.
[18] Risto Miikkulainen,et al. Faster Training by Selecting Samples Using Embeddings , 2019, 2019 International Joint Conference on Neural Networks (IJCNN).
[19] Andrew Y. Ng,et al. Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .
[20] Geoffrey E. Hinton,et al. Learning internal representations by error propagation , 1986 .
[21] John J. Grefenstette,et al. Genetic Search with Approximate Function Evaluation , 1985, ICGA.
[22] Graham W. Taylor,et al. Improved Regularization of Convolutional Neural Networks with Cutout , 2017, ArXiv.
[23] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[24] Quoc V. Le,et al. AutoML-Zero: Evolving Machine Learning Algorithms From Scratch , 2020, ICML.
[25] Kristen Grauman,et al. 2.5D Visual Sound , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[26] Geoffrey E. Hinton,et al. Visualizing Data using t-SNE , 2008 .
[27] Yuan Yu,et al. TensorFlow: A system for large-scale machine learning , 2016, OSDI.
[28] Stefano Soatto,et al. Time Matters in Regularizing Deep Networks: Weight Decay and Data Augmentation Affect Early Learning Dynamics, Matter Little Near Convergence , 2019, NeurIPS.
[29] Risto Miikkulainen,et al. Evolutionary optimization of deep learning activation functions , 2020, GECCO.
[30] Nikolaus Hansen,et al. Completely Derandomized Self-Adaptation in Evolution Strategies , 2001, Evolutionary Computation.
[31] Yike Guo,et al. Semantic Image Synthesis via Adversarial Learning , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[32] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[33] Jian Sun,et al. Identity Mappings in Deep Residual Networks , 2016, ECCV.