暂无分享,去创建一个
[1] Quoc V. Le,et al. On optimization methods for deep learning , 2011, ICML.
[2] Klaus-Robert Müller,et al. Efficient BackProp , 2012, Neural Networks: Tricks of the Trade.
[3] Yann LeCun,et al. Universality in halting time and its applications in optimization , 2015, ArXiv.
[4] Yann LeCun,et al. Explorations on high dimensional landscapes , 2014, ICLR.
[5] Yoram Singer,et al. Train faster, generalize better: Stability of stochastic gradient descent , 2015, ICML.
[6] Michael I. Jordan,et al. Gradient Descent Converges to Minimizers , 2016, ArXiv.
[7] Michael I. Jordan,et al. Gradient Descent Only Converges to Minimizers , 2016, COLT.