暂无分享,去创建一个
Lorenzo Rosasco | Tomaso A. Poggio | Kenji Kawaguchi | Brando Miranda | Qianli Liao | Xavier Boix | Jack Hidary | Hrushikesh Mhaskar | T. Poggio | H. Mhaskar | L. Rosasco | X. Boix | B. Miranda | Q. Liao | J. Hidary | Kenji Kawaguchi
[1] J. Czipszer,et al. Sur l'approximation d'une fonction périodique et de ses dérivées successives par un polynome trigono-métrique et par ses dérivées successives , 1958 .
[2] J. Carr. Applications of Centre Manifold Theory , 1981 .
[3] R. Sverdlove. Inverse problems for dynamical systems , 1981 .
[4] Kurt Hornik,et al. Neural networks and principal component analysis: Learning from examples without local minima , 1989, Neural Networks.
[5] Hrushikesh Narhar Mhaskar,et al. Approximation properties of a multilayered feedforward artificial neural network , 1993, Adv. Comput. Math..
[6] Peter L. Bartlett,et al. Neural Network Learning - Theoretical Foundations , 1999 .
[7] B. Aulbach,et al. The Hartman-Grobman theorem for Carathéodory-type differential equations in Banach spaces , 2000 .
[8] 김희라. Waiting for Godot에 나타난 희망의 구조 , 2003 .
[9] Y. Yao,et al. On Early Stopping in Gradient Descent Learning , 2007 .
[10] Shie Mannor,et al. Robustness and Regularization of Support Vector Machines , 2008, J. Mach. Learn. Res..
[11] Bum Il Hong,et al. Simultaneous Approximation Algorithm Using a Feedforward Neural Network with a Single Hidden Layer , 2009 .
[12] Surya Ganguli,et al. Exact solutions to the nonlinear dynamics of learning in deep linear neural networks , 2013, ICLR.
[13] Lorenzo Rosasco,et al. Learning with Incremental Iterative Regularization , 2014, NIPS.
[14] Kenji Kawaguchi,et al. Deep Learning without Poor Local Minima , 2016, NIPS.
[15] Yoram Singer,et al. Train faster, generalize better: Stability of stochastic gradient descent , 2015, ICML.
[16] Yann LeCun,et al. Singularity of the Hessian in Deep Learning , 2016, ArXiv.
[17] T. Poggio,et al. Deep vs. shallow networks : An approximation theory perspective , 2016, ArXiv.
[18] Lorenzo Rosasco,et al. Optimal Rates for Multi-pass Stochastic Gradient Methods , 2016, J. Mach. Learn. Res..
[19] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[20] Razvan Pascanu,et al. Sharp Minima Can Generalize For Deep Nets , 2017, ICML.
[21] Tomaso A. Poggio,et al. Theory II: Landscape of the Empirical Risk in Deep Learning , 2017, ArXiv.
[22] Tengyu Ma,et al. Identity Matters in Deep Learning , 2016, ICLR.
[23] Nathan Srebro,et al. Exploring Generalization in Deep Learning , 2017, NIPS.
[24] Guillermo Sapiro,et al. Robust Large Margin Deep Neural Networks , 2017, IEEE Transactions on Signal Processing.
[25] Lorenzo Rosasco,et al. Why and when can deep-but not shallow-networks avoid the curse of dimensionality: A review , 2016, International Journal of Automation and Computing.
[26] Nathan Srebro,et al. The Implicit Bias of Gradient Descent on Separable Data , 2017, J. Mach. Learn. Res..
[27] Tomaso A. Poggio,et al. Theory of Deep Learning IIb: Optimization Properties of SGD , 2018, ArXiv.
[28] Tomaso A. Poggio,et al. Fisher-Rao Metric, Geometry, and Complexity of Neural Networks , 2017, AISTATS.