An analysis of training and generalization errors in shallow and deep networks

[1]  Jong Chul Ye,et al.  ${k}$ -Space Deep Learning for Accelerated MRI , 2020, IEEE Transactions on Medical Imaging.

[2]  H. N. Mhaskar,et al.  Function approximation with zonal function networks with activation functions analogous to the rectified linear unit functions , 2017, J. Complex..

[3]  Tomaso A. Poggio,et al.  Theory IIIb: Generalization in Deep Networks , 2018, ArXiv.

[4]  Mikhail Belkin,et al.  Overfitting or perfect fitting? Risk bounds for classification and regression rules that interpolate , 2018, NeurIPS.

[5]  Mikhail Belkin,et al.  To understand deep learning we need to understand kernel learning , 2018, ICML.

[6]  Lorenzo Rosasco,et al.  Theory of Deep Learning III: explaining the non-overfitting puzzle , 2017, ArXiv.

[7]  T. Poggio,et al.  Theory of Deep Learning III : the non-overfitting puzzle , 2018 .

[8]  Hrushikesh N. Mhaskar,et al.  Function approximation with ReLU-like zonal function networks , 2017, ArXiv.

[9]  Hrushikesh N. Mhaskar,et al.  A Deep Learning Approach to Diabetic Blood Glucose Prediction , 2017, Front. Appl. Math. Stat..

[10]  Nathan Srebro,et al.  Exploring Generalization in Deep Learning , 2017, NIPS.

[11]  Noah Golowich,et al.  Musings on Deep Learning: Properties of SGD , 2017 .

[12]  Samy Bengio,et al.  Understanding deep learning requires rethinking generalization , 2016, ICLR.

[13]  T. Poggio,et al.  Deep vs. shallow networks : An approximation theory perspective , 2016, ArXiv.

[14]  Guillermo Sapiro,et al.  Margin Preservation of Deep Neural Networks , 2016, ArXiv.

[15]  Yoram Singer,et al.  Train faster, generalize better: Stability of stochastic gradient descent , 2015, ICML.

[16]  Geoffrey E. Hinton,et al.  Deep Learning , 2015, Nature.

[17]  Hrushikesh Narhar Mhaskar,et al.  Minimum Sobolev norm interpolation with trigonometric polynomials on the torus , 2013, J. Comput. Phys..

[18]  H. Mhaskar,et al.  Applications of classical approximation theory to periodic basis function networks and computational harmonic analysis , 2013 .

[19]  Shivkumar Chandrasekaran,et al.  Minimum Sobolev Norm schemes and applications in image processing , 2010, Electronic Imaging.

[20]  H. N. Mhaskar,et al.  Eignets for function approximation on manifolds , 2009, ArXiv.

[21]  Quoc Thong Le Gia,et al.  Localized Linear Polynomial Operators and Quadrature Formulas on the Sphere , 2008, SIAM J. Numer. Anal..

[22]  F. J. Narcowich,et al.  Approximation with interpolatory constraints , 2001 .

[23]  H. Mhaskar Approximation Theory and Neural Networks , 2001 .

[24]  Allan Pinkus,et al.  Approximation theory of the MLP model in neural networks , 1999, Acta Numerica.

[25]  H. N. Mhaskar,et al.  Neural Networks for Optimal Approximation of Smooth and Analytic Functions , 1996, Neural Computation.

[26]  C. Micchelli,et al.  Degree of Approximation by Neural and Translation Networks with a Single Hidden Layer , 1995 .

[27]  Charles A. Micchelli,et al.  Dimension-independent bounds on the degree of approximation by neural networks , 1994, IBM J. Res. Dev..

[28]  A. Timan Theory of Approximation of Functions of a Real Variable , 1994 .

[29]  G. Golub Matrix computations , 1983 .

[30]  E. Stein Singular Integrals and Differentiability Properties of Functions (PMS-30), Volume 30 , 1971 .

[31]  P. Heywood Trigonometric Series , 1968, Nature.

[32]  I. P. Natanson Constructive function theory , 1964 .

[33]  J. Czipszer,et al.  Sur l'approximation d'une fonction périodique et de ses dérivées successives par un polynome trigono-métrique et par ses dérivées successives , 1958 .

[34]  P. Erdös On Some Convergence Properties of the Interpolation Polynomials , 1943 .