A Universal Analysis of Large-Scale Regularized Least Squares Solutions
暂无分享,去创建一个
[1] Gitta Kutyniok,et al. 1 . 2 Sparsity : A Reasonable Assumption ? , 2012 .
[2] Mihailo Stojnic,et al. A framework to characterize performance of LASSO algorithms , 2013, ArXiv.
[3] Wen Gao,et al. Efficient Generalized Fused Lasso and its Application to the Diagnosis of Alzheimer's Disease , 2014, AAAI.
[4] E. Candès. The restricted isometry property and its implications for compressed sensing , 2008 .
[5] Richard G. Baraniuk,et al. Compressive Sensing , 2008, Computer Vision, A Reference Guide.
[6] Andrea Montanari,et al. Applications of the Lindeberg Principle in Communications and Statistical Learning , 2010, IEEE Transactions on Information Theory.
[7] David L Donoho,et al. Compressed sensing , 2006, IEEE Transactions on Information Theory.
[8] J. Lindeberg. Eine neue Herleitung des Exponentialgesetzes in der Wahrscheinlichkeitsrechnung , 1922 .
[9] Christos Thrampoulidis,et al. Asymptotically exact error analysis for the generalized equation-LASSO , 2015, 2015 IEEE International Symposium on Information Theory (ISIT).
[10] Christos Thrampoulidis,et al. LASSO with Non-linear Measurements is Equivalent to One With Linear Measurements , 2015, NIPS.
[11] B. Hassibi,et al. Precise Error Analysis of the $\ell_2$-LASSO , 2015, 1502.04977.
[12] Marc Lelarge,et al. Fundamental limits of symmetric low-rank matrix estimation , 2016, Probability Theory and Related Fields.
[13] Emmanuel J. Candès,et al. Decoding by linear programming , 2005, IEEE Transactions on Information Theory.
[14] E. Candès,et al. Stable signal recovery from incomplete and inaccurate measurements , 2005, math/0503066.
[15] Jieping Ye,et al. Guaranteed Sparse Recovery under Linear Transformation , 2013, ICML.
[16] Joel A. Tropp,et al. Universality laws for randomized dimension reduction, with applications , 2015, ArXiv.
[17] Andrea Montanari,et al. High dimensional robust M-estimation: asymptotic variance via approximate message passing , 2013, ArXiv.
[18] Z. Bai. METHODOLOGIES IN SPECTRAL ANALYSIS OF LARGE DIMENSIONAL RANDOM MATRICES , A REVIEW , 1999 .
[19] D. Donoho. For most large underdetermined systems of equations, the minimal 𝓁1‐norm near‐solution approximates the sparsest near‐solution , 2006 .
[20] Nicolas Macris,et al. The mutual information in random linear estimation , 2016, 2016 54th Annual Allerton Conference on Communication, Control, and Computing (Allerton).
[21] J. W. Silverstein,et al. Spectral Analysis of Large Dimensional Random Matrices , 2009 .
[22] R. DeVore,et al. A Simple Proof of the Restricted Isometry Property for Random Matrices , 2008 .
[23] Michael A. Saunders,et al. Atomic Decomposition by Basis Pursuit , 1998, SIAM J. Sci. Comput..
[24] Michael Elad,et al. Stable recovery of sparse overcomplete representations in the presence of noise , 2006, IEEE Transactions on Information Theory.
[25] Yoshiyuki Kabashima,et al. Erratum: A typical reconstruction limit of compressed sensing based on Lp-norm minimization , 2009, ArXiv.
[26] Volkan Cevher,et al. Estimation Error of the Lasso , 2016 .
[27] Christos Thrampoulidis,et al. Precise Error Analysis of Regularized $M$ -Estimators in High Dimensions , 2016, IEEE Transactions on Information Theory.
[28] R. Tibshirani,et al. The solution path of the generalized lasso , 2010, 1005.1971.
[29] Noureddine El Karoui,et al. Asymptotic behavior of unregularized and ridge-regularized high-dimensional robust regression estimators : rigorous results , 2013, 1311.2445.
[30] Emmanuel J. Candès,et al. A Probabilistic and RIPless Theory of Compressed Sensing , 2010, IEEE Transactions on Information Theory.
[31] Christos Thrampoulidis,et al. The squared-error of generalized LASSO: A precise analysis , 2013, 2013 51st Annual Allerton Conference on Communication, Control, and Computing (Allerton).
[32] Emmanuel J. Candès,et al. Near-Optimal Signal Recovery From Random Projections: Universal Encoding Strategies? , 2004, IEEE Transactions on Information Theory.
[33] R. Tibshirani,et al. Sparsity and smoothness via the fused lasso , 2005 .
[34] Nicolas Macris,et al. Mutual information for symmetric rank-one matrix estimation: A proof of the replica formula , 2016, NIPS.
[35] Andrea Montanari,et al. The LASSO Risk for Gaussian Matrices , 2010, IEEE Transactions on Information Theory.
[36] Y. Gordon. On Milman's inequality and random subspaces which escape through a mesh in ℝ n , 1988 .