Generalization improvement for regularized least squares classification

In the past decades, regularized least squares classification (RLSC) is a commonly used supervised classification method in the machine learning filed because it can be easily resolved through the simple matrix analysis and achieve a close-form solution. Recently, some studies conjecture that the margin distribution is more crucial to the generalization performance. Moreover, from the view of margin distribution, RLSC only considers the first-order statistics (i.e., margin mean) and does not consider the actual higher-order statistics of margin distribution. In this paper, we propose a novel RLSC which takes into account the actual second-order (i.e., variance) information of margin distribution. It is intuitively expected that small margin variance will improve the generalization performance of RLSC from a geometric view. We incorporate the margin variance into the objective function of RLSC and achieve the optimal classifier by minimizing the margin variance. To evaluate the performance of our algorithm, we conduct a series of experiments on several benchmark datasets in comparison with RLSC, kernel minimum squared error, support vector machine and large margin distribution machine. And the empirical results verify the effectiveness of our algorithm and indicate that the margin distribution is helpful to improve the classification performance.

[1]  Hui Xue,et al.  Semi-supervised classification learning by discrimination-aware manifold regularization , 2015, Neurocomputing.

[2]  Mohamed Cheriet,et al.  Help-Training for semi-supervised support vector machines , 2011, Pattern Recognit..

[3]  Qiang Yang,et al.  Discriminatively regularized least-squares classification , 2009, Pattern Recognit..

[4]  Chih-Jen Lin,et al.  A dual coordinate descent method for large-scale linear SVM , 2008, ICML '08.

[5]  T. Poggio,et al.  Chapter 7 Regularized Least-Squares Classification , 2003 .

[6]  Li Wang,et al.  A modified Tikhonov regularization method , 2015, J. Comput. Appl. Math..

[7]  Robert E. Schapire,et al.  How boosting the margin can also boost classifier complexity , 2006, ICML.

[8]  Mikhail Belkin,et al.  Manifold Regularization: A Geometric Framework for Learning from Labeled and Unlabeled Examples , 2006, J. Mach. Learn. Res..

[9]  Zhi-Hua Zhou,et al.  A Refined Margin Analysis for Boosting Algorithms via Equilibrium Margin , 2011, J. Mach. Learn. Res..

[10]  Vladimir Vapnik,et al.  Statistical learning theory , 1998 .

[11]  Xuegong Zhang,et al.  Kernel MSE algorithm: a unified framework for KFD, LS-SVM and KRR , 2001, IJCNN'01. International Joint Conference on Neural Networks. Proceedings (Cat. No.01CH37222).

[12]  Jing Peng,et al.  SVM vs regularized least squares classification , 2004, ICPR 2004.

[13]  Charles A. Micchelli,et al.  Learning the Kernel Function via Regularization , 2005, J. Mach. Learn. Res..

[14]  Zhi-Hua Zhou Large Margin Distribution Learning , 2014, ANNPR.

[15]  Zhi-Hua Zhou,et al.  On the doubt about margin explanation of boosting , 2010, Artif. Intell..