An L1-Regularized naïVE Bayes-Inspired Classifier for Discarding Redundant and Irrelevant Predictors

The naive Bayes model is a simple but often satisfactory supervised classification method. The original naive Bayes scheme, does, however, have a serious weakness, namely, the harmful effect of redundant predictors. In this paper, we study how to apply a regularization technique to learn a computationally efficient classifier that is inspired by naive Bayes. The proposed formulation, combined with an L1-penalty, is capable of discarding harmful, redundant predictors. A modification of the LARS algorithm is devised to solve this problem. We tackle both real-valued and discrete predictors, assuring that our method is applicable to a wide range of data. In the experimental section, we empirically study the effect of redundant and irrelevant predictors. We also test the method on a high dimensional data set from the neuroscience field, where there are many more predictors than data cases. Finally, we run the method on a real data set than combines categorical with numeric predictors. Our approach is compared with several naive Bayes variants and other classification algorithms (SVM and kNN), and is shown to be competitive.

[1]  Concha Bielza,et al.  Forward stagewise naïve Bayes , 2011, Progress in Artificial Intelligence.

[2]  S. Rosset,et al.  Piecewise linear regularized solution paths , 2007, 0708.2197.

[3]  Pedro M. Domingos,et al.  Beyond Independence: Conditions for the Optimality of the Simple Bayesian Classifier , 1996, ICML.

[4]  Jinbo Bi,et al.  Dimensionality Reduction via Sparse Support Vector Machines , 2003, J. Mach. Learn. Res..

[5]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[6]  Pedro M. Domingos,et al.  On the Optimality of the Simple Bayesian Classifier under Zero-One Loss , 1997, Machine Learning.

[7]  S. Sathiya Keerthi,et al.  A simple and efficient algorithm for gene selection using sparse logistic regression , 2003, Bioinform..

[8]  Pedro Larrañaga,et al.  Feature selection in Bayesian classifiers for the prognosis of survival of cirrhotic patients treated with TIPS , 2005, J. Biomed. Informatics.

[9]  Aurélie Labbe,et al.  Refining gene signatures: a Bayesian approach , 2009, BMC Bioinformatics.

[10]  F. Fleuret Fast Binary Feature Selection with Conditional Mutual Information , 2004, J. Mach. Learn. Res..

[11]  D. Hand,et al.  Idiot's Bayes—Not So Stupid After All? , 2001 .

[12]  A. Ishai,et al.  Distributed and Overlapping Representations of Faces and Objects in Ventral Temporal Cortex , 2001, Science.

[13]  Marc Boullé,et al.  Compression-Based Averaging of Selective Naive Bayes Classifiers , 2007, J. Mach. Learn. Res..