L1-based compression of random forest models

Random forests are effective supervised learning methods applicable to large-scale datasets. However, the space complexity of tree ensembles, in terms of their total number of nodes, is often prohibitive, spe- cially in the context of problems with very high-dimensional input spaces. We propose to study their compressibility by applying a L1-based regu- larization to the set of indicator functions defined by all their nodes. We show experimentally that preserving or even improving the model accuracy while significantly reducing its space complexity is indeed possible.

[1]  J. Freidman,et al.  Multivariate adaptive regression splines , 1991 .

[2]  Leo Breiman,et al.  Bias, Variance , And Arcing Classifiers , 1996 .

[3]  Daniel Hernández-Lobato,et al.  An Analysis of Ensemble Pruning Techniques Based on Ordered Aggregation , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[4]  Laurent Heutte,et al.  On the selection of decision trees in Random Forests , 2009, 2009 International Joint Conference on Neural Networks.

[5]  E.J. Candes,et al.  An Introduction To Compressive Sampling , 2008, IEEE Signal Processing Magazine.

[6]  C. Sitthi-amorn,et al.  Bias , 1993, The Lancet.

[7]  Bogdan E. Popescu,et al.  PREDICTIVE LEARNING VIA RULE ENSEMBLES , 2008, 0811.1679.

[8]  R. Tibshirani,et al.  Forward stagewise regression and the monotone lasso , 2007, 0705.0269.

[9]  Pierre Geurts,et al.  Extremely randomized trees , 2006, Machine Learning.

[10]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[11]  Pierre Geurts,et al.  Some Enhencements of Decision Tree Bagging , 2000, PKDD.

[12]  N. Meinshausen Node harvest: simple and interpretable regression and classication , 2009, 0910.2145.

[13]  N. Meinshausen Forest Garrote , 2009, 0906.3590.

[14]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .