Joint learning and pruning of decision forests

Decision forests, such as Random forest (Breiman, 2001) and Extremely randomized trees (Geurts et al., 2006), are state-of-the-art supervised learning methods. Unfortunately, the size of the trees increases (at worst linearly) with the size of the training set, leading to substantial memory footprints when storing the forests. Several post-processing methods have been proposed to tackle this problem (Meinshausen et al., 2009; Joly et al., 2012; De Vleeschouwer et al., 2015; Ren et al., 2015). Although very e↵ective, they require building the whole forest first and then solving a highdimensional optimization problem to (post-)prune it. In this research, we would like to investigate whether a similar compression e↵ect could be achieved directly during tree growing, without ever requiring the development of the whole forest. For that purpose, we propose a joint learning and pruning (JLP) method which iteratively and greedily deepens the trees by optimizing globally the sequence of nodes to develop, while still choosing locally the splitting variables and cut points at all tree nodes based on the standard score criterion. The algorithm is presented in Section 2 and some results and comparisons with more naive and post-pruning methods are reported in Section 3.

[1]  Pierre Geurts,et al.  L1-based compression of random forest models , 2012, ESANN.

[2]  Laurent Jacques,et al.  Mitigating memory requirements for random trees/ferns , 2015, 2015 IEEE International Conference on Image Processing (ICIP).

[3]  D. Ruppert The Elements of Statistical Learning: Data Mining, Inference, and Prediction , 2004 .

[4]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[5]  Robert Tibshirani,et al.  The Elements of Statistical Learning: Data Mining, Inference, and Prediction, 2nd Edition , 2001, Springer Series in Statistics.

[6]  Jian Sun,et al.  Global refinement of random forest , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[7]  Pierre Geurts,et al.  Extremely randomized trees , 2006, Machine Learning.

[8]  N. Meinshausen Forest Garrote , 2009, 0906.3590.

[9]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.