GPU-accelerated and parallelized ELM ensembles for large-scale regression

Abstract The paper presents an approach for performing regression on large data sets in reasonable time, using an ensemble of extreme learning machines (ELMs). The main purpose and contribution of this paper are to explore how the evaluation of this ensemble of ELMs can be accelerated in three distinct ways: (1) training and model structure selection of the individual ELMs are accelerated by performing these steps on the graphics processing unit (GPU), instead of the processor (CPU); (2) the training of ELM is performed in such a way that computed results can be reused in the model structure selection, making training plus model structure selection more efficient; (3) the modularity of the ensemble model is exploited and the process of model training and model structure selection is parallelized across multiple GPU and CPU cores, such that multiple models can be built at the same time. The experiments show that competitive performance is obtained on the regression tasks, and that the GPU-accelerated and parallelized ELM ensemble achieves attractive speedups over using a single CPU. Furthermore, the proposed approach is not limited to a specific type of ELM and can be employed for a large variety of ELMs.

[1]  R. H. Myers Classical and modern regression with applications , 1986 .

[2]  Ed Anderson,et al.  LAPACK Users' Guide , 1995 .

[3]  James Demmel,et al.  Benchmarking GPUs to tune dense linear algebra , 2008, HiPC 2008.

[4]  Michel Verleysen,et al.  Ensemble Modeling with a Constrained Linear System of Leave-One-Out Outputs , 2010, ESANN.

[5]  James Demmel,et al.  Benchmarking GPUs to tune dense linear algebra , 2008, 2008 SC - International Conference for High Performance Computing, Networking, Storage and Analysis.

[6]  Chee Kheong Siew,et al.  Extreme learning machine: Theory and applications , 2006, Neurocomputing.

[7]  Robert K. L. Gay,et al.  Error Minimized Extreme Learning Machine With Growth of Hidden Nodes and Incremental Learning , 2009, IEEE Transactions on Neural Networks.

[8]  Hongming Zhou,et al.  Optimization method based extreme learning machine for classification , 2010, Neurocomputing.

[9]  Amaury Lendasse,et al.  Adaptive Ensemble Models of Extreme Learning Machines for Time Series Prediction , 2009, ICANN.

[10]  Jack Dongarra,et al.  LAPACK Users' guide (third ed.) , 1999 .

[11]  Yuan Lan,et al.  Ensemble of online sequential extreme learning machine , 2009, Neurocomputing.

[12]  Nikolaos Kourentzes,et al.  Automatic modelling of neural networks for time series prediction - in search of a uniform methodology across varying time frequencies , 2008 .

[13]  Amaury Lendasse,et al.  OP-ELM: Optimally Pruned Extreme Learning Machine , 2010, IEEE Transactions on Neural Networks.

[14]  Narasimhan Sundararajan,et al.  A Fast and Accurate Online Sequential Learning Algorithm for Feedforward Networks , 2006, IEEE Transactions on Neural Networks.

[15]  Erkki Oja,et al.  Solving Large Regression Problems using an Ensemble of GPU-accelerated ELMs , 2010, ESANN.

[16]  Kurt Keutzer,et al.  Fast support vector machine training and classification on graphics processors , 2008, ICML '08.

[17]  Calyampudi R. Rao,et al.  Generalized inverse of a matrix and its applications , 1972 .

[18]  G. Schwarz Estimating the Dimension of a Model , 1978 .

[19]  Guang-Bin Huang,et al.  Extreme learning machine: a new learning scheme of feedforward neural networks , 2004, 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No.04CH37541).

[20]  James Demmel,et al.  LAPACK Users' Guide, Third Edition , 1999, Software, Environments and Tools.

[21]  Amaury Lendasse,et al.  A faster model selection criterion for OP-ELM and OP-KNN: Hannan-Quinn criterion , 2009, ESANN.

[22]  Guang-Bin Huang,et al.  Convex Incremental Extreme Learning Machine , 2007 .

[23]  Benjamin Schrauwen,et al.  An experimental unification of reservoir computing methods , 2007, Neural Networks.

[24]  John Burkardt MATLAB Parallel Computing , 2009 .

[25]  Chee Kheong Siew,et al.  Universal Approximation using Incremental Constructive Feedforward Networks with Random Hidden Nodes , 2006, IEEE Transactions on Neural Networks.

[26]  Chee Kheong Siew,et al.  Real-time learning capability of neural networks , 2006, IEEE Trans. Neural Networks.