A comparison of evolution strategies and backpropagation for neural network training

Abstract This report investigates evolution strategies (ESs, a subclass of evolutionary algorithms) as an alternative to gradient-based neural network training. Based on an empirical comparison of population- and gradient-based search, we derive hints for parameterization and draw conclusions about the usefulness of evolution strategies for this purpose. We will see that ESs can only compete with gradient-based search in the case of small problems and that ESs are good for training neural networks with a non-differentiable activation function. Insights into how evolution strategies behave in search spaces generated by neural networks are offered. Here, we see that for this class of objective functions, the dimensionality of the problem is critical. With increasing numbers of decision variables, the learning becomes more and more difficult for ESs, and an “efficient” parameterization becomes crucial.

[1]  H. Kitano Neurogenetic learning: an integrated method of designing and training neural networks using genetic algorithms , 1994 .

[2]  M. Mandischer,et al.  Comparison of Neural Networks, Evolutionary Techniques and Thermodynamic Group Contribution Methods , 2001 .

[3]  Norbert Hoffmann,et al.  Simulation Neuronaler Netze , 1991 .

[4]  Hiroaki Kitano,et al.  Empirical Studies on the Speed of Convergence of Neural Network Training Using Genetic Algorithms , 1990, AAAI.

[5]  Günter Rudolph,et al.  Contemporary Evolution Strategies , 1995, ECAL.

[6]  Thomas Bäck,et al.  An Overview of Parameter Control Methods by Self-Adaption in Evolutionary Algorithms , 1998, Fundam. Informaticae.

[7]  Hans-Georg Beyer,et al.  Toward a Theory of Evolution Strategies: On the Benefits of Sex the (/, ) Theory , 1995, Evolutionary Computation.

[8]  David B. Fogel,et al.  Alternative Neural Network Training Methods , 1995, IEEE Expert.

[9]  Reinhard Männer,et al.  Parallel Problem Solving from Nature — PPSN III , 1994, Lecture Notes in Computer Science.

[10]  Willfried Wienholt Minimizing the System Error in Feedforward Neural Networks with Evolution Strategy , 1993 .

[11]  Martin Mandischer An empirical investigation of neural networks, evolution strategies, and evolutionary-trained neural networks and their application to some chemical engineering problems , 2000 .

[12]  H. White,et al.  Universal approximation using feedforward networks with non-sigmoid hidden layer activation functions , 1989, International 1989 Joint Conference on Neural Networks.

[13]  Nikolaus Hansen,et al.  Step-Size Adaption Based on Non-Local Use of Selection Information , 1994, PPSN.

[14]  Lawrence Davis,et al.  Training Feedforward Neural Networks Using Genetic Algorithms , 1989, IJCAI.

[15]  Andrew Luk,et al.  Evolution of connection weights combined with local search for multi-layered neural network , 1996, Proceedings of IEEE International Conference on Evolutionary Computation.

[16]  Cornelia Kappler,et al.  Are Evolutionary Algorithms Improved by Large Mutations? , 1996, PPSN.

[17]  Thomas Bäck,et al.  Evolutionary Algorithms in Theory and Practice , 1996 .

[18]  Jacques Periaux,et al.  Genetic Algorithms in Engineering and Computer Science , 1996 .