Adapting Self-Adaptive Parameters in Evolutionary Algorithms

The lognormal self-adaptation has been used extensively in evolutionary programming (EP) and evolution strategies (ES) to adjust the search step size for each objective variable. However, it was discovered in our previous study (K.-H. Liang, X. Yao, Y. Liu, C. Newton, and D. Hoffman, in Evolutionary Programming VII. Proc. of the Seventh Annual Conference on Evolutionary Programming, vol. 1447, edited by V. Porto, N. Saravanan, D. Waagen, and A. Eiben, Lecture Notes in Computer Science, Springer: Berlin, pp. 291–300, 1998) that such self-adaptation may rapidly lead to a search step size that is far too small to explore the search space any further, and thus stagnates search. This is called the loss of step size control. It is necessary to use a lower bound of search step size to avoid this problem. Unfortunately, the optimal setting of lower bound is highly problem dependent. This paper first analyzes both theoretically and empirically how the step size control was lost. Then two schemes of dynamic lower bound are proposed. The schemes enable the EP algorithm to adjust the lower bound dynamically during evolution. Experimental results are presented to demonstrate the effectiveness and efficiency of the dynamic lower bound for a set of benchmark functions.

[1]  Hans-Paul Schwefel,et al.  Evolution and optimum seeking , 1995, Sixth-generation computer technology series.

[2]  Ko-Hsin Liang,et al.  An Experimental Investigation of Self-Adaptation in Evolutionary Programming , 1998, Evolutionary Programming.

[3]  Feller William,et al.  An Introduction To Probability Theory And Its Applications , 1950 .

[4]  Xin Yao,et al.  Evolutionary programming made faster , 1999, IEEE Trans. Evol. Comput..

[5]  Hans-Paul Schwefel,et al.  Evolution and Optimum Seeking: The Sixth Generation , 1993 .

[6]  Kumar Chellapilla,et al.  Combining mutation operators in evolutionary programming , 1998, IEEE Trans. Evol. Comput..

[7]  Heinz Mühlenbein,et al.  Predictive Models for the Breeder Genetic Algorithm I. Continuous Parameter Optimization , 1993, Evolutionary Computation.

[8]  Hans-Georg Beyer,et al.  Toward a Theory of Evolution Strategies: Self-Adaptation , 1995, Evolutionary Computation.

[9]  BeyerHans-Georg Toward a theory of evolution strategies , 1993 .

[10]  Xin Yao,et al.  Fast Evolutionary Programming , 1996, Evolutionary Programming.

[11]  Xin Yao,et al.  An Analysis of Evolutionary Algorithms Based on Neighborhood and Step Sizes , 1997, Evolutionary Programming.

[12]  W. Price Global optimization by controlled random search , 1983 .

[13]  Thomas Bäck,et al.  Evolutionary algorithms in theory and practice - evolution strategies, evolutionary programming, genetic algorithms , 1996 .

[14]  David B. Fogel,et al.  Meta-evolutionary programming , 1991, [1991] Conference Record of the Twenty-Fifth Asilomar Conference on Signals, Systems & Computers.

[15]  A. Griewank Generalized descent for global optimization , 1981 .

[16]  Thomas Bäck,et al.  An Overview of Evolutionary Algorithms for Parameter Optimization , 1993, Evolutionary Computation.

[17]  Thomas Bäck,et al.  An Overview of Evolutionary Computation , 1993, ECML.

[18]  Charles M. Grinstead,et al.  Introduction to probability , 1999, Statistics for the Behavioural Sciences.

[19]  David B. Fogel,et al.  A Comparison of Evolutionary Programming and Genetic Algorithms on Selected Constrained Optimization Problems , 1995, Simul..

[20]  P. A. P. Moran,et al.  An introduction to probability theory , 1968 .

[21]  J. Stapleton Introduction to Probability Theory and Statistical Inference , 1970 .

[22]  D. Ackley A connectionist machine for genetic hillclimbing , 1987 .