Analysis of Noisy Evolutionary Optimization When Sampling Fails

In noisy evolutionary optimization, sampling is a common strategy to deal with noise, which evaluates the fitness of a solution multiple times (called sample size) independently and then uses the average to approximate the true fitness. Previous studies mainly focused on the empirical design of efficient sampling strategies, and the few theoretical analyses mainly proved the effectiveness of sampling with a fixed sample size in some situations. There are many fundamental theoretical issues to be addressed. In this paper, we first investigate the effect of sample size. By analyzing the (1+1)-EA on noisy LeadingOnes, we show that as the sample size increases, the running time can reduce from exponential to polynomial, but then return to exponential. This discloses that a proper sample size is crucial in practice. Then, we investigate what other strategies can work when sampling with any fixed sample size fails. By two illustrative examples, we prove that using parent populations can be better, and if using parent populations is also ineffective, adaptive sampling (i.e., sampling with an adaptive sample size) can work.

[1]  Zhi-Hua Zhou,et al.  Analyzing Evolutionary Optimization in Noisy Environments , 2013, Evolutionary Computation.

[2]  B. Hajek Hitting-time and occupation-time bounds implied by drift analysis with applications , 1982, Advances in Applied Probability.

[3]  Pietro Simone Oliveto,et al.  On the runtime analysis of the Simple Genetic Algorithm , 2014, Theor. Comput. Sci..

[4]  Xu Liu,et al.  An efficient energy aware virtual network migration based on genetic algorithm , 2019, Frontiers of Computer Science.

[5]  Robert Ivor John,et al.  Evolutionary optimisation of noisy multi-objective problems using confidence-based dynamic resampling , 2010, Eur. J. Oper. Res..

[6]  Yang Yu,et al.  Switch Analysis for Running Time Analysis of Evolutionary Algorithms , 2015, IEEE Transactions on Evolutionary Computation.

[7]  Zhuhong Zhang,et al.  Immune Algorithm with Adaptive Sampling in Noisy Environments and Its Application to Stochastic Optimization Problems , 2007, IEEE Computational Intelligence Magazine.

[8]  Jürgen Branke,et al.  Sequential Sampling in Noisy Environments , 2004, PPSN.

[9]  Carsten Witt,et al.  Runtime Analysis of the ( + 1) EA on Simple Pseudo-Boolean Functions , 2006, Evolutionary Computation.

[10]  Yang Yu,et al.  Subset Selection under Noise , 2017, NIPS.

[11]  Wusheng Chou,et al.  Path planning for mobile robot using self-adaptive learning particle swarm optimization , 2016, Science China Information Sciences.

[12]  Dirk Sudholt,et al.  A Simple Ant Colony Optimizer for Stochastic Shortest Path Problems , 2012, Algorithmica.

[13]  Carsten Witt,et al.  Runtime Analysis of the ( μ +1) EA on Simple Pseudo-Boolean Functions , 2006 .

[14]  Luc Devroye,et al.  Combinatorial methods in density estimation , 2001, Springer series in statistics.

[15]  Duc-Cuong Dang,et al.  Efficient Optimisation of Noisy Fitness Functions with Population-based Evolutionary Algorithms , 2015, FOGA.

[16]  Chao Qian,et al.  Distributed Pareto Optimization for Large-Scale Noisy Subset Selection , 2020, IEEE Transactions on Evolutionary Computation.

[17]  Chao Qian,et al.  Running Time Analysis of the (1+1\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$1+1$$\end{document})-EA for OneMax an , 2017, Algorithmica.

[18]  Andrew M. Sutton,et al.  The Compact Genetic Algorithm is Efficient Under Extreme Gaussian Noise , 2017, IEEE Transactions on Evolutionary Computation.

[19]  Xin Yao,et al.  On the Effectiveness of Sampling for Evolutionary Optimization in Noisy Environments , 2014, Evolutionary Computation.

[20]  Benjamin Doerr,et al.  Ants easily solve stochastic shortest path problems , 2012, GECCO '12.

[21]  I. Tyurin Refinement of the upper bounds of the constants in Lyapunov's theorem , 2010 .

[22]  Erick Cantú-Paz,et al.  Adaptive Sampling for Noisy Problems , 2004, GECCO.

[23]  Dorian Nogneng,et al.  A new analysis method for evolutionary optimization of dynamic and noisy objective functions , 2018, GECCO.

[24]  Pietro Simone Oliveto,et al.  Simplified Drift Analysis for Proving Lower Bounds in Evolutionary Computation , 2008, Algorithmica.

[25]  Ujjwal Maulik,et al.  A Survey of Multiobjective Evolutionary Algorithms for Data Mining: Part I , 2014, IEEE Transactions on Evolutionary Computation.

[26]  Stefan Droste,et al.  Design and Management of Complex Technical Processes and Systems by means of Computational Intelligence Methods Analysis of the (1+1) EA for a Noisy OneMax , 2004 .

[27]  Frank Neumann,et al.  Bioinspired computation in combinatorial optimization: algorithms and their computational complexity , 2012, GECCO '12.

[28]  Yang Yu,et al.  Evolutionary Learning: Advances in Theories and Algorithms , 2019 .

[29]  Olivier Teytaud,et al.  Analysis of runtime of optimization algorithms for noisy functions over discrete codomains , 2015, Theor. Comput. Sci..

[30]  Timo Kötzing,et al.  Optimizing expected path lengths with ant colony optimization using fitness proportional update , 2013, FOGA XII '13.

[31]  Xin Yao,et al.  Drift analysis and average time complexity of evolutionary algorithms , 2001, Artif. Intell..

[32]  Andrew M. Sutton,et al.  Robustness of Ant Colony Optimization to Noise , 2015, GECCO.

[33]  Benjamin Doerr,et al.  Multiplicative drift analysis , 2010, GECCO.

[34]  Pietro Simone Oliveto,et al.  Erratum: Simplified Drift Analysis for Proving Lower Bounds in Evolutionary Computation , 2008, PPSN.

[35]  Dirk Sudholt,et al.  On the robustness of evolutionary algorithms to noise: refined results and an example where noise helps , 2018, GECCO.

[36]  Chao Qian,et al.  Towards a Running Time Analysis of the (1+1)-EA for OneMax and LeadingOnes Under General Bit-Wise Noise , 2018, PPSN.

[37]  Yang Yu,et al.  Analysis of Noisy Evolutionary Optimization When Sampling Fails , 2020, Algorithmica.

[38]  Anne Auger,et al.  Theory of Randomized Search Heuristics: Foundations and Recent Developments , 2011, Theory of Randomized Search Heuristics.

[39]  Adam Prügel-Bennett,et al.  Run-Time Analysis of Population-Based Evolutionary Algorithm in Noisy Environments , 2015, FOGA.