Accuracy, Parsimony, and Generality in Evolutionary Learning Systems via Multiobjective Selection

Evolutionary learning systems (also known as Pittsburgh learningclassifier systems) need to balance accuracy and parsimony for evolving high quality general hypotheses. The learning process used in evolutionary learning systems is based on a set of training instances that sample the target concept to be learned. Thus, the learning process may overfit the learned hypothesis to the given set of training instances. In order to address some of these issues, this paper introduces a multiobjective approach to evolutionary learning systems. Thus, we translate the selection of promising hypotheses into a two-objective problem that looks for: (1) accurate (low error), and (2) compact (low complexity) solutions. Using the proposed multiobjective approach a set of compromise hypotheses are spread along the Pareto front. We also introduce a theory of the impact of noise when sampling the target concept to be learned, as well as the appearance of overfitted hypotheses as the result of perturbations on high quality generalization hypotheses in the Pareto front.

[1]  V. Pareto,et al.  Vilfredo Pareto. Cours d’Économie Politique , 1897 .

[2]  Charles Gide,et al.  Cours d'économie politique , 1911 .

[3]  Dipl. Ing. Karl Heinz Kellermayer NUMERISCHE OPTIMIERUNG VON COMPUTER-MODELLEN MITTELS DER EVOLUTIONSSTRATEGIE Hans-Paul Schwefel Birkhäuser, Basel and Stuttgart, 1977 370 pages Hardback SF/48 ISBN 3-7643-0876-1 , 1977 .

[4]  Stephen F. Smith,et al.  Flexible Learning of Problem Solving Heuristics Through Adaptive Search , 1983, IJCAI.

[5]  Leo Breiman,et al.  Classification and Regression Trees , 1984 .

[6]  David E. Goldberg,et al.  Genetic Algorithms with Sharing for Multimodalfunction Optimization , 1987, ICGA.

[7]  David E. Goldberg,et al.  Genetic Algorithms in Search Optimization and Machine Learning , 1988 .

[8]  Kenneth A. De Jong,et al.  Learning Concept Classification Rules Using Genetic Algorithms , 1991, IJCAI.

[9]  C. Janikow Inductive learning of decision rules from attribute-based examples: a knowledge-intensive genetic algorithm approach , 1992 .

[10]  J. Ross Quinlan,et al.  C4.5: Programs for Machine Learning , 1992 .

[11]  John H. Holland,et al.  Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence , 1992 .

[12]  Zbigniew Michalewicz,et al.  Genetic Algorithms + Data Structures = Evolution Programs , 1992, Artificial Intelligence.

[13]  John R. Koza,et al.  Genetic programming - on the programming of computers by means of natural selection , 1993, Complex adaptive systems.

[14]  Lothar Thiele,et al.  Genetic Programming and Redundancy , 1994 .

[15]  Walter Alden Tackett,et al.  Recombination, selection, and the genetic construction of computer programs , 1994 .

[16]  Kalyanmoy Deb,et al.  Muiltiobjective Optimization Using Nondominated Sorting in Genetic Algorithms , 1994, Evolutionary Computation.

[17]  L. Altenberg EMERGENT PHENOMENA IN GENETIC PROGRAMMING , 1994 .

[18]  Peter Nordin,et al.  Complexity Compression and Evolution , 1995, ICGA.

[19]  Stewart W. Wilson Classifier Fitness Based on Accuracy , 1995, Evolutionary Computation.

[20]  Thomas Bäck,et al.  Generalized Convergence Models for Tournament- and (mu, lambda)-Selection , 1995, ICGA.

[21]  David E. Goldberg,et al.  Genetic Algorithms, Tournament Selection, and the Effects of Noise , 1995, Complex Syst..

[22]  Hans-Paul Schwefel,et al.  Parallel Problem Solving from Nature — PPSN IV , 1996, Lecture Notes in Computer Science.

[23]  Tobias Blickle,et al.  Evolving Compact Solutions in Genetic Programming: A Case Study , 1996, PPSN.

[24]  Thomas G. Dietterich What is machine learning? , 2020, Archives of Disease in Childhood.

[25]  Christopher J. Merz,et al.  UCI Repository of Machine Learning Databases , 1996 .

[26]  D. Fogel Evolutionary algorithms in theory and practice , 1997, Complex..

[27]  F. Jim Pareto-optimality in Fuzzy Modeling , 1998 .

[28]  Riccardo Poli,et al.  Fitness Causes Bloat: Mutation , 1997, EuroGP.

[29]  Terence Soule,et al.  Effects of Code Growth and Parsimony Pressure on Populations in Genetic Programming , 1998, Evolutionary Computation.

[30]  Ian H. Witten,et al.  Generating Accurate Rule Sets Without Global Optimization , 1998, ICML.

[31]  Jude W. Shavlik,et al.  Machine Learning: Proceedings of the Fifteenth International Conference , 1998 .

[32]  David A. Van Veldhuizen,et al.  Evolutionary Computation and Convergence to a Pareto Front , 1998 .

[33]  Elisabet Golobardes,et al.  Automatic diagnosis with genetic algorithms and case-based reasoning , 1999, Artif. Intell. Eng..

[34]  Ian H. Witten,et al.  Data mining: practical machine learning tools and techniques, 3rd Edition , 1999 .

[35]  Ester Bernadó-Mansilla,et al.  A Study of a Genetic Classifier System Based on the Pittsburgh Approach on a Medical Domain , 1999, IEA/AIE.

[36]  Stewart W. Wilson Get Real! XCS with Continuous-Valued Inputs , 1999, Learning Classifier Systems.

[37]  Eckart Zitzler,et al.  Evolutionary algorithms for multiobjective optimization: methods and applications , 1999 .

[38]  Lothar Thiele,et al.  Comparison of Multiobjective Evolutionary Algorithms: Empirical Results , 2000, Evolutionary Computation.

[39]  Karl Rihaczek,et al.  1. WHAT IS DATA MINING? , 2019, Data Mining for the Social Sciences.

[40]  Carlos A. Coello Coello,et al.  An updated survey of GA-based multiobjective optimization techniques , 2000, CSUR.

[41]  Kalyanmoy Deb,et al.  A Fast Elitist Non-dominated Sorting Genetic Algorithm for Multi-objective Optimisation: NSGA-II , 2000, PPSN.

[42]  Gary B. Lamont,et al.  Multiobjective Evolutionary Algorithms: Analyzing the State-of-the-Art , 2000, Evolutionary Computation.

[43]  Vili Podgorelec,et al.  Fighting Program Bloat with the Fractal Complexity Measure , 2000, EuroGP.

[44]  William B. Langdon,et al.  Quadratic Bloat in Genetic Programming , 2000, GECCO.

[45]  Zbigniew Michalewicz,et al.  Evolutionary Computation 2 , 2000 .

[46]  Kenneth A. De Jong,et al.  Evolving Behaviors for Cooperating Agents , 2000, ISMIS.

[47]  Antonio F. Gómez-Skarmeta,et al.  Accurate, Transparent, and Compact Fuzzy Models for Function Approximation and Dynamic Modeling through Multi-objective Evolutionary Optimization , 2001, EMO.

[48]  Lothar Thiele,et al.  Multiobjective genetic programming: reducing bloat using SPEA2 , 2001, Proceedings of the 2001 Congress on Evolutionary Computation (IEEE Cat. No.01TH8546).

[49]  Xavier Llorà,et al.  XCS and GALE: A Comparative Study of Two Learning Classifier Systems on Data Mining , 2001, IWLCS.

[50]  Xavier Llorà,et al.  Inducing Partially-Defined Instances with Evolutionary Algorithms , 2001, ICML.

[51]  Xavier Llorà,et al.  Knowledge-independent data mining with fine-grained parallel evolutionary algorithms , 2001 .

[52]  Marco Laumanns,et al.  SPEA2: Improving the strength pareto evolutionary algorithm , 2001 .

[53]  D. Goldberg,et al.  Minimal Achievable Error in the LED problem , 2002 .

[54]  Ian H. Witten,et al.  Data mining: practical machine learning tools and techniques with Java implementations , 2002, SGMD.

[55]  Kalyanmoy Deb,et al.  A fast and elitist multiobjective genetic algorithm: NSGA-II , 2002, IEEE Trans. Evol. Comput..

[56]  Terence Soule,et al.  Exons and Code Growth in Genetic Programming , 2002, EuroGP.

[57]  Martin V. Butz,et al.  Anticipatory Learning Classifier Systems , 2002, Genetic Algorithms and Evolutionary Computation.

[58]  Xavier Llorà,et al.  Bounding the Effect of Noise in Multiobjective Learning Classifier Systems , 2003, Evolutionary Computation.

[59]  William B. Langdon,et al.  Some Considerations on the Reason for Bloat , 2002, Genetic Programming and Evolvable Machines.

[60]  J. Ross Quinlan,et al.  Induction of Decision Trees , 1986, Machine Learning.

[61]  D. Kibler,et al.  Instance-based learning algorithms , 2004, Machine Learning.

[62]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.