Generalization in the XCSF Classifier System: Analysis, Improvement, and Extension

We analyze generalization in XCSF and introduce three improvements. We begin by showing that the types of generalizations evolved by XCSF can be influenced by the input range. To explain these results we present a theoretical analysis of the convergence of classifier weights in XCSF which highlights a broader issue. In XCSF, because of the mathematical properties of the Widrow-Hoff update, the convergence of classifier weights in a given subspace can be slow when the spread of the eigenvalues of the autocorrelation matrix associated with each classifier is large. As a major consequence, the system's accuracy pressure may act before classifier weights are adequately updated, so that XCSF may evolve piecewise constant approximations, instead of the intended, and more efficient, piecewise linear ones. We propose three different ways to update classifier weights in XCSF so as to increase the generalization capabilities of XCSF: one based on a condition-based normalization of the inputs, one based on linear least squares, and one based on the recursive version of linear least squares. Through a series of experiments we show that while all three approaches significantly improve XCSF, least squares approaches appear to be best performing and most robust. Finally we show how XCSF can be extended to include polynomial approximations.

[1]  Daniele Loiacono,et al.  Extending XCSF beyond linear approximation , 2005, GECCO '05.

[2]  M. F. Fuller,et al.  Practical Nonparametric Statistics; Nonparametric Statistical Inference , 1973 .

[3]  Stewart W. Wilson,et al.  XCS with Computable Prediction in Multistep Environments , 2005 .

[4]  Larry Bull,et al.  Accuracy-based Neuro And Neuro-fuzzy Classifier Systems , 2002, GECCO.

[5]  Pier Luca Lanzi,et al.  An Analysis of Generalization in the XCS Classifier System , 1999, Evolutionary Computation.

[6]  Daniele Loiacono,et al.  XCS with computed prediction in multistep environments , 2005, GECCO '05.

[7]  Martin V. Butz,et al.  Strong, Stable, and Reliable Fitness Pressure in XCS due to Tournament Selection , 2005, Genetic Programming and Evolvable Machines.

[8]  Larry Bull,et al.  Backpropagation in Accuracy-Based Neural Learning Classifier Systems , 2005, IWLCS.

[9]  E GoldbergDavid,et al.  Generalization in the XCSF Classifier System , 2007 .

[10]  Bernard Widrow,et al.  Adaptive switching circuits , 1988 .

[11]  Larry Bull,et al.  A Neural Learning Classifier System with Self-Adaptive Constructivism for Mobile Robot Control , 2006, Artificial Life.

[12]  Richard S. Sutton,et al.  Introduction to Reinforcement Learning , 1998 .

[13]  Stewart W. Wilson Generalization in the XCS Classifier System , 1998 .

[14]  Daniele Loiacono,et al.  Computed prediction in binary multistep problems , 2008, 2008 IEEE Congress on Evolutionary Computation (IEEE World Congress on Computational Intelligence).

[15]  Stewart W. Wilson Classifiers that approximate functions , 2002, Natural Computing.

[16]  Stewart W. Wilson Mining Oblique Data with XCS , 2000, IWLCS.

[17]  Pier Luca Lanzi,et al.  Mining interesting knowledge from data with the XCS classifier system , 2001 .

[18]  Graham C. Goodwin,et al.  Adaptive filtering prediction and control , 1984 .

[19]  S. Glantz,et al.  Primer of Applied Regression & Analysis of Variance , 1990 .

[20]  Larry Bull,et al.  A neural learning classifier system with self-adaptive constructivism , 2003, The 2003 Congress on Evolutionary Computation, 2003. CEC '03..

[21]  David E. Goldberg,et al.  Genetic Algorithms in Search Optimization and Machine Learning , 1988 .

[22]  L. Baird Reinforcement Learning Through Gradient Descent , 1999 .

[23]  Stewart W. Wilson Classifier Fitness Based on Accuracy , 1995, Evolutionary Computation.

[24]  Peter Dayan,et al.  Technical Note: Q-Learning , 2004, Machine Learning.

[25]  Justin A. Boyan,et al.  Least-Squares Temporal Difference Learning , 1999, ICML.

[26]  Vidroha Debroy,et al.  Genetic Programming , 1998, Lecture Notes in Computer Science.

[27]  Stewart W. Wilson Function approximation with a classifier system , 2001 .

[28]  Daniele Loiacono,et al.  XCS with computed prediction in continuous multistep environments , 2005, 2005 IEEE Congress on Evolutionary Computation.

[29]  Larry Bull,et al.  Prediction calculation in accuracy-based neural learning classifier systems , 2004 .

[30]  Larry Bull,et al.  Robotic experimentation using constructivism within a neural learning classifier system , 2003 .

[31]  Stewart W. Wilson Classifier Systems for Continuous Payoff Environments , 2004, GECCO.

[32]  John H. Holland,et al.  Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence , 1992 .

[33]  S. Haykin,et al.  Adaptive Filter Theory , 1986 .

[34]  William H. Press,et al.  Numerical Recipes in C The Art of Scientific Computing , 1995 .

[35]  Andrew G. Barto,et al.  Linear Least-Squares Algorithms for Temporal Difference Learning , 2005, Machine Learning.

[36]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[37]  Martin V. Butz,et al.  Gradient Descent Methods in Learning Classifier Systems , 2004 .

[38]  Martin V. Butz,et al.  Gradient descent methods in learning classifier systems: improving XCS performance in multistep problems , 2005, IEEE Transactions on Evolutionary Computation.

[39]  Martin V. Butz,et al.  Tournament Selection: Stable Fitness Pressure in XCS , 2003, GECCO.

[40]  Daniele Loiacono,et al.  XCS with computed prediction for the learning of Boolean functions , 2005, 2005 IEEE Congress on Evolutionary Computation.

[41]  Martin J. Oates,et al.  A Preliminary Investigation of Modified XCS as a Generic Data Mining Tool , 2001, IWLCS.

[42]  Peter Dayan,et al.  Q-learning , 1992, Machine Learning.

[43]  Simon Haykin,et al.  Neural Networks: A Comprehensive Foundation , 1998 .

[44]  Martin V. Butz,et al.  An algorithmic description of XCS , 2000, Soft Comput..