[1]
Jose L. Aguilar and Mariela Cerrada. Reliability-Centered Maintenance Methodology-Based Fuzzy Classifier System Design for Fault Tolerance. In Koza et al. [529], page 621. One page paper.

[2]
Jose Aguilar and Mariela Cerrada. Fuzzy classifier system and genetic programming on system identification problems. In Lee Spector, Erik D. Goodman, Annie Wu, W.B. Langdon, Hans-Michael Voigt, Mitsuo Gen, Sandip Sen, Marco Dorigo, Shahram Pezeshk, Max H. Garzon, and Edmund Burke, editors, Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2001), pages 1245-1251, San Francisco, California, USA, 7-11 July 2001. Morgan Kaufmann.

[3]
J. Aguilar-Ruiz, J. Riquelme, and M. Toro. Evolutionary learning of hierarchical decision rules. IEEE Transactions on Systems, Man and Cybernetics, Part B, 33(2):324-331, 2003.

[4]
Manu Ahluwalia and Larry Bull. A Genetic Programming-based Classifier System. In Banzhaf et al. [32], pages 11-18.
In this paper we present initial results from the use of Genetic Programming in Holland's learning Classifier System architecture. Rule are of the form condition(s)/action in which the conditions are represented as binary strings and the actions are represented by S-expressions. The basic Classifier System architecture used here in more analogous to Wilson's ZCS system than Holland's original formalism since no internal message list exists; stimulus-response rules are developed here. Using two well-known classification tasks it is shown that our approach can develop useful feature extractors for the K-nearest-neighbour algorithm. We also show that the use of niche-based evolutionary search can improve performance.

[5]
Sameer Alam, Kamran Shafi, Hussein A. Abbass, and Micheal Barlow. Evolving Air Traffic Scenarios for the Evaluation of Conflict Detection Models. In 6th Eurocontrol Innovative Research Workshop, Eurocontrol Experimental Research Centre, France, Dec 4-6, 2007, 2007.

[6]
Sameer Alam, Kamran Shafi, Hussein A. Abbass, and Micheal Barlow. Forming an Intelligent Ensemble of Conflict Detection Algorithms in Free Flight by Data Mining the Scenario Space. Transportation Research Part-C: Emerging Technologies, 2009. In Press.

[7]
Rudolf F. Albrecht, Nigel C. Steele, and Colin R. Reeves, editors. Proceedings of the International Conference on Artificial Neural Nets and Genetic Algorithms. Spring-Verlag, 1993.

[8]
Peter J. Angeline, Zbyszek Michalewicz, Marc Schoenauer, Xin Yao, and Ali Zalzala, editors. Proceedings of the 1999 Congress on Evolutionary Computation CEC99, Washington (DC), 1999. IEEE Press.

[9]
Plamen Angelov. Evolving Rule-based Models. A tool for design of flexible adaptive systems, volume 92 of Studies in fuzziness and soft computing. Springer-Verlag, 2002.

[10]
G. Armano, M. Marchesi, and A. Murru. Nxcs: Hybrid approach to stock indexes forecasting. In Shu-Heng Chen, editor, Genetic Algorithms and Genetic Programming in Computational Finance, chapter 6, pages 125-158. Kluwer, 2002.

[11]
G. Armano, A. Murru, and F. Roli. Stock market prediction by a mixture of genetic-neural experts. International Journal of Pattern Recognition and Artificial Intelligence, 16(5):501-526, 2002.

[12]
G. Armano, M. Marchesi, and A. Murru. A hybrid genetic-neural architecture for stock indexes forecasting. To appear in the Int. Journal of Information Sciences, 2004.

[13]
W. Brian Arthur, John H. Holland, Blake LeBaron, Richard Palmer, and Paul Talyer. Asset Pricing Under Endogenous Expectations in an Artificial Stock Market. Technical report, Santa Fe Institute, 1996. This is the original version of LeBaron1999a.

[14]
Jaume Bacardit and Josep M. Garrell. Evolution of adaptive discretization intervals for A rule-based genetic learning system. In W. B. Langdon, E. Cantú-Paz, K. Mathias, R. Roy, D. Davis, R. Poli, K. Balakrishnan, V. Honavar, G. Rudolph, J. Wegener, L. Bull, M. A. Potter, A. C. Schultz, J. F. Miller, E. Burke, and N. Jonoska, editors, GECCO 2002: Proceedings of the Genetic and Evolutionary Computation Conference, page 677. Morgan Kaufmann Publishers, 2002.

[15]
Jaume Bacardit and Josep Maria Garrell. Evolving multiple discretizations with adaptive intervals for a Pittsburgh rule-based learning classifier system. In E. Cantú-Paz, J. A. Foster, K. Deb, D. Davis, R. Roy, U.-M. O'Reilly, H.-G. Beyer, R. Standish, G. Kendall, S. Wilson, M. Harman, J. Wegener, D. Dasgupta, M. A. Potter, A. C. Schultz, K. Dowsland, N. Jonoska, and J. Miller, editors, Genetic and Evolutionary Computation -- GECCO-2003, pages 1818-1831, Berlin, 2003. Springer-Verlag.

[16]
Jaume Bacardit and Josep Maria Garrell. Bloat control and generalization pressure using the minimum description length principle for a pittsburgh approach learning classifier system. In Tim Kovacs, Xavier LLòra, Keiki Takadama, Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson, editors, Learning Classifier Systems. International Workshops, IWLCS 2003-2005, Revised Selected Papers, volume 4399 of LNCS, pages 59-79. Springer, 2007.

[17]
Jaume Bacardit and Natalio Krasnogor. Empirical evaluation of ensemble techniques for a pittsburgh learning classifier system. In Jaume Bacardit, Ester Bernadó-Mansilla, Martin Butz, Tim Kovacs, Xavier Llorà, and Keiki Takadama, editors, Learning Classifier Systems. 10th and 11th International Workshops (2006-2007), volume 4998/2008 of Lecture Notes in Computer Science, pages 255-268. Springer, 2008.
Ensemble techniques have proved to be very successful in boosting the performance of several types of machine learning methods. In this paper, we illustrate its usefulness in combination with GAssist, a Pittsburgh-style Learning Classifier System. Two types of ensembles are tested. First we evaluate an ensemble for consensus prediction. In this case several rule sets learnt using GAssist with different initial random seeds are combined using a flat voting scheme in a fashion similar to bagging. The second type of ensemble is intended to deal more efficiently with ordinal classification problems. That is, problems where the classes have some intrinsic order between them and, in case of misclassification, it is preferred to predict a class that is close to the correct one within the class intrinsic order. The ensemble for consensus prediction is evaluated using 25 datasets from the UCI repository. The hierarchical ensemble is evaluated using a Bioinformatics dataset. Both methods significantly improve the performance and behaviour of GAssist in all the tested domains.

[18]
Jaume Bacardit, David E. Goldberg, and Martin V. Butz. Improving the performance of a pittsburgh learning classifier system using a default rule. In Tim Kovacs, Xavier LLòra, Keiki Takadama, Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson, editors, Learning Classifier Systems. International Workshops, IWLCS 2003-2005, Revised Selected Papers, volume 4399 of LNCS, pages 291-307. Springer, 2007.

[19]
J. Bacardit, M. Stout, J.D. Hirst, and N. Krasnogor. Data mining in proteomics with learning classifier systems. In L. Bull, E. Bernadó Mansilla, and J. Holmes, editors, Learning Classifier Systems in Data Mining, pages 17-46. Springer, 2008.

[20]
J. Bacardit, E.K. Burke, and N. Krasnogor. Improving the scalability of rule-based evolutionary learning. Memetic Computing, 1(1), to appear in 2009.

[21]
Jaume Bacardit. Pittsburgh Genetic-Based Machine Learning in the Data Mining era: Representations, generalization, and run-time. PhD thesis, Universitat Ramon Llull, 2004.

[22]
Thomas Bäck, David B. Fogel, and Zbigniew Michalewicz, editors. Handbook of Evolutionary Computation. Institute of Physics Publishing and Oxford University Press, 1997. http://www.iop.org/Books/Catalogue/.

[23]
Thomas Bäck, Ulrich Hammel, and Hans-Paul Schwefel. Evolutionary computation: Comments on the history and current state. IEEE Transactions on Evolutionary Computation, 1(1):3-17, 1997.
Evolutionary computation has started to receive significant attention during the last decade, although the origins can be traced back to the late 1950's. This article surveys the history as well as the current state of this rapidly growing field. We describe the purpose, the general structure, and the working principles of different approaches, including genetic algorithms (GA) [with links to genetic programming (GP) and classifier systems (CS)], evolution strategies (ES), and evolutionary programming (EP) by analysis and comparison of their most important constituents (i.e., representations, variation operators, reproduction, and selection mechanism). Finally, we give a brief overview on the manifold of application domains, although this necessarily must remain incomplete.

[24]
Thomas Bäck, editor. Proceedings of the 7th International Conference on Genetic Algorithms (ICGA97). Morgan Kaufmann, 1997.

[25]
Jalal Baghdadchi. A Classifier Based Learning Model for Intelligent Agents. In Whitely et al. [878], page 870. One page poster paper.
The objective of this study is to synthesize a learning model capable of successful and effective operation in hard-to-model environments. Here, we are presenting a structurally simple and functionally flexible model. The model follows the learning patterns experienced by the humans. The novelty of the adaptive model lies on the knowledge base, dual learning strategy, and flexible reasoning. The knowledge based is allowed to grow for as long as the agent lives. Learning is brought about by the interaction between two qualitatively different activities leaving long-term and short-term marks on the behavior of the agent.

[26]
Anthony J. Bagnall and G. D. Smith. Using an Adaptive Agent to Bid in a Simplified Model of the UK Market in Electricity. In Banzhaf et al. [32], page 774. One page poster paper.
This work has no abstract.

[27]
Anthony J. Bagnall and G. D. Smith. An Adaptive Agent Model for Generator Company Bidding in the UK Power Pool. In Proceedings of Artificial Evolution, 4th European Conference (AE'99), volume 1829 of Lecture Notes in Computer Science, pages 191-203. Springer, 2000.
This paper describes an autonomous adaptive agent model of the UK market in electricity, where the agents represent electricity generating companies. We briefly describe the UK market in electricity generation, then detail the simplifications we have made. Our current model consists of a single adaptive agent bidding against several nonadaptive agents. The adaptive agent uses a hierarchical agent structure with two Learning Classifier Systems to evolve market bidding rules to meet two objectives. We detail how the agent interacts with its environment, the particular problems this environment presents to the agent and the agent and classifier architectures we used in our experiments. We present the results and conclude that using our structure can improve performance.

[28]
A.J. Bagnall and Z.V. Zatuchna. On the classification of maze problems. In L. Bull and T. Kovacs, editors, Applications of Learning Classifier Systems, pages 307-316. Springer, 2005.

[29]
Anthony J. Bagnall. A Multi-Adaptive Agent Model of Generator Bidding in the UK Market in Electricity. In Whitely et al. [878], pages 605-612.
A model of the UK market in electricity combining key factors influencing generator bidding is proposed and a hierarchical multi-objective adaptive agent architecture using case based reasoning and learning classifier systems is described. Experimentation shows that the adaptive agents learn bidding strategies that have been observed in the real world, and that in some market scenarios the agents appear to be learning the benefits of cooperating to receive increased long term rewards. The potential of the adaptive agent model is illustrated by experimentation with an alternative market structure.

[30]
N. R. Ball. Towards the Development of Cognitive Maps in Classifier Systems. In Albrecht et al. [7], pages 712-718.
Classifier systems are well tested vehicles for implementing genetic algorithms in machine learning environments. This paper presents a novel system architecture that transforms a classifier system's knowledge representation from message-based structures to self-organizing neural networks. These networks have been integrated with a classifier system to produce a Hybrid Learning System (HLS) that exhibits adaptive behaviour when driven by low level environmental feedback. Problems are represented within HLS as objects characterized by environmental features. Objects controlled by the system have preset goals set against a subset of their features and the system has to achieve these goals by developing a behavioural repertoire that efficiently explores and exploits the problem environment. Three types of knowledge structures evolve during this adaptive process: a cognitive map of useful regularities within the environment (encoded in a self-organizing network); classifier behaviour calibrated against feature states and targets (encoded in a set of self-organizing feature maps); a population of complex behaviours (evolved from a gene pool supplied as part of the initial problem specification).

[31]
Sanghamitra Bandyopadhyay and Sankar K. Pal. Classification and Learning Using Genetic Algorithms. Applications in Bioinformatics and Web Intelligence. Natural Computing Series. Springer, 2007.

[32]
Wolfgang Banzhaf, Jason Daida, Agoston E. Eiben, Max H. Garzon, Vasant Honavar, Mark Jakiela, and Robert E. Smith, editors. Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-99). Morgan Kaufmann, 1999.

[33]
Flavio Baronti, Alessandro Passaro, and Antonina Starita. Post-processing clustering to decrease variability in xcs induced rulesets. In Tim Kovacs, Xavier LLòra, Keiki Takadama, Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson, editors, Learning Classifier Systems. International Workshops, IWLCS 2003-2005, Revised Selected Papers, volume 4399 of LNCS, pages 80-92. Springer, 2007.

[34]
Alwyn Barry. The Emergence of High Level Structure in Classifier Systems - A Proposal. Irish Journal of Psychology, 14(3):480-498, 1993.

[35]
Alwyn Barry. Hierarchy Formulation Within Classifiers System -- A Review. In Goodman et al. [379], pages 195-211.
Whilst the development of Learning Classifier Systems has produced excellent results in some fields of application, it has been widely noted that problems emerge when seeking to establish higher levels of knowledge (see Barry (1993) for relevant review). Tsotsos (1995) suggests that research into the operation of the Visual Cortex shows a hierarchical decomposition of processing more structured than a simple Subsumption Architecture arrangement. Whilst the LCS can provide both memory and planning by the use of tags an rule chains, it provides a flat rule space. Various approached have been taken to introducing structure to the LCS. We examine these approaches and identify three major lines of research: multiple interacting LCS, Single LCS with a structured population; and Structured Encoding of Rules. We illustrate that the first two of these areas have been interpreted quite differently, and seek to draw out common findings from the different approaches. We round off our examination of the area by a more detailed look at the work of Dorigo and Schnepf (1992), using a Hybrid Classifier System to examine the performance claims of Dorigo and Schnepf's architecture.

[36]
Alwyn Barry. Aliasing in XCS and the Consecutive State Problem: 1 -- Effects. In Banzhaf et al. [32], pages 19-26.
Whilst XCS (Wilson, 1998) has been shown to be more robust and reliable than previous LCS implementations (Kovacs, 1996, 1997), Lanzi(1997) identified a potential problem in the application of XCS to certain simple multi-step non Markovian environments. The 'Aliasing Problem' occurs when the environment provides the same message for two states in environmental positions that generate different constant payoffs. This prevents classifiers forming a correct payoff prediction for that message. This paper introduces a sub-class of the aliasing problem termed the 'Consecutive StateProblem' and uses the subclass to identify the effects of consecutive state aliasing on the learning of the State * Action * Payoff mapping within XCS. It is shown that aliasing states can prevent the formation of classifiers covering preceding states due to the trade-off of accuracy for match set occupancy made by the classifiers covering the aliasing states. This can be prevented by identifying a condition encoding which makes such match set ' piracy' improbable. However, under conditions of intense competition for population space where the classifier covering the aliased states cannot gain additional match set occupancy these classifiers will not be maintained within the population. Barry (1999) uses these findings to identify a solution to the Consecutive State Problem which is less heavyweight than the more general solution proposed by Lanzi (1997, 1998).

[37]
Alwyn Barry. Aliasing in XCS and the Consecutive State Problem: 2 -- Solutions. In Banzhaf et al. [32], pages 27-34.
The 'Aliasing Problem' within XCS (Wilson,1995, 1998), first identified by Lanzi (1997), does not only appear whenever the aliased states occur in separate environmental locations but also when they occur consecutively (Barry,1999). Lanzi (1997, 1998) introduced a mechanism that could solve the Aliasing Problem through the use of memory mechanisms within XCS (Wilson, 1995; Cliff and Ross,1994). Whilst this mechanism is a solution to the general problem of aliasing, it is a heavyweight solution. By limiting the scope of a solution to the Consecutive State Problem, which is shown to be a sub-problem of the Aliasing Problem, a simpler solution is proposed, and is shown to adequately address this problem. The application of a potential solution utilising explicit action duration identification is discussed and shown to be inadequate both as a solution to the Consecutive State Problem and for more general use within XCS.

[38]
Alwyn Barry. Specifying Action Persistence within XCS. In Whitely et al. [878], pages 50-57.
In investigating the Consecutive State Problem within XCS (Barry, 1999) it was suggested that a possible solution lay in allowing the XCS to persist with a single action over the aliased states. It was shown that this technique was sufficient to overcome the Consecutive State Problem as long as mechanisms were also provided which prevented the persistent application of 'Null Actions'. An alternative solution based on the work of Cobb and Grefenstette (1991) was discussed which sought to extend the action of each classifier so that each classifier could specify the duration that the action should be applied for. It was noted that this was an inadequate solution for the Consecutive State Problem because XCS would still explore the possibility of an action which persisted into but not beyond the aliased states. This work now applies these ideas to a number of non-aliased multiple step environments. It demonstrates that, given a suitable exploration strategy, action persistence can be utilised within XCS to enable the selection of a pathway to a reward state which entails the minimum number of different actions. It is also shown that a modification to the learning mechanism restores the ability of XCS to select the pathway to a reward state with the minimum number of steps whilst minimising the number of actions used.

[39]
Alwyn Barry. XCS Performance and Population Structure within Multiple-Step Environments. PhD thesis, Queens University Belfast, 2000.
Within Michigan-style Learning Classifier Systems based upon Holland's model (Holland et al 1986) support for learning in delayed-reward multiple-step environments was through the co-operation of classifiers within rule-chains. Despite the successful use of this LCS model in direct-reward environments (Wilson, 1985, 1987; Parodi and Bonelli, 1990; Holmes, 1997; Dorigo andColombetti, 1994), the application of LCS to delayed reward Markovian environments has been problematic. There is now a persuasive body of evidence that suggests that the use of strength as a fitness metric for the Genetic Algorithm (Kovacs and Kerber, 2000; Kovacs, 2000a), the use of rule-chains to establish multiple-step policies (Riolo, 1987b, 1989a; Forrest and Miller, 1990; Compiani et al, 1990), and the lack of mechanisms to encourage the development of co-operative populations (Booker, 1988; Smith, 1991; Smith and Golberg, 1991) all contribute to its inability within these environments. XCS (Wilson, 1995, 1998) presents solutions to each of these issues and initial results have shown the considerable promise of the approach (Kovacs, 1996, 1997; Lanzi, 1999c; Saxon and Barry, 1999a). In this work it is shown that whilst the XCS action-chaining mechanisms are effective for short action-chains, the combination of the use of discounted payoff and generalisation prevents XCS from learning optimal solutions in environments requiring even moderately sized action chains. In response it is hypothesised that the structuring of the solution, possibly hierarchically, can be used to reduce the required action chain length. A framework for hierarchical LCS research is proposed using a review of previous LCS hierarchical or structured approaches (Barry, 1993, 1996), and this work is compared to developments within the Reinforcement Learning community. Within a hierarchical solution low-level action chains may suffer when re-used if different payments are given to the action chains. An investigation into the Aliasing Problem (Lanzi, 1998a) reveals a subset of the problem, termed the Consecutive State Problem (Barry, 1999a), that will admit to a simple solution, which is empirically demonstrated (Barry, 1999b). It is shown that XCS is also able to learn the optimal state * action * duration * payoffmapping when a mechanism providing persistent actions is added (Barry, 2000), and that although this cannot be used as a solution to the aliasing problem it does provide a means of increasing the range of action chains. Two forms of preidentified hierarchical structures are introduced and it is shown that these allow multiple XCS instances to learn a hierarchical model that can be applied to operate successfully within environments requiring long action chains.

[40]
Dr. Alwyn Barry. A hierarchical xcs for long path environments. In Lee Spector, Erik D. Goodman, Annie Wu, W.B. Langdon, Hans-Michael Voigt, Mitsuo Gen, Sandip Sen, Marco Dorigo, Shahram Pezeshk, Max H. Garzon, and Edmund Burke, editors, Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2001), pages 913-920, San Francisco, California, USA, 7-11 July 2001. Morgan Kaufmann.

[41]
Alwyn M. Barry. The stability of long action chains in xcs. Journal of Soft Computing, 6(3-4):183-199, 2002.

[42]
Alwyn Barry. Limits in long path learning with XCS. In E. Cantú-Paz, J. A. Foster, K. Deb, D. Davis, R. Roy, U.-M. O'Reilly, H.-G. Beyer, R. Standish, G. Kendall, S. Wilson, M. Harman, J. Wegener, D. Dasgupta, M. A. Potter, A. C. Schultz, K. Dowsland, N. Jonoska, and J. Miller, editors, Genetic and Evolutionary Computation -- GECCO-2003, volume 2724 of LNCS, pages 1832-1843. Springer-Verlag, 2003.

[43]
Richard J. Bauer. Genetic Algorithms and Investment Strategies. Wiley Finance Editions. John Wiley & Sons, 1994.

[44]
Eric Baum and Igor Durdanovic. An Evolutionary Post Production System. In Proceedings of the International Workshop on Learning Classifier Systems (IWLCS-2000), in the Joint Workshops of SAB 2000 and PPSN 2000 [487]. Extended abstract.
Paper is an extended abstract

[45]
Eric Baum and Igor Durdanovic. An Artificial Economy of Post Production Systems. In Lanzi et al. [546], pages 3-20.

[46]
Eric Baum. Towards a model of intelligence as an economy of agents. Machine Learning, 35(2):155-185, 1999.
A market-based algorithm is presented which autonomously apportions complex tasks to multiple cooperating agents giving each agent the motivation of improving performance of the whole system. A specific model, called “The Hayek Machine” is proposed and tested on a simulated Blocks World (BW) planning problem. Hayek learns to solve more complex BW problems than any previous learning algorithm. Given intermediate reward and simple features, it has learned to efficiently solve arbitrary BW problems. The Hayek Machine can also be seen as a model of evolutionary economics.

[47]
Richard K. Belew and Stephanie Forrest. Learning and Programming in Classifier Systems. Machine Learning, 3:193-223, 1988.
Both symbolic and subsymbolic models contribute important insights to our understanding of intelligent systems. Classifier systems are low-level learning systems that are also capable of supporting representations at the symbolic level. In this paper, we explore in detail the issues surrounding the integration of programmed and learned knowledge in classifier-system representations, including comprehensibility, ease of expression, explanation predictability, robustness, redundancy, stability, and the use of analogical representations. We also examine how these issues speak to the debate between symbolic and subsymbolic paradigms. We discuss several dimensions for examining the tradeoffs between programmed and learned representations, and we propose an optimization model for constructing hybrid systems that combine positive aspects of each paradigm.

[48]
Richard K. Belew and Michael Gherrity. Back Propagation for the Classifier System. In Schaffer [718], pages 275-281.
Connectionist networks and the Classifier System (CFS) provide two important examples of massively parallel knowledge representation systems for which successful learning algorithms have been developed. We present a construction that shows how the behavior of a large class of connectionist networks can be reproduced in a CFS. We then use this construction derive a version of the Rumelhart et al.'s back-propagation algorithm for the CFS, and conclude with remarks about critical differences between connectionist networks and the CFS that this analysis highlights.

[49]
L. Beltrametti, R. Fiorentini, L. Marengo, and R. Tamborini. A learning-to-forecast experiment on the foreign exchange market with a classifier system. Journal of Economic Dynamics and Control, 21(8):1543-1575, 1997.
This paper reports on an experiment of learning and forecasting on the foreign exchange market by means of an Artificial Intelligence methodology (a 'Classifier System') which simulates learning and adaptation in complex and changing environments. The experiment has been run for two different exchange rates, the US dollar-D mark rate and the US dollar-yen rate, representative of two possibly different market environments. A fictitious ''artificial agent'' is first trained on a monthly data base from 1973 to 1990, and then tested out-of-sample from 1990 to 1992. Its forecasting performance is then compared with the performance of decision rules which follow the prescription of various economic theories on exchange rate behaviour, and the performance of forecasts given by VAR estimations of the exchange-rate's determinants.

[50]
E. Bernadó, X., Llorà, and J.M. Garrell. XCS and GALE: a Comparative Study of Two Learning Classifier Systems with Six Other Learning Algorithms on Classification Tasks. In Proceedings of the 4th International Workshop on Learning Classifier Systems (IWLCS-2001), pages 337-341, 2001. Short version published in Genetic and Evolutionary Compution Conference (GECCO2001).

[51]
Ester Bernadó, Xavier Llorà, and Josep M. Garrell. XCS and GALE: A Comparative Study of Two Learning Classifier Systems on Data Mining. In Lanzi et al. [546], pages 115-132.

[52]
Ester Bernadó-Mansilla and Josep M. Garrell-Guiu. Accuracy-Based Learning Classifier Systems: Models, Analysis and Applications to Classification Tasks. Evolutionary Computation, 11(3):209-238, 2003.

[53]
Ester Bernadó-Mansilla and T.K. Ho. Domain of competence of XCS classifier system in complexity measurement space. IEEE Trans. Evolutionary Computation, 9(1):82-104, 2005.

[54]
Hugues Bersini and Francisco J. Varela. Hints for Adaptive Problem Solving Gleaned From Immune Networks. In Schwefel and Männer [728], pages 343-354.

[55]
Janine Beunings, Ludwig Bölkow, Bernd Heydemann, Biruta Kresling, Claus-Peter Lieckfeld, Claus Mattheck, Werner Nachtigall, Josef Reichholf, Bertram J. Schmidt, Veronika Straaß, and Reinhard Witt. Bionik: Natur als Vorbild. WWF Dokumentationen. PRO FUTURA Verlag, 1993.

[56]
J. Biondi. Robustness and evolution in an adaptive system application on classification task. In Albrecht et al. [7], pages 463-470.
In this paper, we proposed an approach to a single-step Classifier System, in which the useful population is built by progressively specializing classifiers. It has been applied to a classification task in a medical domain. To permit the system to explore alternatives without making decisions earlier in learning stages, all the classifiers that might be selected are triggered and receive the resulting reward corresponding to their action. The payoff function involves the classifier's performance, its specificity and the system's performance (its robustness). Genetic operators are activated with a probability which depends on the system's robustness. During the test stages, no further learning takes place and the system's performance is measured by the percentage of correct classification made on the second set of examples. When the measure of performance is the highest, the population is stabilized and contains the correct classifiers (the payoff function and genetic operators have no more effect on classifiers). This approach achieves convergency more quickly and makes it possible to have a final accurate population without over-specializing.

[57]
Andrea Bonarini and Filippo Basso. Learning to compose fuzzy behaviors for autonomous agents. Int. Journal of Approximate Reasoning, 17(4):409-432, 1997.

[58]
Andrea Bonarini, Marco Dorigo, V. Maniezzo, and D. Sorrenti. AutonoMouse: An Experiment in Grounded Behaviors. In Proceedings of GAA91 -- Second Italian Workshop on Machine Learning, Bari, Italy, 1991.

[59]
Andrea Bonarini, Claudio Bonacina, and Matteo Matteucci. Fuzzy and crisp representation of real-valued input for learning classifier systems. In Wu [923], pages 228-235.
We discuss some issues concerning the application of learning classifier systems to real-valued applications. In particular, we focus on the possibility of classifying data by crisp and fuzzy intervals, showing the effect of their granularity on the learning performance. We introduce the concept of sensorial cluster and we discuss the difference between cluster aliasing and perceptual aliasing. We show the impact of different choices on the performance of both crisp and fuzzy learning classifier systems applied to a mobile, autonomous, robotic agent.

[60]
Andrea Bonarini, Claudio Bonacina, and Matteo Matteucci. Fuzzy and Crisp Representations of Real-valued Input for Learning Classifier Systems. In Lanzi et al. [544], pages 107-124.
We discuss some issues concerning the application of learning classifier systems to real-valued applications. In particular, we focus on the possibility of classifying data by crisp and fuzzy intervals, showing the effect of their granularity on the learning performance. We introduce the concept of sensorial cluster and we discuss the difference between cluster aliasing and perceptual aliasing. We show the impact of different choices on the performance of both crisp and fuzzy learning classifier systems applied to a mobile, autonomous, robotic agent.

[61]
Andrea Bonarini. ELF: Learning Incomplete Fuzzy Rule Sets for an Autonomous Robot. In Hans-Jürgen Zimmermann, editor, First European Congress on Fuzzy and Intelligent Technologies -- EUFIT'93, volume 1, pages 69-75, Aachen, D, September 1993. Verlag der Augustinus Buchhandlung.

[62]
Andrea Bonarini. Evolutionary Learning of General Fuzzy Rules with Biased Evaluation Functions: Competition and Cooperation. Proc. 1st IEEE Conf. on Evolutionary Computation, pages 51-56, 1994.

[63]
Andrea Bonarini. Learning Behaviors Represented as Fuzzy Logic Controllers. In Hans-Jürgen Zimmermann, editor, Second European Congress on Intelligent Techniques and Soft Computing - EUFIT'94, volume 2, pages 710-715, Aachen, D, 1994. Verlag der Augustinus Buchhandlung.

[64]
Andrea Bonarini. Extending Q-learning to Fuzzy Classifier Systems. In Marco Gori and Giovanni Soda, editors, Proceedings of the Italian Association for Artificial Intelligence on Topics in Artificial Intelligence, volume 992 of LNAI, pages 25-36, Berlin, 1995. Springer.

[65]
Andrea Bonarini. Delayed Reinforcement, Fuzzy Q-Learning and Fuzzy Logic Controllers. In Herrera and Verdegay [423], pages 447-466.
In this paper, we discuss situations arising with reinforcement learning algorithms, when the reinforcement is delayed. The decision to consider delayed reinforcement is typical in many applications, and we discuss some motivations for it. Then, we summarize Q-Learning, a popular algorithm to deal with delayed reinforcement, and its recent extensions to use it to learn fuzzy logic structures (Fuzzy Q-Learning). Moreover, we present how a reinforcement learning algorithm we have developed in the past (ELF - Evolutionary Learning of Fuzzy rules) implements an extension of the popular Q-Learning algorithm for the distribution of delayed reinforcement when the controller to be learnt is a Fuzzy Logic Controller (FLC). Finally, we present some examples of the application of ELF to learning FLCs that implement behaviors for an autonomous agent.

[66]
Andrea Bonarini. Delayed Reinforcement, Fuzzy Q-Learning and Fuzzy Logic Controllers. In F. Herrera and J. L. Verdegay, editors, Genetic Algorithms and Soft Computing, (Studies in Fuzziness, 8), pages 447-466, Berlin, D, 1996. Physica-Verlag.

[67]
Andrea Bonarini. Evolutionary Learning of Fuzzy rules: competition and cooperation. In W. Pedrycz, editor, Fuzzy Modelling: Paradigms and Practice, pages 265-284. Norwell, MA: Kluwer Academic Press, 1996. ftp://ftp.elet.polimi.it/pub/Andrea.Bonarini/ELF/ELF-Pedrycz.ps.gz.
We discuss the problem of learning fuzzy rules using Evolutionary Learning techniques, such as Genetic Algorithms and Learning Classifier Systems. We present ELF, a system able to evolve a population of fuzzy rules to obtain a sub-optimal Fuzzy Logic Controller. ELF tackles some of the problems typical of the Evolutionary Learning approach: competition and cooperation between fuzzy rules, evolution of general fuzzy rules, imperfect reinforcement programs, fast evolution for real-time applications, dynamic evolution of the focus of the search. We also present some of the results obtained from the application of ELF to the development of Fuzzy Logic Controllers for autonomous agents and for the classical cart-pole problem.

[68]
Andrea Bonarini. Anytime learning and adaptation of fuzzy logic behaviors. Adaptive Behavior, 5(3-4):281-315, 1997.

[69]
Andrea Bonarini. Reinforcement Distribution to Fuzzy Classifiers. In Proceedings of the IEEE World Congress on Computational Intelligence (WCCI) -- Evolutionary Computation, pages 51-56. IEEE Computer Press, 1998.

[70]
Andrea Bonarini. Comparing reinforcement learning algorithms applied to crisp and fuzzy learning classifier systems. In Banzhaf et al. [32], pages 52-59.
We have implemented a tool to compare different modules of Reinforcement Learning algorithms applied to Learning Classifier Systems (LCS). We focus on three main classes of modules: credit assignment modules, exploration policies, and evolutionary strategies. For each class we have implemented many of the proposals we can find in the literature and also some new algorithms that we have designed. In this paper, we present the results of the application of our tool to both fuzzy and crisp LCSs that learn behaviors for simulated autonomous agents. Fuzzy LCSs can be considered a successful approach to cope with real-valued input and output in a real environment. A lot of investigations can be done with this tool in this experimental setting. This paper is focused on the comparison among different credit assignment algorithms and on their performance in learning both crisp and fuzzy models. Our experiments show that the more complex credit assignment algorithms (such as, for instance, the TD(lambda) generally have better performance than the more basic (such as Q-learning or Bucket Brigade) also when applied to LCSs. Moreover, fuzzy LCSs seem to require a larger computational effort, but also show more robustness.

[71]
Andrea Bonarini. An Introduction to Learning Fuzzy Classifier Systems. In Lanzi et al. [544], pages 83-104.
We present a class of Learning Classifier Systems that learn fuzzy rule-based models, instead of interval-based or Boolean models. We discuss some motivations to consider Learning Fuzzy Classifier Systems (LFCS) as a promising approach to learn mappings from real-valued input to real-valued output, basing on data interpretation implemented by fuzzy sets. We describe some of the approaches explicitly or implicitly referring to this research area, presented in literature since the beginning of the last decade. We also show how the general LFCS model can be considered as a framework for a wide range of systems, each implementing in a different way the modules composing the basic architecture. We also mention some of the applications of LFCS presented in literature, which show the potentialities of this type of systems. Finally, we introduce a general methodology to extend reinforcement distribution algorithms usually not designed to learn fuzzy models. This opens new application possibilities.

[72]
Pierre Bonelli and Alexandre Parodi. An Efficient Classifier System and its Experimental Comparison with two Representative learning methods on three medical domains. In Booker and Belew [74], pages 288-295.
In this paper, we describe a Classifier System, Newboole, and we present its experimental comparison with two widely used learning algorithms, CN2 (logic reduction system) and Back Propagation (neural net), on three medical domains. The experimental results, obtained in the context of learning from preclassified examples, demonstrate two main points: firstly, that all three systems perform very closely on the induction tasks with a slight advantage for the Back Propagation algorithm. Secondly, that a Classifier System can provide comprehensive solutions in the form of a reasonable number of ``symbolic'' decision rules, which is not the case using Back Propagation.

[73]
Pierre Bonelli, Alexandre Parodi, Sandip Sen, and Stewart W. Wilson. NEWBOOLE: A Fast GBML System. In International Conference on Machine Learning, pages 153-159, San Mateo, California, 1990. Morgan Kaufmann.
Genetics based machine learning systems are considered by a majority of machine learners as slow rate learning systems. In this paper, we propose an improvement of Wilson's classifier system BOOLE that show how Genetics based machine learning systems learning rates can be greatly improved. This modification consists in a change of the reinforcement component. We then compare the respective performances of this modified BOOLE, called NEWBOOLE, and a neural net using back propagation on a difficult boolean learning task, the multiplexer function. The results of this comparison show that NEWBOOLE obtains significantly faster learning rates.

[74]
Lashon B. Booker and Richard K. Belew, editors. Proceedings of the 4th International Conference on Genetic Algorithms (ICGA91). Morgan Kaufmann, July 1991.

[75]
Lashon B. Booker, David E. Goldberg, and John H. Holland. Classifier Systems and Genetic Algorithms. Artificial Intelligence, 40:235-282, 1989.
Classifier systems are massively parallel, message-passing, rule-based systems that learn through credit assignment (the bucket brigade algorithm) and rule discovery (the genetic algorithm). They typically operate in environments that exhibit one or more of the following characteristics: (1) perpetually novel events accompanied by large amounts of noisy or irrelevant data; (2) continual, often real-time requirements for action; (3) implicitly or inexactly defined goals; and (4) sparse payoff or reinforcement obtainable only through long action sequences. Classifier systems are designed to absorb new information continuously from such environments, devising sets of competing hypotheses (expressed as rules) without disturbing significantly capabilities already acquired. This paper reviews the definition, theory, and extant applications of classifier systems, comparing them with other machine learning techniques, and closing with a discussion of advantages, problems, and possible extensions of classifier systems.

[76]
Lashon B. Booker, Rick L. Riolo, and John H. Holland. Learning and Representation in Classifier Systems. In Vassant Honavar and Leonard Uhr, editors, Artificial Intelligence and Neural Networks, pages 581-613. Academic Press, 1994.

[77]
Lashon B. Booker. Intelligent Behavior as an Adaptation to the Task Environment. PhD thesis, The University of Michigan, 1982.

[78]
Lashon B. Booker. Improving the performance of genetic algorithms in classifier systems. In Grefenstette [389], pages 80-92.
Classifier systems must continuously infer useful categories and other generalizations in the form of classifier taxa from the steady stream of messages received and transmitted. This paper describes ways to use the genetic algorithm more effectively in discovering such patterns. Two issues are addressed. First, a flexible criterion is advocated for deciding when a message matches a classifier taxon. This is shown to improve performance over a wide range of categorization problems. Second, a restricted mating policy and crowding algorithm are introduced. These modifications lead to the growth and dynamic management of subpopulations correlated with the various pattern categories in the environment.

[79]
Lashon B. Booker. Classifier Systems that Learn Internal World Models. Machine Learning, 3:161-192, 1988.
Most classifier systems learn a collection of stimulus-response rules, each of which directly acts on the problem-solving environment and accrues strength proportional to the overt reward expected from the behavioral sequences in which the rule participates. GOFER is an example of a classifier system that builds an internal model of its environment, using rules to represent objects, goals, and relationships. The model is used to direct behavior, and learning is triggered whenever the model proves to be an inadequate basis for generating behavior in a given situation. This means that overt external rewards are not necessarily the only or the most useful source of feedback for inductive change. GOFER is tested in a simple two-dimensional world where it learns to locate food and avoid noxious stimulation.

[80]
Lashon B. Booker. Triggered rule discovery in classifier systems. In Schaffer [718], pages 265-274.
Recent work by Quinlan (1988) and Grefenstette (1988) has raised doubts about the ability of classifier systems to learn concepts or long temporal sequences of rules efficiently. This paper shows how the use of learning triggers can greatly increase the performance of a classifier system to the point that it compares much more favorably with other learning systems. We introduce a new classifier system called Gofer-1 that demonstrates how to trigger rule discovery in an effective manner.

[81]
Lashon B. Booker. Instinct as an Inductive Bias for Learning Behavioral Sequences. In Meyer and Wilson [613], pages 230-237.

[82]
Lashon B. Booker. Representing Attribute-Based Concepts in a Classifier System. In Rawlins [674], pages 115-127.
Legitimate concerns have been raised about the expressive adequacy of the classifier language. This paper shows that many of those concerns stem from the inadequacies of the binary encodings typically used with classifier systems, not the classifier language per se. In particular, we describe some straightforward binary encodings for attribute-based instance spaces. These encodings give classifier systems the ability to represent ordinal and nominal attributes as expressively as most symbolic machine learning systems, without sacrificing the building blocks required by the genetic algorithm.

[83]
Lashon B. Booker. Viewing Classifier Systems as an Integrated Architecture. In Collected Abstracts for the First International Workshop on Learning Classifier System (IWLCS-92) [486]. October 6-8, NASA Johnson Space Center, Houston, Texas.
Paper is an extended abstract

[84]
Lashon B. Booker. Do We Really Need to Estimate Rule Utilities in Classifier Systems?. In Wu [923], pages 236-241.
Classifier systems have traditionally used explicit measures of utility (strength, predicted payoff, accuracy, etc.) to quantify the performance and fitness of classifier rules. Much of the effort in designing and implementing these systems has focused on getting these utilities ``right''. One alternative worth exploring is the idea of using endogenous fitness; that is, reinforcing successful performance with ``resources'' that rules need in order to reproduce. Under this regime, the best rules are those that accumulate the most resources over their lifetime and, consequently, have the most offspring. This paper describes a classifier system designed along these lines. Rules have no associated utility measure, just a resource reservoir. When enough resources have been accumulated, the rule reproduces and the reservoir is reduced. Preliminary tests of this system on the multiplexor problem show that it performs as well as utility based classifier systems such as XCS.

[85]
Lashon B. Booker. Classifier systems, endogenous fitness, and delayed reward: A preliminary investigation. In Proceedings of the International Workshop on Learning Classifier Systems (IWLCS-2000), in the Joint Workshops of SAB 2000 and PPSN 2000 [487]. Extended abstract.
Paper is an extended abstract

[86]
Lashon B. Booker. Do We Really Need to Estimate Rule Utilities in Classifier Systems? In Lanzi et al. [544], pages 125-142.
Classifier systems have traditionally used explicit measures of utility (strength, predicted payoff, accuracy, etc.) to quantify the performance and fitness of classifier rules. Much of the effort in designing and implementing these systems has focused on getting these utilities ``right''. One alternative worth exploring is the idea of using endogenous fitness; that is, reinforcing successful performance with ``resources'' that rules need in order to reproduce. Under this regime, the best rules are those that accumulate the most resources over their lifetime and, consequently, have the most offspring. This paper describes a classifier system designed along these lines. Rules have no associated utility measure. Instead, each rule has one or more reservoirs that can be used to store resources. When enough resources have been accumulated, a rule utilizes some of its resources to reproduce and the reservoir level is reduced accordingly. Preliminary tests of this system on the multiplexor problem show that it performs as well as utility-based classifier systems such as XCS.

[87]
Lashon B. Booker. Classifier systems, endogenous fitness, and delayed rewards: A preliminary investigation. In Lee Spector, Erik D. Goodman, Annie Wu, W.B. Langdon, Hans-Michael Voigt, Mitsuo Gen, Sandip Sen, Marco Dorigo, Shahram Pezeshk, Max H. Garzon, and Edmund Burke, editors, Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2001), pages 921-926, San Francisco, California, USA, 7-11 July 2001.
Previous work has shown the potential advantages of using endogenous fitness schemes in classifier systems. The basic idea behind endogenous fitness is to reinforce successful system performance with ``resources'' that rules need in order to reproduce. Instead of storing explicit quantitative estimates of performance, each rule has one or more reservoirs that are used to store resources. When enough resources have been accumulated, a rule utilizes some of its resources to reproduce and the reservoir level is reduced accordingly. This paper extends this concept to accommodate environments having delayed rewards. Reinforcement learning techniques for solving average-reward Markovian decision processes are combined with a simple endogenous fitness scheme in a classifier system. We describe initial tests of this approach on state space search problems used in previous classifier system studies.

[88]
Lashon B. Booker. A new approach to encoding actions in classifier systems. 2001.
The classifier system framework is a general-purpose approach to learning and representation designed to exhibit non-brittle behavior in complex, continually varying environments. Broadly speaking, classifier systems are expected to avoid brittle behavior because they implement processes that build and refine models of the environment. One of the most important of these processes is categorization. As Holland [5] has pointed out (p. 598) "Categorization is the system's major weapon for combating the enviironment's perpetual novelty. The system must readily generate categories for input messages, and it must be able to generate categories relevant to its internal processes". Research in classifier systems has focused almost exclusively on finding generalizations for input messages. However, generalizations of actions will also be required in order to build effective models of the environment. This paper introduces a new encoding for actions in classifier rules that lends itself to representing abstract actions.

[89]
Lashon B. Booker. Adaptive value function approximations in classifier systems. In GECCO '05: Proceedings of the 2005 workshops on Genetic and evolutionary computation, pages 90-91. ACM, 2005.

[90]
Lashon B. Booker. Approximating value functions in classifier systems. In L. Bull and T. Kovacs, editors, Foundations of Learning Classifier Systems, volume 183/2005 of Studies in Fuzziness and Soft Computing, pages 45-61. Springer, 2005.
While there has been some attention given recently to the issues of function approximation using learning classifier systems (e.g. [13, 3]), few studies have looked at the quality of the value function approximation computed by a learning classifier system when it solves a reinforcement learning problem [1, 8]. By contrast, considerable attention has been paid to this issue in the reinforcement learning literature [12]. One of the fundamental assumptions underlying algorithms for solving reinforcement learning problems is that states and state-action pairs have well-defined values that can be computed and used to help determine an optimal policy. The quality of those approximations is a critical factor in determining the success of many algorithms in solving reinforcement learning problems.

[91]
Lashon B. Booker. Adaptive value function approximations in classifier systems. In Tim Kovacs, Xavier LLòra, Keiki Takadama, Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson, editors, Learning Classifier Systems. International Workshops, IWLCS 2003-2005, Revised Selected Papers, volume 4399 of LNCS, pages 219-238. Springer, 2007.

[92]
Gavin Brown, Tim Kovacs, and James Marshall. UCSpv: Principled Voting in UCS Rule Populations. In Hod Lipson et al., editor, GECCO'07: the Genetic and Evolutionary Computation Conference, pages 1774-1781. ACM, 2007.

[93]
Will Browne, Karen Holford, Carolynne Moore, and John Bullock. The implementation of a learning classifier system for parameter identification by signal processing of data from steel strip downcoilers. In A. T. Augousti, editor, Software in Measurement Conference, number 96-147. IEE Computer and Control Division, 1996.
The aim of this project is to improve the quality and consistency of coiling in a steel hot strip mill. Signal processing will be used to gather information on the parameter characteristics of the mill downcoilers as an aid to operator and engineering decision making. The Artificial Intelligence (AI) paradigm of Learning Classifier Systems (LCS) is proposed for the signal processing.

[94]
Will Browne, Karen Holford, Carolynne Moore, and John Bullock. A Practical Application of a Learning Classifier System for Downcoiler Decision Support in a Steel Hot Strip Mill. Ironmaking and Steelmaking, 25(1):33-41, 1997. Engineering Doctorate Seminar '97. Swansea, Wales, Sept. 2nd, 1997.
The aim of this project is to improve the quality and consistency of coiling in a hot strip mill at British Steel Strip Products, Integrated Works. The Artificial Intelligence paradigm of Learning Classifier Systems is proposed for the processing of plant data. The stochastic computational technique of LCS will produce off-line rules to aid operator and engineering decision making. These rules link the plant inputs (plant condition, strip properties and associated variables) to coil outputs (presentation - including telescoping and pinching) in a form that is capable of being verified and validated. This is central to the initial operation, where on-line data will produce off-line rules that are critically evaluated by a human operator before implementation. Improvements to a basic LCS, that allow operation on industrial data, are detailed. Initial experimental results show that the technique of LCS has the potential to become a very useful tool for processing industrial data. Improvements in availability, coil presentation and ultimately customer satisfaction will result in a cost benefit to British Steel Plc.

[95]
Will Browne, Karen Holford, Carolynne Moore, and John Bullock. A Practical Application of a Learning Classifier System in a Steel Hot Strip Mill. In Smith et al. [770], pages 611-614.
The aim of this project is to improve the quality and consistency of coiling in a steel hot strip mill at British Steel Strip Products, Integrated Works. The artificial intelligence paradigm of learning classifier systems (LCS) is proposed for the processing of plant data. Improvements to a basic LCS, that allow operation on industrial data, are detailed. Initial experimental results show that the technique of LCS has the potential to become a very useful for processing industrial data. The stochastic computational technique will produce off-line rules to aid operator and engineering decision making. Improvements in availability, coil presentation and ultimately customer satisfaction will result in cost benefits to British Steel Plc.

[96]
Will Browne, Karen Holford, and Carolynne Moore. An Industry-Based Development of the Learning Classifier System Technique. In I. C. Parmee, editor, Adaptive Computing in Design and Manufacture (ACDM 2000), pages 199-210. Springer-Verlag, 1999.
This paper describes the development of an Industrial Learning Classifier System for application in the steel industry. The real domain problem was the prediction and diagnosis of product quality issues in a Steel Hot Strip Mill. The properties of the data from this environment include multi-modality (much parameter interaction), poor separation between fault levels and high dimensionality (many parameters). The method to develop the Learning Classifier System technique, based on deterministic simulated data, is presented. The advances made in the technique that assist in its functionality in this type of industrial environments are given. The novel methods developed are core to the Learning Classifier System technique and are not 'fixes' for given problems. They address the fitness measure, encoding alphabet, population scope, phases of training, genetic operators, life limits and removal of taxation schemes. These improvements allow the industrial LCS to function correctly in the simulated domain. Encouraging results from diagnosis of real data are presented; however, further work is needed for greater accuracy and to allow the prediction function to be used on-line. Learning Classifier Systems represent a potentially useful tool that combines the transparency of symbolic approaches (such as Decision Trees) with the learning ability of connectionist approaches (such as Artificial Neural Networks) to machine learning.

[97]
Will Browne, Karen Holford, Carolynne Moore, and John Bullock. An Industrial Learning Classifier System: The Importance of Pre-Processing Real Data and Choice of Alphabet. Engineering Applications of Artificial Intelligence, 13(1):25-36, 2000.
Learning Classifier Systems (LCS) have received considerable attention in the research community, yet few have been applied in practice. This paper describes the development of a LCS for monitoring data produced by a hot strip mill at British Steel Strip Products. The problems associated with applying a theoretical technique in a practical environment are discussed with particular attention being given to the pre processing of voluminous real data. The appropriate choice of alphabet for the LCS is also discussed, with a comparison of two alphabets, namely the ternary alphabet and the real numbered approach, being included.

[98]
Will Browne. The Development of an Industrial Learning Classifier System for Application to a Steel Hot Strip Mill. PhD thesis, University of Wales, Cardiff, 1999.
The search for continual improvement in industry has identified the resource of plant data. Learning Classifier Systems (LCSs) were anticipated to be capable of exploiting plant data for cost benefit. The initial LCS performed poorly on simulated data as complex domains caused greedy and unstable performance. The aim of the project was to develop the LCS technique into a robust tool for application to industry. The utilisation of performance methods where appropriate, was achieved by splitting the LCS rule-base into the three training phases. Another advance was to separate the fitness measure into component functions, thus enabling optimal control of the LCS. Combining rule accuracy with the degree of domain match allowed the rule discovery to evenly search all niches of the rule base, whilst still exerting a generalisation pressure. Motivated by experiments with real data a morphing genetic operator to improve search rates, an evaluation limit to enable graceful improvements of hierarchies and a child limit to prevent convergence to a sub-optimum performance level were created. Implementing a real numbered alphabet simplified rule interpretation, automatically adjusted condition ranges to avoid aliasing and formed correct rule boundaries. Further simplification of the internal parameters removed all taxation, which greatly simplified the use of the industrial LCS. Optimum prediction and correct diagnosis of the complex simulated data was achieved. The real data sets from British Steel governed plant conditions and output quality. Diagnosis of the input-output relationships that could assist operators, engineers and managers was possible and contained encouraging results. However, inadequacies in data quality and the technique allowed only 80 prediction, which was insufficient confidence for plant predictive use. Although the LCS technique is still not fully developed, the effective learning, transparency and co-operation in rules has many potential benefits for industry.

[99]
Larry Bull and Terence C. Fogarty. Coevolving Communicating Classifier Systems for Tracking. In Albrecht et al. [7], pages 522-527.
In this paper we suggest a general approach to using the genetic algorithm (GA)[1] to evolve complex control systems. It has been shown [2] that although the GA may be used to evolve simple controllers, it is not able to cope with the evolution of controllers for more complex problems. We present an architecture of co-evolving communicating classifier systems [3] as a general solution to this, where the only restriction is that each classifier system is responsible for one simple behaviour. Thus the ecology of sub-problems evolves its own organisational structure at the same time its constituents evolve their solutions. Whether this structure ends up as a democratic soup, a hierarchy, or something in between, is determined by co-evolution rather than prescribed a priori by a creator. We use the trail following ``tracker task'' to compare the performance of a single classifier, responsible for the control of the whole system, evolved for this task with the performance of a co-evolved controller using our approach. The resulting interactions of the classifier system are also examined.

[100]
Larry Bull and Terence C. Fogarty. Evolving Cooperative Communicating Classifier Systems. In A. V. Sebald and L. J. Fogel, editors, Proceedings of the Third Annual Conference on Evolutionary Programming, pages 308-315, 1994.

[101]
Larry Bull and Terence C. Fogarty. Parallel Evolution of Communicating Classifier Systems. In Proceedings of the 1994 IEEE Conference on Evolutionary Computing, pages 680-685. IEEE, 1994.

[102]
Larry Bull and Terence C. Fogarty. Evolutionary Computing in Cooperative Multi-Agent Systems. In Sandip Sen, editor, Proceedings of the 1996 AAAI Symposium on Adaptation, Coevolution and Learning in Multi-Agent Systems, pages 22-27. AAAI, 1996.

[103]
Larry Bull and Terence C. Fogarty. Evolutionary Computing in Multi-Agent Environments: Speciation and Symbiogenesis. In H-M. Voigt, W. Ebeling, I. Rechenberg, and H.-P. Schwefel, editors, Parallel Problem Solving from Nature -- PPSN IV, pages 12-21. Springer-Verlag, 1996.

[104]
Larry Bull and O. Holland. Internal and External Representations: A Comparison in Evolving the Ability to Count. In Proceedings of the First Annual Society for the Study of Artificial Intelligence and Simulated Behaviour Robotics Workshop, pages 11-14, 1994.

[105]
Larry Bull and Jacob Hurst. Self-Adaptive Mutation in ZCS Controllers. In Proceedings of the EvoNet Workshops - EvoRob 2000, pages 339-346. Springer, 2000.

[106]
Larry Bull and Jacob Hurst. ZCS redux. Evolutionary Computation, 10(2):185-205, 2002.

[107]
Larry Bull and Tim Kovacs, editors. Foundations of Learning Classifier Systems, volume 183 of Studies in Fuzziness and Soft Computing. Springer, 2005.

[108]
Larry Bull and Toby O'Hara. Accuracy-based neuro and neuro-fuzzy classifier systems. In W. B. Langdon, E. Cantú-Paz, K. Mathias, R. Roy, D. Davis, R. Poli, K. Balakrishnan, V. Honavar, G. Rudolph, J. Wegener, L. Bull, M. A. Potter, A. C. Schultz, J. F. Miller, E. Burke, and N. Jonoska, editors, GECCO 2002: Proceedings of the Genetic and Evolutionary Computation Conference, pages 905-911. Morgan Kaufmann Publishers, 9-13 July 2002.

[109]
L. Bull and R. Preen. On dynamical genetic programming: Random boolean networks in learning classifier systems. In To appear in Proceedings of the EvoNet Workshops: EuroGP. Springer, 2009.

[110]
Larry Bull and Matt Studley. Consideration of multiple objectives in neural learning classifier systems. In H.-P. Schwefel J.-J. Merelo Guervós, P. Adamidis, H.-G. Beyer, J.-L. Fernández-Villacañas, editor, Parallel Problem Solving from Nature - PPSN VII, 7th International Conference, Granada, Spain, September 7-11, 2002. Proceedings, number 2439 in Lecture Notes in Computer Science, LNCS, page 549 ff. Springer-Verlag, 2002.

[111]
Larry Bull, Terence C. Fogarty, S. Mikami, and J. G. Thomas. Adaptive Gait Acquisition using Multi-agent Learning for Wall Climbing Robots. In Automation and Robotics in Construction XII, pages 80-86, 1995.

[112]
Larry Bull, Terence C. Fogarty, and M. Snaith. Evolution in Multi-agent Systems: Evolving Communicating Classifier Systems for Gait in a Quadrupedal Robot. In Eshelman [298], pages 382-388.

[113]
Larry Bull, Jacob Hurst, and Andy Tomlinson. Mutation in Classifier System Controllers. In et al. [299], pages 460-467.
The use and benefits of self-adaptive mutation operators are well-known within evolutionary computing. In this paper we examine the use of self-adaptive mutation in Michigan-style Classifier Systems with the aim of improving their performance as controllers for autonomous mobile robots. Initially, we implement the operator in the ZCS classifier and examine its performance in two ``animat'' environments. It is shown that, although no significant increase in performance is seen over results presented in the literature using a fixed rate of mutation, the operator adapts to approximately this rate regardless of the initial range. The same operator is then implemented in the more sophisticated XCS classifier, with its performance examined on another animat task. Again it is shown that no real improvements in performance are obtained over previous results with a fixed mutation rate, but that the operator adapts to a suitable rate.

[114]
Larry Bull, Pier Luca Lanzi, and Wolfgang Stolzmann (guest editors). Journal of soft computing, special issue on learning classifier systems, 6(3-4), 2002.

[115]
Larry Bull, Dave Wyatt, and Ian Parmee. Towards the use of XCS in interactive evolutionary design. In W. B. Langdon, E. Cantú-Paz, K. Mathias, R. Roy, D. Davis, R. Poli, K. Balakrishnan, V. Honavar, G. Rudolph, J. Wegener, L. Bull, M. A. Potter, A. C. Schultz, J. F. Miller, E. Burke, and N. Jonoska, editors, GECCO 2002: Proceedings of the Genetic and Evolutionary Computation Conference, page 951. Morgan Kaufmann Publishers, 2002.

[116]
L. Bull, P.L. Lanzi, and T. O'Hara. Anticipation mappings for learning classifier systems. In Proceedings of the 2007 congress on evolutionary computation (CEC2007), pages 2133-214. IEEE, 2007.

[117]
L. Bull, M. Studley, T. Bagnall, and I. Whittley. On the use of rule-sharing in learning classifier system ensembles. IEEE Trans. Evolutionary Computation, 11:496-502, 2007.

[118]
Larry Bull. Artificial Symbiology: evolution in cooperative multi-agent environments. PhD thesis, University of the West of England, 1995.
Nature is full of examples of both inter and intra species; from the workings of ant colonies to the cleaning symbiosis seen between the Pederson shrimp and the fish of the Bahamas. The fields of Artificial Intelligence and Artificial Life have consequently focused on these phenomenon as a means of dealing with complex systems in which agents must cooperate to achieve certain goals. In this thesis we examine the performance of the Genetic Algorithm when applied to systems of this type. That is, we examine the use of Evolutionary Computing techniques within cooperative multiagent environments. In the process we investigate some aspects of the natural phenomenon of symbiosis on which we base many elements of the work, in particular conditions under which various aspects of symbiotic associations occur. In extending the Genetic Algorithm to cooperative multiagent environments we introduce two macro-level operators (megamutations) to allow for greater integration between agents; the forming of hereditary endo-symbiosis and the horizontal transfer genes between such symbionts. Our results indicate that hereditary endo-symbiosis will form between agents evolving from within a window of the chaotic region of their attribute space and that gene transfer will occur from within a larger overlapping window. These operators are used within a generic rule-based framework, a simplified version of Pittsburgh-style Classifier Systems, which we alter to allow for direct systemic communication to evolve between the thus represented agents. We find that uninterpreted communication protocols will emerge between such agents using our framework. This work therefore contributes to the implementation of the Genetic Algorithm within complex systems.

[119]
Larry Bull. On ZCS in Multi-agent Environments. In A. E. Eiben, T. Baeck, M. Schoenauer, and H.-P. Schwefel, editors, Proceedings Parallel Problem Solving From Nature (PPSN-V), volume 1498 of Lecture Notes in Computer Science, pages 471-480. Springer-Verlag, 1998.
This paper examines the performance of the ZCS Michigan-style classifier system in multi-agent environments. Using an abstract multi-agent model the effects of varying aspects of the performance, reinforcement and discovery components are examined. It is shown that small modifications to the basic ZCS architecture can improve its performance in environments with significant inter-agent dependence. Further, it is suggested that classifier systems have characteristics which make them more suitable to such non-stationary problem domains in comparison to other forms of reinforcement learning. Results from initial use of ZCS as an adaptive economic trading agent within an artificial double-auction market are then presented, with the findings from the abstract model shown to improve the efficiency of the traders and hence the overall market.

[120]
Larry Bull. On Evolving Social Systems. Computational and Mathematical Organization Theory, 5(3):281-298, 1999.

[121]
Larry Bull. On using ZCS in a Simulated Continuous Double-Auction Market. In Banzhaf et al. [32], pages 83-90.
This paper presents results from on-going investigations into the performance of the Michigan-style classifier system in a complex multi-agent environment. Using a simplified model of a continuous double-auction market place the use of ZCS as an adaptive economic trading agent is examined. It is shown that a number of small changes to the basic system greatly improves its performance, resulting in improvements in the overall efficiency of the market. It is also shown that the role of the rule-discovery component of the classifier system is particularly critical in such a closely-coupled multi-agent environment.

[122]
Larry Bull. Simple markov models of the genetic algorithm in classifier systems: Accuracy-based fitness. In Proceedings of the International Workshop on Learning Classifier Systems (IWLCS-2000), in the Joint Workshops of SAB 2000 and PPSN 2000 [487]. Extended abstract.
Paper is an extended abstract

[123]
Larry Bull. Simple markov models of the genetic algorithm in classifier systems: Multi-step tasks. In Proceedings of the International Workshop on Learning Classifier Systems (IWLCS-2000), in the Joint Workshops of SAB 2000 and PPSN 2000 [487]. Extended abstract.
Paper is an extended abstract

[124]
Larry Bull. Lookahead And Latent Learning In ZCS. In W. B. Langdon, E. Cantú-Paz, K. Mathias, R. Roy, D. Davis, R. Poli, K. Balakrishnan, V. Honavar, G. Rudolph, J. Wegener, L. Bull, M. A. Potter, A. C. Schultz, J. F. Miller, E. Burke, and N. Jonoska, editors, GECCO 2002: Proceedings of the Genetic and Evolutionary Computation Conference, pages 897-904, New York, 9-13 July 2002. Morgan Kaufmann Publishers.

[125]
Larry Bull. On accuracy-based fitness. Journal of Soft Computing, 6(3-4):154-161, 2002.

[126]
Larry Bull, editor. Applications of Learning Classifier Systems. Springer, 2004.

[127]
Larry Bull. A simple payoff-based learning classifier system. In Xin Yao et al., editor, Parallel Problem Solving from Nature - PPSN VIII, pages 1032-1041. Springer Verlag, 2004.

[128]
Larry Bull. Two Simple Learning Classifier Systems. In Larry Bull and Tim Kovacs, editors, Foundations of Learning Classifier Systems, number 183 in Studies in Fuzziness and Soft Computing, pages 63-90. Springer-Verlag, 2005.

[129]
Larry Bull. On lookahead and latent learning in simple lcs. In Jaume Bacardit, Ester Bernadó-Mansilla, Martin Butz, Tim Kovacs, Xavier Llorà, and Keiki Takadama, editors, Learning Classifier Systems. 10th and 11th International Workshops (2006-2007), volume 4998/2008 of Lecture Notes in Computer Science, pages 154-168. Springer, 2008.
Learning Classifier Systems use evolutionary algorithms to facilitate rule- discovery, where rule fitness is traditionally payoff based and assigned under a sharing scheme. Most current research has shifted to the use of an accuracy-based scheme where fitness is based on a rule’s ability to predict the expected payoff from its use. Learning Classifier Systems that build anticipations of the expected states following their actions are also a focus of current research. This paper presents a simple but effective learning classifier system of this last type, using payoff-based fitness, with the aim of enabling the exploration of their basic principles, i.e., in isolation from the many other mechanisms they usually contain. The system is described and modelled, before being implemented. Comparisons to an equivalent accuracy-based system show similar performance. The use of self-adaptive mutation in such systems in general is then considered.

[130]
L. Bull. On dynamical genetic programming: Simple boolean networks in learning classifier systems. To appear in International Journal of Parallel, Emergent and Distributed Systems, 2009.

[131]
Martin V. Butz and David E. Goldberg. Bounding the population size in XCS to ensure reproductive opportunities. In E. Cantú-Paz, J. A. Foster, K. Deb, D. Davis, R. Roy, U.-M. O'Reilly, H.-G. Beyer, R. Standish, G. Kendall, S. Wilson, M. Harman, J. Wegener, D. Dasgupta, M. A. Potter, A. C. Schultz, K. Dowsland, N. Jonoska, and J. Miller, editors, Genetic and Evolutionary Computation -- GECCO-2003, volume 2724 of LNCS, pages 1844-1856. Springer-Verlag, 2003.

[132]
Martin V. Butz and Martin Pelikan. Analyzing the evolutionary pressures in xcs. In Lee Spector, Erik D. Goodman, Annie Wu, W.B. Langdon, Hans-Michael Voigt, Mitsuo Gen, Sandip Sen, Marco Dorigo, Shahram Pezeshk, Max H. Garzon, and Edmund Burke, editors, Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2001), pages 935-942, San Francisco, California, USA, 7-11 July 2001. Morgan Kaufmann.

[133]
M.V. Butz and M. Pelikan. Studying XCS/BOA learning in boolean functions: structure encoding and random boolean functions. In M. Cattolico et al., editor, Genetic and evolutionary computation conference, GECCO 2006, pages 1449-1456. ACM, 2006.

[134]
Martin Butz and Wolfgang Stolzmann. Action-Planning in Anticipatory Classifier Systems. In Wu [923], pages 242-249.
Learning consists in the acquisition of knowledge. In Reinforcement Learning this is knowledge about how to reach a maximum of environmental reward. We are interested in the acquisition of knowledge that consists in having expectations of behavioral consequences. Behavioral consequences depend on the current situation, so it is necessary to learn in which situation S which behavior/reaction R leads to which behavioral consequences C. In other words, SRC units are learned. It was the psychologist Edward Tolman (1932) who firstly stated that animals learn SRC units. Seward (1949) proved that rats are able to learn in the absence of reward and confirmed Tolman's assumption. Learning in the absence of reinforcement is called `latent learning' and cannot be explained by usual reinforcement learning techniques. In the field of Learning Classifier Systems (LCS) latent learning is realized in Riolo's CFSC2 (Riolo, 1991) and Stolzmann's ACS (Stolzmann, 1997, 1998). Both authors prove the performance of their learning algorithms with a simulation of Seward's experiment. This experiment consists in a learning phase without any reward followed by a test phase where the rats have to use the knowledge they acquired during the learning phase to do action-planning. Action-planning and latent learning occur at different times. This paper focuses on the integration of action-planning and latent learning in ACS. Using an example about learning of the hand-eye co-ordination of a robot arm in conjunction with a camera it will be shown, that a combination of action-planning and latent learning in ACS induces a substantial reduction of the number of trials which are required to learn a complete model of a prototypically environment.

[135]
Martin V. Butz and Stewart W. Wilson. An Algorithmic Description of XCS. Technical Report 2000017, Illinois Genetic Algorithms Laboratory, 2000.
A concise description of the XCS classifier system's parameters, structures, and algorithms is presented as an aid to research. The algorithms are written in modularly structured pseudo code with accompanying explanations.

[136]
Martin V. Butz and Stewart W. Wilson. An Algorithmic Description of XCS. In Lanzi et al. [545], pages 253-272.

[137]
Martin V. Butz and Stewart W. Wilson. An algorithmic description of xcs. Journal of Soft Computing, 6(3-4):144-153, 2002.

[138]
Martin Butz, David E. Goldberg, and Wolfgang Stolzmann. New challenges for an ACS: Hard problems and possible solutions. Technical Report 99019, University of Illinois at Urbana-Champaign, Urbana, IL, October 1999.
An Anticipatory Classifier System (ACS) is a learning mechanism based on learning classifier systems and the cognitive model of ``Anticipatory Behavioral Control''. By comparing perceived consequences with its own expectations (anticipations), an ACS is able to learn in multi-step environments. To date, the ACS has proven its abilities in various problems of that kind. It is able to learn latently (i.e. to learn without getting any reward) and it is able to distinguish between non-Markov states. Additionally, an ACS is capable of incrementally building a cognitive map that can be used to do action-planning. Although the ACS has proven to scale up in suitable environments, it depends on certain environmental properties. It believes itself to be the only agent that can change the perceptions received from an environment. Any environmental change is considered and believed to be caused by the executed actions. The ACS learns from the changes by using fixed mechanisms. This paper reveals the properties of an environment that the current ACS assumes to be given. By investigating the problems of the current ACS when violating these properties we believe that this investigation will immediately serve for a better understanding of the ACS and lead to many ideas to improve the current ACS. We will propose some ideas and discuss the important ones in more detail.

[139]
Martin Butz, David E. Goldberg, and Wolfgang Stolzmann. The anticipatory classifier system and genetic generalization. Technical Report 2000032, Illinois Genetic Algorithms Laboratory, 2000.
The anticipatory classifier system (ACS) combines the learning classifier system framework with the learning theory of anticipatory behavioral control. The result is an evolutionary system that builds an environmental model and further applies reinforcement learning techniques to form an optimal behavioral policy in the model. After providing some background as well as outlining the objectives of the system, we explain in detail all involved current processes. Furthermore, we analyze the deficiency of over-specialization in the anticipatory learning process (ALP), the main learning mechanism in the ACS. Consequently, we introduce a genetic algorithm (GA) to the ACS that is meant for generalization of over-specialized classifiers. We show that it is possible to form a symbiosis between a directed specialization and a genetic generalization mechanism achieving a learning mechanism that evolves a complete, accurate, and compact description of a perceived environment. Results in three different environmental settings confirm the usefulness of the genetic algorithm in the ACS. Finally, we discuss future research directions with the ACS and anticipatory systems in general.

[140]
Martin V. Butz, David E. Goldberg, and Wolfgang Stolzmann. Introducing a Genetic Generalization Pressure to the Anticipatory Classifier System -- Part 1: Theoretical Approach. In Whitely et al. [878], pages 34-41. Also Technical Report 2000005 of the Illinois Genetic Algorithms Laboratory.
The Anticipatory Classifier System (ACS) is a learning classifier system that is based on the cognitive mechanism of anticipatory behavioral control. Besides the common reward learning, the ACS is able to learn latently (i.e. to learn in an environment without getting any reward) which is not possible with reinforcement learning techniques. Furthermore, it forms a complete internal representation of the environment and thus it is able to use cognitive processes such as reasoning and planning. Latest research observed that the ACS is not generating accurate, maximally general rules reliably (i.e. rules which are accurate and also as general as possible), but it is sometimes generating over-specialized rules. This paper shows how a genetic algorithm can be used to overcome this present pressure of over-specification in the ACS mechanism with a genetic generalization pressure. The ACS works then as a hybrid which learns latently, forms a cognitive map, and evolves accurate, maximally general rules.

[141]
Martin V. Butz, David E. Goldberg, and Wolfgang Stolzmann. Introducing a Genetic Generalization Pressure to the Anticipatory Classifier System -- Part 2: Performance Analysis. In Whitely et al. [878], pages 42-49. Also Technical Report 2000006 of the Illinois Genetic Algorithms Laboratory.
The Anticipatory Classifier System (ACS) is able to form a complete internal representation of an environment. Unlike most other classifier system and reinforcement learning approaches, it is able to learn latently (i.e. to learn in an environment without getting any reward) and to form an internal model of the perceived environment. After the observation that the model is not necessarily maximally general a genetic generalization pressure was introduced to the ACS. This paper focuses on the different mechanisms in the anticipatory learning process, which resembles the specification pressure, and in the genetic algorithm, which realizes the genetic generalization pressure. The capability of generating maximally general rules and evolving a completely converged population is investigated in detail. Furthermore, the paper approaches a first comparison with the XCS classifier system in different mazes and the multiplexer problem.

[142]
Martin V. Butz, David E. Goldberg, and Wolfgang Stolzmann. Investigating Generalization in the Anticipatory Classifier System. In Proceedings of Parallel Problem Solving from Nature (PPSN VI), 2000. Also technical report 2000014 of the Illinois Genetic Algorithms Laboratory.
Recently, a genetic algorithm (GA) was introduced to the Anticipatory Classifier System (ACS) which surmounted the occasional problem of over-specification of rules. This paper investigates the resulting generalization capabilities further by monitoring in detail the performance of the ACS in the highly challenging multiplexer task. Moreover, by comparing the ACS to the XCS in this task it is shown that the ACS generates accurate, maximally general rules and its population converges to those rules. Besides the observed ability of latent learning and the formation of an internal environmental representation, this ability of generalization adds a new advantage to the ACS in comparison with similar approaches.

[143]
Martin V. Butz, David E. Goldberg, and Wolfgang Stolzmann. Probability-enhanced predictions in the anticipatory classifier system. In Proceedings of the International Workshop on Learning Classifier Systems (IWLCS-2000), in the Joint Workshops of SAB 2000 and PPSN 2000 [487]. Extended abstract.
Paper is an extended abstract

[144]
Martin V. Butz, Tim Kovacs, Pier Luca Lanzi, and Stewart W. Wilson. How XCS Evolves Accurate Classifiers. In Lee Spector, Erik D. Goodman, Annie Wu, W. B. Langdon, Hans-Michael Voigt, Mitsuo Gen, Sandip Sen, Marco Dorigo, Shahram Pezeshk, Max H Garzon, and Edmund Burke, editors, GECCO-2001: Proceedings of the Genetic and Evolutionary Computation Conference, pages 927-934. Morgan Kaufmann, 2001.

[145]
Martin V. Butz, Kumara Sastry, and David E. Goldberg. Tournament selection: Stable fitness pressure in XCS. In E. Cantú-Paz, J. A. Foster, K. Deb, D. Davis, R. Roy, U.-M. O'Reilly, H.-G. Beyer, R. Standish, G. Kendall, S. Wilson, M. Harman, J. Wegener, D. Dasgupta, M. A. Potter, A. C. Schultz, K. Dowsland, N. Jonoska, and J. Miller, editors, Genetic and Evolutionary Computation -- GECCO-2003, volume 2724 of LNCS, pages 1857-1869. Springer-Verlag, 2003.

[146]
Martin Butz, Tim Kovacs, Pier Luca Lanzi, and Stewart W. Wilson. Toward a theory of generalization and learning in xcs. IEEE Transactions on Evolutionary Computation, 8(1):8-46, 2004.
Takes initial steps toward a theory of generalization and learning in the learning classifier system XCS. We start from Wilson's generalization hypothesis, which states that XCS has an intrinsic tendency to evolve accurate, maximally general classifiers. We analyze the different evolutionary pressures in XCS and derive a simple equation that supports the hypothesis theoretically. The equation is tested with a number of experiments that confirm the model of generalization pressure that we provide. Then, we focus on the conditions, termed "challenges," that must be satisfied for the existence of effective fitness or accuracy pressure in XCS. We derive two equations that suggest how to set the population size and the covering probability so as to ensure the development of fitness pressure. We argue that when the challenges are met, XCS is able to evolve problem solutions reliably. When the challenges are not met, a problem may provide intrinsic fitness guidance or the reward may be biased in such a way that the problem will still be solved. The equations and the influence of intrinsic fitness guidance and biased reward are tested on large Boolean multiplexer problems. The paper is a contribution to understanding how XCS functions and lays the foundation for research on XCS's learning complexity.

[147]
M.V. Butz, D.E. Goldberg, and P.L. Lanzi. Bounding learning time in XCS. In Genetic and evolutionary computation (GECCO 2004), volume 3103/2004 of LNCS, pages 739-750. Springer, 2004.
It has been shown empirically that the XCS classifier system solves typical classification problems in a machine learning competitive way. However, until now, no learning time estimate has been derived analytically for the system. This paper introduces a time estimate that bounds the learning time of XCS until maximally accurate classifiers are found. We assume a domino convergence model in which each attribute is successively specialized to the correct value. It is shown that learning time in XCS scales polynomially in problem length and problem complexity and thus in a machine learning competitive way.

[148]
M.V. Butz, D.E. Goldberg, and P.L. Lanzi. Computational complexity of the XCS classifier system. In Larry Bull and Tim Kovacs, editors, Foundations of Learning Classifier Systems, number 183 in Studies in Fuzziness and Soft Computing, pages 91-126. Springer-Verlag, 2005.

[149]
M.V. Butz, D.E. Goldberg, and P.L. Lanzi. Gradient descent methods in learning classifier systems: improving XCS performance in multistep problems. IEEE Trans. Evolutionary Computation, 9(5):452-473, 2005.

[150]
M.V. Butz, M. Pelikan, X. Llorà, and D.E. Goldberg. Extracted global structure makes local building block processing effective in XCS. In H.G. Beyer and U.M. O'Reilly, editors, Genetic and evolutionary computation conference, GECCO 2005, pages 655-662. ACM, 2005.

[151]
M.V. Butz, P.L. Lanzi, and S.W. Wilson. Hyper-ellipsoidal conditions in xcs: rotation, linear approximation, and solution structure. In M. Cattolico, editor, Proc. genetic and evolutionary computation conference (GECCO 2006), pages 1457-1464. ACM, 2006.

[152]
M.V. Butz, M. Pelikan, X. Llorà, and D.E. Goldberg. Automated global structure extraction for effective local building block processing in XCS. Evolutionary Computation, 14(3):345-380, 2006.

[153]
Martin V. Butz, David E. Goldberg, and Pier Luca Lanzi. Effect of pure error-based fitness in xcs. In Tim Kovacs, Xavier LLòra, Keiki Takadama, Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson, editors, Learning Classifier Systems. International Workshops, IWLCS 2003-2005, Revised Selected Papers, volume 4399 of LNCS, pages 104-114. Springer, 2007.

[154]
M.V. Butz, D.E. Goldberg, P.L. Lanzi, and K. Sastry. Problem solution sustenance in XCS: Markov chain analysis of niche support distributions and the impact on computational complexity. Genetic Programming and Evolvable Machines, 8(1):5-37, 2007.

[155]
Martin V. Butz, Pier Luca Lanzi, Xavier Llorá, and Daniele Loiacono. An Analysis of Matching in Learning Classifier Systems. In Proceedings of the 10th annual conference on Genetic and evolutionary computation, pages 1349-1356. ACM, 2008.
We investigate rule matching in learning classifier systems for problems involving binary and real inputs. We consider three rule encodings: the widely used character-based encoding, a specificity-based encoding, and a binary encoding used in Alecsys. We compare the performance of the three algorithms both on matching alone and on typical test problems. The results on matching alone show that the population generality influences the performance of the matching algorithms based on string representations in different ways. Character-based encoding becomes slower and slower as generality increases, specificity-based encoding becomes faster and faster as generality increases. The results on typical test problems show that the specificity-based representation can halve the time required for matching but also that binary encoding is about ten times faster on the most difficult problems. Moreover, we extend specificity-based encoding to real-inputs and propose an algorithm that can halve the time require for matching real inputs using an interval-based representation.

[156]
M.V. Butz, P.L. Lanzi, and S.W. Wilson. Function approximation with xcs: Hyperellipsoidal conditions, recursive least squares, and compaction. IEEE Trans. Evolutionary Computation, 12(3):355-376, 2008.

[157]
M.V. Butz, P. Stalph, and P.L. Lanzi. Self-adaptive mutation in xcsf. In GECCO '08: Proceedings of the 10th annual conference on Genetic and evolutionary computation, pages 1365-1372. ACM, 2008.
Recent advances in XCS technology have shown that self-adaptive mutation can be highly useful to speed-up the evolutionary progress in XCS. Moreover, recent publications have shown that XCS can also be successfully applied to challenging real-valued domains including datamining, function approximation, and clustering. In this paper, we combine these two advances and investigate self-adaptive mutation in the XCS system for function approximation with hyperellipsoidal condition structures, referred to as XCSF in this paper. It has been shown that XCSF solves function approximation problems with an accuracy, noise robustness, and generalization capability comparable to other statistical machine learning techniques and that XCSF outperforms simple clustering techniques to which linear approximations are added. This paper shows that the right type of self-adaptive mutation can further improve XCSF's performance solving problems more parameter independent and more reliably. We analyze various types of self-adaptive mutation and show that XCSF with self-adaptive mutation ranges,differentiated for the separate classifier condition values, yields most robust performance results. Future work may further investigate the properties of the self-adaptive values and may integrate advanced self-adaptation techniques.

[158]
Martin V. Butz. An Implementation of the XCS classifier system in C. Technical Report 99021, The Illinois Genetic Algorithms Laboratory, 1999.
The XCS classifier system was developed by Wilson (1995). The learning mechanism is based on the accuracy of its reward prediction. This method leads to the formation of accurate most general classifiers. This paper explains how to download, compile and use the XCS code version 1.0 written in ANSI C. It discusses how to select various parameter settings, how to add and remove certain procedures in the XCS, how to apply the XCS in the multiplexer environment and diverse woods environments, and how to add code to apply the XCS in other environments. The code provides the mechanisms introduced by Wilson (1995) and the enhancements published by Wilson (1998).

[159]
Martin V. Butz. XCSJava 1.0: An Implementation of the XCS classifier system in Java . Technical Report 2000027, Illinois Genetic Algorithms Laboratory, 2000.
The XCSJava 1.0 implementation of the XCS classifier system in Java is freely available from the IlliGAL anonymous ftp-site. The implementation covers the basic features of the XCS classifier system and provides a multiplexer and maze environment for testing purposes. This paper explains how to download, compile, and run the code. Moreover, it explains the object oriented approach in the implementation and the possible parameter manipulation as well as the environmental interface to hook in other test environments. Additionally to the source code, an executable package of the version as well as an XCSJava 1.0 API documentation is provided.

[160]
Martin V. Butz. An Algorithmic Description of ACS2. In Lanzi et al. [546], pages 211-229.

[161]
Martin V. Butz. Anticipatory learning classifier systems. Kluwer Academic Publishers, 2002.

[162]
Martin V. Butz. Biasing Exploration in an Anticipatory Learning Classifier System. In Lanzi et al. [546], pages 3-22.

[163]
M.V. Butz. Kernel-based, ellipsoidal conditions in the real-valued xcs classifier system. In H.G. Beyer et al., editor, Proc. genetic and evolutionary computation conference (GECCO 2005), pages 1835-1842. ACM, 2005.

[164]
M.V. Butz. Rule-Based Evolutionary Online Learning Systems: A Principled Approach to LCS Analysis and Design. Studies in Fuzziness and Soft Computing. Springer-Verlag, 2006.
This book offers a comprehensive introduction to learning classifier systems (LCS) – or more generally, rule-based evolutionary online learning systems. LCSs learn interactively – much like a neural network – but with an increased adaptivity and flexibility. This book provides the necessary background knowledge on problem types, genetic algorithms, and reinforcement learning as well as a principled, modular analysis approach to understand, analyze, and design LCSs. The analysis is exemplarily carried through on the XCS classifier system – the currently most prominent system in LCS research. Several enhancements are introduced to XCS and evaluated. An application suite is provided including classification, reinforcement learning and data-mining problems. Reconsidering John Holland’s original vision, the book finally discusses the current potentials of LCSs for successful applications in cognitive science and related areas.

[165]
Alessio Camilli and Roberto Di Meglio. Sistemi a classificatori su architetture a parallelismo massiccio. Technical report, Univ. Delgi Studi di Pisa, 1989.

[166]
Alessio Camilli, Roberto Di Meglio, F. Baiardi, M. Vanneschi, D. Montanari, and R. Serra. Classifier System Parallelization on MIMD Architectures. Technical Report 3/17, CNR, 1990.

[167]
Alessio Camilli. Classifier systems in massively parallel architectures. Master's thesis, University of Pisa, 1990. (In Italian).

[168]
Y. J. Cao, N. Ireson, Larry Bull, and R. Miles. Design of a Traffic Junction Controller using a Classifier System and Fuzzy Logic. In Bernd Reusch, editor, Proceedings of the Sixth International Conference on Computational Intelligence, Theory and Applications (6th Fuzzy Days), volume 1625 of LNCS, pages 342-353. Springer-Verlag, 1999.

[169]
Y. J. Cao, N. Ireson, L. Bull, and R. Miles. Distributed Learning Control of Traffic Signals. In Proceedings of the EvoNet Workshops - EvoSCONDI 2000, pages 117-126. Springer, 2000.

[170]
Y. J. Cao, N. Ireson, L. Bull, and R. Miles. An evolutionary intelligent agents approach to traffic signal control. International Journal of Knowledge-based Intelligent Engineering Systems, 5(4):279-289, 2001.

[171]
A. Carbonaro, G. Casadei, and A. Palareti. Genetic Algorithms and Classifier Systems in Simulating a Cooperative Behavior. In Albrecht et al. [7], pages 479-483.
Genetic Algorithms and Classifier Systems are often used in biologic-like and evolutionary behaviors' simulations. The basic example is Wood7 Wilson's world. In this environment it is interesting to study some problems: Can evolve the cooperative behaviors of organisms present in the world? How and when do the behaviors evolve? Some preliminary results show the conditions under that cooperative behavior rules are developing rapidly. Particularly we have pointed out the likely of following observations: a. The cooperative behavior develops more easily if the initial population starts from the same point. b. It exists some thresholds under that the cooperative behavior can't evolve; these thresholds depend to the population size.

[172]
Brian Carse and Terence C. Fogarty. A delayed-action classifier system for learning in temporal environments. In Proceedings of the 1st IEEE Conference on Evolutionary Computation, volume 2, pages 670-673, 1994.

[173]
Brian Carse and Terence C. Fogarty. A Fuzzy Classifier System Using the Pittsburgh Approach. In Davidor and Schwefel [228], pages 260-269.
This paper describes a fuzzy classifier system using the Pittsburgh model. In this model genetic operations and fitness assignment apply to complete rule-sets, rather than to individual rules, thus overcoming the problem of conflicting individual and collective interests of classifiers. The fuzzy classifier system presented here dynamically adjusts both membership functions and fuzzy relations. A modified crossover operator for particular use in Pittsburgh-syle fuzzy classifier systems, with variable length rule-sets, is introduced and evaluated. Experimental results of the new system, which appear encouraging, are presented and discussed.

[174]
Brian Carse and A. G. Pipe. A Framework For Evolving Fuzzy Classifier Systems Using Genetic Programming. In 14th Int. Conf. of Florida Artificial Intelligence Research Society FLAIRS-2001, pages 465-469. AAAI Press, 2001.

[175]
Brian Carse and A. G. Pipe. X-FCS: a fuzzy classifier systems using accuracy based fitness -- first results. In Procs. Int. Conf. In Fuzzy Logic and Technology, EUSFLAT, pages 195-198, 2001.

[176]
Brian Carse, Terence C. Fogarty, and A. Munro. Distributed Adaptive Routing Control in Communications Networks using a Temporal Fuzzy Classifier System. In Proceedings of the Fifth IEEE Conference on Fuzzy Systems, pages 2203-2207. IEEE, 1996.

[177]
Brian Carse, Terence C. Fogarty, and A. Munro. Evolutionary Learning of Controllers using Temporal Fuzzy Classifier Systems. In I. C. Parmee, editor, Proceedings of the Second Conference on Adaptive Computing in Engineering Design and Control, pages 174-180, 1996.

[178]
Brian Carse, Terence C. Fogarty, and A. Munro. Evolving fuzzy rule based controllers using genetic algorithms. International Journal for Fuzzy Sets and Systems, 80:273-293, 1996.

[179]
Brian Carse, Terence C. Fogarty, and A. Munro. The Temporal Fuzzy Classifier System and its Application to Distributed Control in a Homogeneous Multi-Agent ecology. In Goodman et al. [379], pages 76-86.
A fuzzy classifier system is described which explicitly represents time in the classifier syntax by augmenting individual classifiers with temporal tags. This feature allows the learning algorithm - in this case the genetic algorithm - to explore and exploit temporal features of the environment in which the classifier system might be expected to operate. The proposed temporal fuzzy classifier system is applied to a multi-agent distributed control task - adaptive distributed rooting in packet-switched communications networks.

[180]
Brian Carse, Terence C. Fogarty, and Alistair Munro. Evolving Temporal Fuzzy Rule-Bases for Distributed Routing Control in Telecommunication Networks. In Herrera and Verdegay [423], pages 467-488.
Distributed routing control in telecommunication networks is a challenging problem. A networked assembly of geographically dispersed routing controllers are required to route traffic across the network in such a way so as to avoid congestion. Measured state information for each controller is delayed and necessarily available only on occasion. Interactions between routing controllers are highly non-linear and instability is a serious problem. A hybrid technique for distributing routing is proposed based on a synthesis of shortest-path routing, machine learning and fuzzy control. An architecture is described for a novel temporal fuzzy classifier system which forms the basis for each routing controller. Experimental results are presented which compare the new technique with two extant routing methods -- non-adaptive shortest-hop routing and adaptive shortest-path routing.

[181]
Brian Carse, Terence C. Fogarty, and Alistair Munro. Artificial evolution of fuzzy rule bases which represent time: A temporal fuzzy classifier system. International Journal of Intelligent Systems, 13(issue 10-11):905-927, 1998.

[182]
Brian Carse. Learning Anticipatory Behaviour Using a Delayed Action Classifier System. In Fogarty [324], pages 210-223.
To manifest anticipatory behaviour that goes beyond simple stimulus-response, classifier systems must evolve internal reasoning processes based on couplings via internal messages. A major challenge that has been encountered in engendering internal reasoning processes in classifier systems has been the discovery and maintenance of long classifier chains. This paper proposes a modified version of the traditional classifier system, called the delayed action classifier system (DACS), devised specifically for learning of anticipatory or predictive behaviour. DACS operates by delaying the action (i.e. posting of messages) of appropriately tagged, matched classifiers by a number of execution cycles which is encoded on the classifier. Since classifier delays are encoded on the classifier genome, a GA is able to explore simultaneously the spaces of actions and delays. Results of experiments comparing DACS to a traditional classifier system in terms of the dynamics of classifier reinforcement and system performance using the bucket brigade are presented and examined. Experiments comparing DACS with a traditional classifier system, which appear encouraging, for a simple prediction problem are described and considered. Areas for further work using the delayed-action classifier notion are suggested and briefly discussed.

[183]
G. Casadei, A. Palareti, and G. Proli. Classifier System in Traffic Management. In Albrecht et al. [7], pages 620-627.
The systems of controlling and improving traffic movement have been studied for several years now. The usefulness of these systems is that they can modify and change the lights signals of traffic lights. It is not enough to intervene when the situation has reached a critical point such as a traffic jam. The system has to work out how the traffic will flow. The ideal solution would be a system that works out and foresees the situation on the roads based on a model of motorists' behaviour. This research shows how to best utilise the classifier systems so that it would be possible to create a model that is similar to that of the real world.

[184]
J. Casillas, B. Carse, and L. Bull. Fuzzy-XCS: a michigan genetic fuzzy system. IEEE Trans. Fuzzy Systems, 15:536-550, 2007.

[185]
Proceedings of the 2000 Congress on Evolutionary Computation (CEC00). IEEE Press, 2000.

[186]
Proceedings of the 2001 Congress on Evolutionary Computation (CEC01). IEEE Press, 2001.

[187]
Keith W. Chalk and George D. Smith. The Co-evolution of Classifier Systems in a Competitive Environment. Poster presented at AISB94. Authors were from the University of East Anglia, U.K.

[188]
Keith Chalk and George D. Smith. Multi-Agent Classifier Systems and the Iterated Prisoner's Dilemma. In Smith et al. [770], pages 615-618.
This paper describes experiments using multiple classifier system (CS) agents to play the iterated prisoner's dilemma (IPD) under various conditions. Our main interest is in how, and under what circumstances, cooperation is most likely to emerge through competition between these agents. Experiments are conducted with agents playing fixed strategies and other agents individually and in tournaments, with differing CS parameters. Performance improves when reward is stored and averaged over longer periods, and when a genetic algorithm (GA) is used more frequently. Increasing the memory of the system improves performance to a point, but long memories proved difficult to reinforce fully and performed less well.

[189]
Sin Man Cheang, Kin Hong Lee, and Kwong Sak Leung. Data classification using genetic parallel programming. In E. Cantú-Paz, J. A. Foster, K. Deb, D. Davis, R. Roy, U.-M. O'Reilly, H.-G. Beyer, R. Standish, G. Kendall, S. Wilson, M. Harman, J. Wegener, D. Dasgupta, M. A. Potter, A. C. Schultz, K. Dowsland, N. Jonoska, and J. Miller, editors, Genetic and Evolutionary Computation -- GECCO-2003, volume 2724 of LNCS, pages 1918-1919. Springer-Verlag, 2003.
A novel Linear Genetic Programming (LGP) paradigm called Genetic Parallel Programming (GPP) has been proposed to evolve parallel programs based on a Multi-ALU Processor. It is found that GPP can evolve parallel programs for Data Classification problems. In this paper, five binary-class UCI Machine Learning Repository databases are used to test the effectiveness of the proposed GPP-classifier. The main advantages of employing GPP for data classification are: 1) speeding up evolutionary process by parallel hardware fitness evaluation; and 2) discovering parallel algorithms automatically. Experimental results show that the GPP-classifier evolves simple classification programs with good generalization performance. The accuracies of these evolved classifiers are comparable to other existing classification algorithms.

[190]
A. P. Chen and Y. H. Chang. Using extended classifier system to forecast S&P futures based on contrary sentiment indicators. Evolutionary Computation, 3:2084-2090, 2005.

[191]
Hung-Ming Chen and Shinn-Ying Ho. Designing an optimal evolutionary fuzzy decision tree for data mining. In Lee Spector, Erik D. Goodman, Annie Wu, W.B. Langdon, Hans-Michael Voigt, Mitsuo Gen, Sandip Sen, Marco Dorigo, Shahram Pezeshk, Max H. Garzon, and Edmund Burke, editors, Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2001), pages 943-950, San Francisco, California, USA, 7-11 July 2001. Morgan Kaufmann.

[192]
C.C. Chen and C.C. Wong. Self-generating rule-mapping fuzzy controller design using a genetic algorithm. IEE Proc. Control Theory Appl., 149(2):143-148, 2002.

[193]
An Pin Chen, Yi-Chang Chen, and Wen-Chuan Tseng. Applying Extending Classifier System to Develop an Option-Operation Suggestion Model of Intraday Trading -- An Example of Taiwan Index Option. In Knowledge-Based Intelligent Information and Engineering Systems, 9th International Conference (KES 2005), volume 3681 of Lecture Notes in Computer Science, pages 27-33. Springer, 2005.

[194]
Tung Wan Cheng, Wen Chih Tsai, and An Pin Chen. Strategy of futures trading mechanism using extended classifier system. In Intelligent Systems, 2004. Proceedings. 2004 2nd International IEEE Conference, volume 2, pages 503-507, 2004.
Nowadays, many artificial intelligent trading models divided the process in three separate subprocesses: trading, validation and application, but these models cannot meet the request of today's trading environment. A new online learning algorithm, extended classifier system (XCS) is used in futures extended classifier trading mechanism (FXCTM) to satisfy traders' requirement. This paper verifies that FXCTM provides a very good forecast ability in future market trading performance. Also, this paper discusses about how the population set of XCS affects the result of the model. Finally, the simulation results show that this model could get an obvious profit from futures market.

[195]
Pawel Cichosz and Jan J. Mulawka. GBQL: A novel genetics-based reinforcement learning architecture. In Proceedings of the Third European Congress on Intelligent Techniques and Soft Computing (EUFIT'95), 1995.
This research attempts to integrate the existing ideas in two fields: reinforcement learning algorithms based on the methods of temporal differences (TD), in particular Q-learning, and genetics-based machine learning, in particular classifier systems (CS). Close relations between the bucket brigade credit assignment algorithm used in classifier systems and TD methods, several widely realized drawbacks of CS, and good theoretical properties of TD, gave the initial motivation for developing a learning architecture that would combine TD-based temporal credit assignment algorithms with genetics-based adaptive knowledge representation. This paper presents a simple instantiation of this idea, called GBQL (Genetics-Based Q-Learning). This learning architecture may be expected to be a promising alternative for stimulus-response classifier systems on one hand, and for the implementations of Q-learning using other knowledge representation methods (e.g., connectionist networks) on the other hand.

[196]
Pawel Cichosz and Jan J. Mulawka. Faster temporal credit assignment in learning classifier systems. In Proceedings of the First Polish Conference on Evolutionary Algorithms (KAE-96), 1996.
Classifier systems are genetics-based learning systems using the paradigm of reinforcement learning. In the most challenging case of delayed reinforcement, it involves a difficult temporal credit assignment problem. Standard classifier systems solve this problem using the bucket brigade algorithm. In this paper we show how to make the temporal credit assignment process faster by augmenting this algorithm by some refinements borrowed from a related field of reinforcement learning algorithms based on the methods of temporal differences (TD). These algorithms usually converge significantly faster if they are used in combination with TD(lambda). As a natural consequence of the easily noticeable similarity between the bucket brigade and TD(0), the BB(lambda) algorithm is derived, using the standard technique of eligibility traces. The TTD(lambda,m) procedure, which eliminates eligibility traces and implements an approximation of TD(lambda) in a computationally efficient way, has also been ported to the context of classifier systems, yielding the TBB(lambda,m) algorithm. The two resulting novel algorithms provide promising and, strangely enough, completely unexplored so far possibilities of making learning classifier systems learn faster under the conditions of reinforcement delay.

[197]
Pawel Cichosz. Reinforcement learning algorithms based on the methods of temporal differences. Master's thesis, Institute of Computer Science, Warsaw University of Technology, 1994.
The reinforcement learning paradigm differs significantly from the traditional supervised learning paradigm. An agent in each particular input situation must generate an action. Then it receives a reinforcement value from the environment, providing a measure of the agent's performance. The task for the agent is to maximize the reinforcement values it receives in long term. Reinforcement learning agents are adaptive, reactive, and self-improving. To formulate a particular task as a reinforcement learning task one just has to design an appropriate reinforcement function, specifying the goal of the task. This makes the paradigm widely applicable, especially in such domains as game playing, automatic control, and robotics. The reinforcement value received by the agent at a particular time step may reflect the positive or negative consequences of actions taken several steps before. In order to deal with such delayed reinforcement one needs some algorithms for temporal credit assignment. This thesis concentrates on a class of algorithms based on Sutton's temporal differences (TD) methods. The AHC and Q-learning algorithms are well known instances of this class. The TTD procedure is proposed for the efficient and general implementation of this class of algorithms, as an alternative to the traditional so called eligibility traces implementation, which is found to suffer from both inefficiency and lack of generality. Important practical issues in using these algorithms are discussed. The problem of learning with multidimensional actions is addressed and a simple way to generalize appropriately TD-based algorithms is presented. It is argued that existing one-dimensional algorithms are hardly applicable to tasks with vector actions and that the proposed extensions, despite their simplicity, constitute a promising approach to this problem, though they require further work. A novel genetics-based reinforcement learning architecture is introduced. It combines Q-learning with genetics-based knowledge representation and rule discovery mechanisms. For the class of learning problems considered in this thesis, it can be a promising alternative to Holland's classifier systems with the bucket brigade temporal credit assignment algorithm. For all described algorithms experimental results are presented, illustrating their performance. Several important open problems in reinforcement learning are identified and directions for future research are outlined.

[198]
Pawel Cichosz. Reinforcement Learning by Truncating Temporal Differences. PhD thesis, Department of Electronics and Information Technology, Warsaw University of Technology, 1997.
The paradigm of reinforcement learning provides an appealing framework for developing intelligent adaptive systems. The learner interacts with a possibly unknown and stochastic environment by observing its states and performing actions. It receives scalar reinforcement, or reward values, which provide a relative measure of the quality of the executed actions. The learner's task is to identify an optimal decision policy, i.e., a state-action mapping that leads to the maximization of the rewards it receives in the long term. Reinforcement values may be sparse and delayed with respect to the actions which contributed to them. A common approach to learning from delayed rewards is to use TD(lambda) methods for predicting future rewards in each state. Q-learning is currently the most popular and best theoretically understood TD-based reinforcement learning algorithm, but a variety of other related algorithms can be used. There have been a few impressive practical applications of reinforcement learning, but the existing algorithms still suffer from important deficiencies. This thesis examines possible ways of overcoming some of them, and thus making it easier to develop successful intelligent systems based on the reinforcement learning paradigm. Probably the most painful problem to be addressed is the relatively slow convergence of reinforcement learning algorithms. Although using TD(lambda>0) is known to usually give a considerable learning speedup, in practice TD(0) is still often used, because positive lambda increases the computational expense enormously, particularly for realistic tasks, with large state spaces. This is because TD(lambda) is implemented using eligibility traces, maintained and updated at each time step for all states. In this thesis the effects of the eligibility traces implementation are analyzed and an alternative implementation is derived, called the TTD procedure, which closely approximates TD(lambda) in a computationally efficient way, so that one can use lambda>0 at essentially the same cost as TD(0). This novel technique is theoretically shown to be approximately equivalent to, and empirically demonstrated to perform at least as well as eligibility traces, while it gives impressive computational savings. This is the major contribution of the dissertation around which the remaining contributions are concentrated. The theoretical analysis of TTD leads to a number of additional interesting results. The proposed technique is shown to be covered by the existing TD(lambda) convergence theory, by proving its error-reduction property. It is extended to variable lambda, to allow one to select lambda values adaptively, which has been suggested by some prior work. Reinforcement learning speedup techniques proposed by other authors, based on experience replay, are shown to be equivalent to special variable lambda forms of TTD. A TTD analog is derived for replacing eligibility traces, recently proposed by other authors and argued to have some important advantages over traditional, accumulating eligibility traces. Finally, a version of TTD is presented for average-reward reinforcement learning, alternative to the standard discounted-reward framework, adopted by this work. To apply reinforcement learning algorithms to tasks with large, especially continuous state spaces, it is usually necessary to combine them with learning function approximators to generalize over the state space. For a particular, but widely used class of function approximators, known as parameter-estimation methods, TTD is rederived in a gradient form and shown to be equivalent to the corresponding version of eligibility traces. Empirical results are presented for the combination of TTD and CMAC, a function approximator particularly well suited to reinforcement learning, which show that it learns successfully and requires much less computation than eligibility traces.

[199]
Dave Cliff and Seth G. Bullock. Adding `Foveal Vision' to Wilson's Animat. Adaptive Behavior, 2(1):47-70, 1993.
Different animals employ different strategies for sampling sensory data. The strategies are often closely constrained by environmental considerations, such as the animal's ecological niche. In animals that can see, differences in sampling strategy manifest themselves as differences in field of view and in spatially variant sampling (so-called ``foveal'' vision). In analysing adaptive behaviour in animals, or attempting to design autonomous robots, mechanisms for exploring variations in sensory sampling strategy will be required. This paper describes our work exploring a minimal system for investigating the effects of variations in patterns of sensory sampling. We have re-implemented Wilson's (1986) animat, and then experimented with altering its sensory sampling pattern (i.e. its sensory field). Empirical results are presented which demonstrate that alterations in the sensory field pattern can have a significant effect on the animat's observable behaviour (and hence also on the internal mechanisms which generate the behaviours). Analysis of our results involves characterising the interaction between the animat's sensory field and the environment within which the animat resides. We found that the animat's observed behaviour can, at least in part, be explained as a result of the animat cautiously moving in a manner which maximises the generation of new information from the environment over time. The paper concludes with a discussion of the generality of the results, and reflections on the prospects for further work.

[200]
Dave Cliff and Susi Ross. Adding Temporary Memory to ZCS. Adaptive Behavior, 3(2):101-150, 1994. Also technical report: ftp://ftp.cogs.susx.ac.uk/pub/reports/csrp/csrp347.ps.Z.
In a recent paper, Wilson (1994b) described a `zeroth-level' classifier system (ZCS). ZCS employs a reinforcement learning technique comparable to Q-Learning (Watkins, 1989). This paper presents results from the first reconstruction of ZCS. Having replicated Wilson's results, we extend ZCS in a manner suggested by Wilson: the original formulation of ZCS has no memory mechanisms, but Wilson (1994b) suggested how internal `temporary memory' registers could be added. We show results from adding one-bit and two-bit memory registers to ZCS. Our results demonstrate that ZCS can efficiently exploit memory facilities in non-Markov environments. We also show that the memoryless ZCS can converge on near-optimal stochastic solutions in non-Markov environments. Following the discussion of adding memory, we present results from trials using ZCS in Markov environments requiring increasingly long chains of actions before reward is received. Our results indicate that inaccurate over-general classifiers can interact with the classifier-generation mechanisms to cause catastrophic breakdowns in overall system performance. Basing classifier fitness on accuracy may alleviate this problem. We conclude that the memory mechanism in its current form is unlikely to scale well for situations requiring large amounts of temporary memory. Nevertheless, the ability to find stochastic solutions when there is insufficient memory might offset this problem to some extent.

[201]
Dave Cliff and Susi Ross. Adding Temporary Memory to ZCS. Technical Report CSRP347, School of Cognitive and Computing Sciences, University of Sussex, 1995. ftp://ftp.cogs.susx.ac.uk/pub/reports/csrp/csrp347.ps.Z.

[202]
Dave Cliff, Philip Husbands, Jean-Arcady Meyer, and Stewart W. Wilson, editors. From Animals to Animats 3. Proceedings of the Third International Conference on Simulation of Adaptive Behavior (SAB94). A Bradford Book. MIT Press, 1994.

[203]
H. G. Cobb and John J. Grefenstette. Learning the persistence of actions in reactive control rules. In Proceedings 8th International Machine Learning Workshop, pages 293-297. Morgan Kaufmann, 1991.
This paper explores the effect of explicitly searching for the persistence of each decision in a time-dependent sequential decision task. In prior studies, Grefenstette, et al, show the effectiveness of SAMUEL, a genetic algorithm-based system, in solving a simulation problem where an agent learn show to evade a predator that is in pursuit. In their work, an agent applies a control action at each time step. This paper examines a reformulation of the problem: the agent learns not only the level of response of a control action, but also how long to apply that control action. By examining this problem, the work shows that it is appropriate to choose a representation of the state space that compresses time information when solving a time-dependent sequential decision problem. By compressing time information, critical events in the decision sequence become apparent.

[204]
Philippe Collard and Cathy Escazut. Relational Schemata: A Way to Improve the Expressiveness of Classifiers. In Eshelman [298], pages 397-404.

[205]
Marco Colombetti and Marco Dorigo. Learning to Control an Autonomous Robot by Distributed Genetic Algorithms. In Roitblat and Wilson [700], pages 305-312.

[206]
Marco Colombetti and Marco Dorigo. Robot Shaping: Developing Situated Agents through Learning. Technical Report TR-92-040, International Computer Science Institute, Berkeley, CA, 1993.
Learning plays a vital role in the development of situated agents. In this paper, we explore the use of reinforcement learning to "shape" a robot to perform a predefined target behavior. We connect both simulated and real robots to Alecsys, a parallel implementation of a learning classifier system with an extended genetic algorithm. After classifying different kinds of Animat-like behaviors, we explore the effects on learning of different types of agent's architecture (monolithic, flat and hierarchical) and of training strategies. In particular, hierarchical architecture requires the agent to learn how to coordinate basic learned responses. We show that the best results are achieved when both the agent's architecture and the training strategy match the structure of the behavior pattern to be learned. We report the results of a number of experiments carried out both in simulated and in real environments, and show that the results of simulations carry smoothly to real robots. While most of our experiments deal with simple reactive behavior, in one of them we demonstrate the use of a simple and general memory mechanism. As a whole, our experimental activity demonstrates that classifier systems with genetic algorithms can be practically employed to develop autonomous agents.

[207]
Marco Colombetti and Marco Dorigo. Training Agents to Perform Sequential Behavior. Technical Report TR-93-023, International Computer Science Institute, Berkeley, CA, September 1993.
This paper is concerned with training an agent to perform sequential behavior. In previous work we have been applying reinforcement learning techniques to control a reactive robot. Obviously, a pure reactive system is limited in the kind of interactions it can learn. In particular, it can only learn what we call pseudo-sequences, that is sequences of actions in which the transition signal is generated by the appearance of a sensorial stimulus. We discuss the difference between pseudo-sequences and proper sequences, and the implication that these differences have on training procedures. A result of our research is that, in case of proper sequences, for learning to be successful the agent must have some kind of memory; moreover it is often necessary to let the trainer and the learner communicate. We study therefore the influence of communication on the learning process. First we consider trainer-to-learner communication introducing the concept of reinforcement sensor, which let the learning robot explicitly know whether the last reinforcement was a reward or a punishment; we also show how the use of this sensor induces the creation of a set of error recovery rules. Then we introduce learner-to-trainer communication, which is used to disambiguate indeterminate training situations, that is situations in which observation alone of the learner behavior does not provide the trainer with enough information to decide if the learner is performing a right or a wrong move. All the design choices we make are discussed and compared by means of experiments in a simulated world.

[208]
Marco Colombetti and Marco Dorigo. Training agents to perform sequential behavior. Adaptive Behavior, 2(3):247-275, 1994. ftp://iridia.ulb.ac.be/pub/dorigo/journals/IJ.06-ADAP94.ps.gz.
This paper is concerned with training an agent to perform sequential behavior. In previous work we have been applying reinforcement learning techniques to control a reactive agent. Obviously, a pure reactive system is limited in the kind of interactions it can learn. In particular, it can learn what we call pseudo-sequences, that is sequences of actions in which each action is selected on the basis of current sensory stimuli; on the contrary, it cannot learn proper sequences, in which actions have to be selected also on the basis of some internal state. Moreover, it is a result of our research that effective learning of proper sequences is improved by letting the agent and the trainer communicate. First we consider trainer-to-agent communication, introducing the concept of reinforcement sensor, which lets the learning robot explicitly know whether the last reinforcement was a reward or a punishment; we also show how the use of this sensor makes error recovery rules emerge. Then we introduce agent-to-trainer communication, which is used to disambiguate ambiguous training situations, that is situations in which the observation of the agent's behavior does not provide the trainer with enough information to decide whether the agent's move is right or wrong. We also show an alternative solution of the problem of ambiguous situations, which involves learning to coordinate behavior in a simpler, unambiguous setting, and then transferring what has been learnt to a more complex situation. All the design choices we make are discussed and compared by means of experiments in a simulated world.

[209]
Marco Colombetti and Marco Dorigo. Verso un'ingegneria del comportamento. Rivista di Automatica, Elettronica e Informatica, 83(10), 1996. In Italian.

[210]
Marco Colombetti and Marco Dorigo. Evolutionary Computation in Behavior Engineering. In Evolutionary Computation: Theory and Applications, chapter 2, pages 37-80. World Scientific Publishing Co.: Singapore, 1999. Also Technical Report. TR/IRIDIA/1996-1, IRIDIA, Université Libre de Bruxelles.

[211]
Marco Colombetti, Marco Dorigo, and G. Borghi. Behavior Analysis and Training: A Methodology for Behavior Engineering. IEEE Transactions on Systems, Man and Cybernetics, 26(6):365-380, 1996.

[212]
Marco Colombetti, Marco Dorigo, and G. Borghi. Robot shaping: The HAMSTER Experiment. In M. Jamshidi et al., editor, Proceedings of ISRAM'96, Sixth International Symposium on Robotics and Manufacturing, May 28-30, Montpellier, France, 1996.

[213]
M. Compiani, D. Montanari, R. Serra, and P. Simonini. Asymptotic dynamics of classifier systems. In Schaffer [718], pages 298-303.
Classifier systems are discussed as high-dimensional dynamical systems. Their learning abilities and long term behavior are analyzed in a letter prediction task domain. We find that the system can develop different types of solutions, sometimes heavily relying on its dynamical properties. A taxonomy of the system solutions is outlined, and some problems due to the activity of the genetic operators are discussed, as well as ways to solve or alleviate them.

[214]
M. Compiani, D. Montanari, R. Serra, and G. Valastro. Classifier systems and neural networks. In Parallel Architectures and Neural Networks-First Italian Workshop, pages 105-118. World Scientific, Teaneck, NJ, 1989.

[215]
M. Compiani, D. Montanari, R. Serra, and P. Simonini. Learning and Bucket Brigade Dynamics in Classifier Systems. In Special issue of Physica D (Vol. 42) [292], pages 202-212.
Classifier systems are rule-based adaptive systems whose learning capabilities emerge from processes of selection and competition within a population of rules (classifiers). These processes are ruled by the values of numerical variables which measure the fitness of each rule. The system's adaptivity is ensured by a fitness reallocation mechanism (the bucket brigade algorithm) and by genetic algorithms which are responsible for the internal dynamics of the system. In this paper we discuss classifier systems as dynamical systems, the main focus being on the asymptotic dynamics due to the bucket brigade abstracting from the action of the genetics. This topic is discussed with reference to a specific task domain, in which the system is used as a detector of statistical properties of periodic or fluctuating external environments. We also describe a major consequence of the genetics on the bucket brigade dynamics, namely the proliferation of individual rules into subpopulations of equivalent classifiers: we then show that this can eventually lead to undesired stochastic behavior or to the destabilizatiion of correct solutions devised by the system.

[216]
Clare Bates Congdon. Classification of epidemiological data: A comparison of genetic algorithm and decision tree approaches. In Proceedings of the 2000 Congress on Evolutionary Computation (CEC00) [185], pages 442-449.
This paper describes an application of genetic algorithms (GA's) to classify epidemiological data, which is often challenging to classify due to noise and other factors. For such complex data (that requires a large number of very specific rules to achieve a high accuracy), smaller rule sets, composed of more general rules, may be preferable, even if they are less accurate. The GA presented here allows the user to encourage smaller rule sets by setting a parameter. The rule sets found are also compared to those created by standard decision-tree algorithms. The results illustrate tradeoffs involving the number of rules, descriptive accuracy, predictive accuracy, and accuracy in describing and predicting positive examples across different rule sets.

[217]
A.L. Corcoran and S. Sen. Using real-valued genetic algorithms to evolve rule sets for classification. In Proceedings of the IEEE Conference on Evolutionary Computation, pages 120-124. IEEE Press, 1994.

[218]
O. Cordón, F. Herrera, E. Herrera-Viedma, and M. Lozano. Genetic Algorithms and Fuzzy Logic in Control Processes. Technical Report DECSAI-95109, University of Granada, Granada, Spain, 1995.
In this paper we describe the genetic algorithms and fuzzy logic, focusing them as tools to model control processes and to design intelligent and automatic control systems. We describe the application of genetic algorithms to design fuzzy logic controllers, as well as the learning classifier systems and their development in a fuzzy environment, the fuzzy learning classifier systems.

[219]
Oscar Cordón, Francisco Herrera, Frank Hoffmann, and Luis Magdalena. Genetic Fuzzy Systems. World Scientific, 2001.
In recent years, a great number of publications have explored the use of genetic algorithms as a tool for designing fuzzy systems. Genetic Fuzzy Systems explores and discusses this symbiosis of evolutionary computation and fuzzy logic. The book summarizes and analyzes the novel field of genetic fuzzy systems, paying special attention to genetic algorithms that adapt and learn the knowledge base of a fuzzy-rule-based system. It introduces the general concepts, foundations and design principles of genetic fuzzy systems and covers the topic of genetic tuning of fuzzy systems. It also introduces the systems: the Michigan, Pittsburgh and Iterative-learning methods. Finally, it explores hybrid genetic fuzzy systems such as genetic fuzzy clustering or genetic neuro-fuzzy systems and describes a number of applications from different areas. Genetic Fuzzy System represents a comprehensive treatise on the design of the fuzzy-rule-based systems using genetic algorithms, both from a theoretical and a practical perspective. It is a valuable compendium for scientists and engineers concerned with research and applications in the domain of fuzzy systems and genetic algorithms.

[220]
H. Brown Cribbs III and Robert E. Smith. Classifier System Renaissance: New Analogies, New Directions. In Koza et al. [527], pages 547-552.
Learning classifier systems (LCSs) have existed for nearly twenty years (Holland & Reitman, 1978). Research efforts in reinforcement learning (RL), evolutionary computation (EC), and neural networks have enhanced the original LCS paradigm. New thoughts from these areas have created a ``renaissance'' period for the LCS. This paper highlights some key LCS advancements and the fields that inspired them. One inspiration, from neural networks, is examined for a novel LCS approach to autonomous mobile robots. A simple, LCS-controlled robot simulation is presented. This simulation shows the potential benefits of combined biological paradigms and the hybridization of ideas in the LCS. Future directions for LCS reseach are discussed.

[221]
Henry Brown Cribbs III and Robert E. Smith. What Can I do with a Learning Classifier System? In C. Karr and L. M. Freeman, editors, Industrial Applications of Genetic Algorithms, pages 299-320. CRC Press, 1998.
The learning classifier system (LCS) is an application of the genetic algorithm to machine learning. Artificial neural networks (ANNs) perform mappings of input vectors to outputs much the same way a LCS does. This chapter introduces the LCS paradigm and provides literature references for future investigation. Through the use of LCS principles, an ANN becomes a variable structure production system, capable of making complex input-output mappnigs that are similar to a LCS. The evolutionary process of a single ANN facilitates a broad understanding of how evolution may help rule-based (or neuron-based) systems. An evolutionary approach to ANN structure is reviewed. Its similarities to the LCS are discussed. A simple extension to Smith and Cribbs' (1994) and Cribbs' (1995) work in an ANN and LCS analogy is presented. The experiment presented removes the nonlinerarity of the ANN's output layer to assess the nonlinear effects of the GA's partitioning within the hidden layer. The results indicate that GA-induced nonlinearity actively participates in the solution of a difficult Boolean problem -- the six multiplexor problem.

[222]
Walling Cyre. Learning grammars with a modified classifier system. In David B. Fogel, Mohamed A. El-Sharkawi, Xin Yao, Garry Greenwood, Hitoshi Iba, Paul Marrow, and Mark Shackleton, editors, Proceedings of the 2002 Congress on Evolutionary Computation CEC2002, pages 1366-1371. IEEE Press, 2002.

[223]
Hai H. Dam, Kamran Shafi, and Hussein A. Abbass. Can Evolutionary Computation Handle Large Datasets? A Study into Network Intrusion Detection. In Proceedings of the 18th Australian Joint Conference on Artificial Intelligence, Lecture Notes in Computer Science (LNCS), 3809, pages 1092-1095. Springer, Heidelberg, 2005.

[224]
H.H. Dam, H.A. Abbass, and C. Lokan. DXCS: an XCS system for distributed data mining. In H.G. Beyer and U.M. O'Reilly, editors, Genetic and evolutionary computation conference, GECCO 2005, pages 1883-1890, 2005.

[225]
Hai Huong Dam, Hussein A. Abbass, Chris Lokan, and Xin Yao. Neural-based learning classifier systems. IEEE Trans. Knowl. Data Eng., 20(1):26-39, 2008.

[226]
Martin Danek and Robert E. Smith. XCS applied to mapping FPGA architectures. In W. B. Langdon, E. Cantú-Paz, K. Mathias, R. Roy, D. Davis, R. Poli, K. Balakrishnan, V. Honavar, G. Rudolph, J. Wegener, L. Bull, M. A. Potter, A. C. Schultz, J. F. Miller, E. Burke, and N. Jonoska, editors, GECCO 2002: Proceedings of the Genetic and Evolutionary Computation Conference, pages 912-919. Morgan Kaufmann Publishers, 2002.

[227]
Dipankar Dasgupta and Fabio A. Gonzalez. Evolving complex fuzzy classifier rules using a linear tree genetic representation. In Lee Spector, Erik D. Goodman, Annie Wu, W.B. Langdon, Hans-Michael Voigt, Mitsuo Gen, Sandip Sen, Marco Dorigo, Shahram Pezeshk, Max H. Garzon, and Edmund Burke, editors, Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2001), pages 299-305, San Francisco, California, USA, 7-11 July 2001. Morgan Kaufmann.

[228]
Y. Davidor and H.-P. Schwefel, editors. Parallel Problem Solving From Nature -- PPSN III, volume 866 of Lecture Notes in Computer Science, Berlin, 1994. Springer Verlag.

[229]
Lawrence Davis and David Orvosh. The Mating Pool: A Testbed for Experiments in the Evolution of Symbol Systems. In Eshelman [298], pages 405-412.

[230]
Lawrence Davis and D. K. Young. Classifier Systems with Hamming Weights. In Proceedings of the Fifth International Conference on Machine Learning, pages 162-173. Morgan Kaufmann, 1988.

[231]
Lawrence Davis, Stewart W. Wilson, and David Orvosh. Temporary Memory for Examples can Speed Learning in a Simple Adaptive System. In Roitblat and Wilson [700], pages 313-320.

[232]
Lawrence Davis, Chunsheng Fu, and Stewart W. Wilson. An incremental multiplexer problem and its uses in classifier system research. In Lanzi et al. [546], pages 23-31.

[233]
Lawrence Davis. Mapping Classifier Systems into Neural Networks. In Proceedings of the Workshop on Neural Information Processing Systems 1, pages 49-56, 1988.

[234]
Lawrence Davis, editor. Genetic Algorithms and Simulated Annealing, Research Notes in Artificial Intelligence. Pitman Publishing: London, 1989.

[235]
Lawrence Davis. Mapping Neural Networks into Classifier Systems. In Schaffer [718], pages 375-378.
Neural networks are machine learning systems based on simple, localized responses to external stimuli. They can respond to the same stimuli that classifier systems respond to, and they alter their internal structure on the basis of reinforcement from an external source. The learning techniques used by researchers in the neural network field have traditionally been quite different from those used by genetic algorithm researchers, however. The tension between these similarities and differences have led researchers to wonder what the formal relationship between the two systems is. This is one of two papers showing that there is a sense in which these two types of machine learning systems are equivalent. In a companion paper, it is shown that any classifier system may be transformed into a neural network that is isomorphic in function. In this paper, it is shown that any neural network can be transformed into a classifier system that is isomorphic in function, although several modifications must be made to standard classifier system practice for this transformation to work. The present paper also considers a different transformation procedure described by Belew and Gherrity that accomplishes this task in a different way. The paper concludes with a discussion of these transformation procedures and their import.

[236]
Lawrence Davis. Covering and Memory in Classifier Systems. In Collected Abstracts for the First International Workshop on Learning Classifier System (IWLCS-92) [486]. October 6-8, NASA Johnson Space Center, Houston, Texas.
Paper is an extended abstract

[237]
Lawrence Davis. A formal relationship between ant colony optimizers and classifier systems. In Tim Kovacs, Xavier LLòra, Keiki Takadama, Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson, editors, Learning Classifier Systems. International Workshops, IWLCS 2003-2005, Revised Selected Papers, volume 4399 of LNCS, pages 258-269. Springer, 2007.

[238]
Devon Dawson and Benjoe Juliano. Modifying xcs for size-constrained systems. International Journal on Neural and Mass-Parallel Computing and Information Systems, 2003.

[239]
Devon Dawson. Improving extended classifier system performance in resource-constrained configurations. Master's thesis, California State University, Chico, 2002.

[240]
Devon Dawson. Improving performance in size-constrained extended classifier systems. In E. Cantú-Paz, J. A. Foster, K. Deb, D. Davis, R. Roy, U.-M. O'Reilly, H.-G. Beyer, R. Standish, G. Kendall, S. Wilson, M. Harman, J. Wegener, D. Dasgupta, M. A. Potter, A. C. Schultz, K. Dowsland, N. Jonoska, and J. Miller, editors, Genetic and Evolutionary Computation -- GECCO-2003, pages 1870-1881, Berlin, 2003. Springer-Verlag.

[241]
Bart de Boer. Classifier Systems: a useful approach to machine learning?. Master's thesis, Leiden University, 1994. ftp://ftp.wi.leidenuniv.nl/pub/CS/MScTheses/deboer.94.ps.gz.
Classifier systems are sub-symbolic or dynamic approaches to machine learning. These systems have been studied rather extensively. In this thesis some theoretical results about the long-term behaviour and the computational abilities of classifier systems are derived. Then some experiments are undertaken. The first experiment entails the implementation of a simple logic function, a multiplexer in a simple classifier system. It is shown that this task can be learned very well. The second task that is taught to the system is a mushroom-classification problem that has been researched with other learning systems. It is shown that this task can be learned. The last problem is the parity problem. First it is shown that this problem does not scale linearly with its number of bits in a straightforward classifier system. An attempt is made to solve it with a multilayer classifier-system, but this is found to be almost impossible. Explanations are given of why this should be the case. Then some thought is given to analogies between classifier systems and neural networks. It is indicated that there are mappings between certain classifier systems and certain neural networks. It is suggested that this is a main concern for future classifier systems research.

[242]
Kenneth A. De Jong, William M. Spears, and Dianna F. Gordon. Using Genetic Algorithms for Concept Learning. Machine Learning, 3:161-188, 13.
In this article, we explore the use of genetic algorithms (GAs) as a key element in the design and implementation of robust concept learning systems. We describe and evaluate a GA-based system called GABIL that continually learns and refines concept classification rules from its interaction with the environment. The use of GAs is motivated by recent studies showing the effects of various forms of bias built into different concept learning systems, resulting in systems that perform well on certain concept classes (generally, those well matched to the biases) and poorly on others. By incorporating a GA as the underlying adaptive search mechanism, we are able to construct a concept learning system that has a simple, unified architecture with several important features. First, the system is surprisingly robust even with minimal bias. Second, the system can be easily extended to incorporate traditional forms of bias found in other concept learning systems. Finally, the architecture of the system encourages explicit representation of such biases and, as a result, provides for an important additional feature: the ability to dynamically adjust system bias. The viability of this approach is illustrated by comparing the performance of GABIL with that of four other more traditional concept learners (AQ14, C4.5, ID5R, and IACL) on a variety of target concepts. We conclude with some observations about the merits of this approach and about possible extensions.

[243]
Kenneth A. De Jong. Learning with Genetic Algorithms: An Overview. Machine Learning, 3:121-138, 1988.
Genetic algorithms represent a class of adaptive search techniques that have been intensively studied in recent years. Much of the interest in genetic algorithms is due to the fact that they provide a set of efficient domain-independent search heuristics which are a significant improvement over traditional ``weak methods'' without the need for incorporating highly domain-specific knowledge. There is now considerable evidence that genetic algorithms are useful for global function optimization and NP-hard problems. Recently, there has been a good deal of interest in using genetic algorithms for machine learning problems. This paper provides a brief overview of how one might use genetic algorithms as a key element in learning systems.

[244]
Michael de la Maza. A SEAGUL Visits the Race Track. In Schaffer [718], pages 208-212.
SEAGUL (Sensitive Evolutionary Adaptable Genetic Unique Learner) is a genetic algorithm that creates production rules that pick the winners of horse races. SEAGUL uses data from the Daily Racing Form, a newspaper that is found at all race tracks and is available to the general public, to generate these rules. SEAGUL deviates from orthodox genetic algorithms in several areas. It has a pre-defined procedure for generating the initial population, it creates inviolable components that cannot be modified through mutation, it does not use the bucket brigade algorithm, and it optimizes its rule set by analyzing variables individually and then collectively.

[245]
K. Deb and W.M. Spears. In: T. back and d.b. fogel and t. michalewicz (eds). evolutionary computation 2: Advanced algorithms and operators, 93-100. institute of physics publishing, 2000.

[246]
Daniel Derrig and James Johannes. Deleting End-of-Sequence Classifiers. In John R. Koza, editor, Late Breaking Papers at the Genetic Programming 1998 Conference, University of Wisconsin, Madison, Wisconsin, USA, July 1998. Stanford University Bookstore.

[247]
Daniel Derrig and James D. Johannes. Hierarchical Exemplar Based Credit Allocation for Genetic Classifier Systems. In Koza et al. [529], pages 622-628.

[248]
L. Desjarlais and Stephanie Forrest. Linked learning in classifier systems: A control architecture for mobile robots. In Collected Abstracts for the First International Workshop on Learning Classifier System (IWLCS-92) [486]. October 6-8, NASA Johnson Space Center, Houston, Texas.
Paper is an extended abstract

[249]
John C. Determan and James A. Foster. A genetic algorithm for expert system rule generation. In Lee Spector, Erik D. Goodman, Annie Wu, W.B. Langdon, Hans-Michael Voigt, Mitsuo Gen, Sandip Sen, Marco Dorigo, Shahram Pezeshk, Max H. Garzon, and Edmund Burke, editors, Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2001), page 757, San Francisco, California, USA, 7-11 July 2001. Morgan Kaufmann.

[250]
P. Devine, R. Paton, and M. Amos. Adaptation of Evolutionary Agents in Computational Ecologies. In BCEC-97, Sweden, 1997.

[251]
Federico Divina and Elena Marchiori. Evolutionary concept learning. In W. B. Langdon, E. Cantú-Paz, K. Mathias, R. Roy, D. Davis, R. Poli, K. Balakrishnan, V. Honavar, G. Rudolph, J. Wegener, L. Bull, M. A. Potter, A. C. Schultz, J. F. Miller, E. Burke, and N. Jonoska, editors, GECCO 2002: Proceedings of the Genetic and Evolutionary Computation Conference, pages 343-350, New York, 9-13 July 2002. Morgan Kaufmann Publishers.

[252]
F. Divina, M. Keijzer, and E. Marchiori. A method for handling numerical attributes in ga-based inductive concept learners. In Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2003), pages 898-908. Springer-Verlag, 2003.

[253]
Phillip William Dixon, David W. Corne, and Martin John Oates. A preliminary investigation of modified xcs as a generic data mining tool. In Lanzi et al. [546], pages 133-150.

[254]
P.W. Dixon, D. Corne, and M.J. Oates. A ruleset reduction algorithm for the XCS learning classifier system. In P.L. Lanzi, W. Stolzmann, and S.W. Wilson, editors, Learning classifier systems, 5th international workshop (IWLCS 2002), volume 2661 of LNCS, pages 20-29. Springer, 2002.

[255]
Jean-Yves Donnart and Jean-Arcady Meyer. A hierarchical classifier system implementing a motivationally autonomous animat. In Cliff et al. [202], pages 144-153.
This work describes a control architecture based on a hierarchical classifier system. This architecture, which uses both reactive and planning rules, implements a motivationally autonomous animat that chooses the actions it will perform according to the expected consequences of the alternatives. The adaptive faculties of this animat are illustrated through various examples.

[256]
Jean-Yves Donnart and Jean-Arcady Meyer. Hierarchical-map Building and Self-positioning with MonaLysa. Adaptive Behavior, 5(1):29-74, 1996.
This paper describes how an animat endowed with the MonaLysa control architecture can build a cognitive map that merges into a hierarchical framework not only topological links between landmarks, but also higher-level structures, control information, and metric distances and orientations. The paper also describes how the animat can use such a map to locate itself, even if it is endowed with noisy dead-reckoning capacities. MonaLysa's mapping and self-positioning capacities are illustrated by results obtained in three different environments and four noise-level conditions. These capacities appear to be gracefully degraded when the environment grows more challenging and when the noise level increases. In the discussion, the current approach is compared to others with similar objectives, and directions for future work are outlined.

[257]
Jean-Yves Donnart and Jean-Arcady Meyer. Learning Reactive and Planning Rules in a Motivationally Autonomous Animat. IEEE Transactions on Systems, Man and Cybernetics - Part B: Cybernetics, 26(3):381-395, 1996.
This work describes a control architecture based on a hierarchical classifier system. This system, which learns both reactive and planning rules, implements a motivationally autonomous animat that chooses the actions it performs according to its perception of the external environment, to its physiological or internal state, to the consequences of its current behavior, and to the expected consequences of its future behavior. The adaptive faculties of this architecture are illustrated within the context of a navigation task, through various experiments with a simulated and a real robot.

[258]
Jean-Yves Donnart and Jean-Arcady Meyer. Spatial Exploration, Map Learning, and Self-Positioning with MonaLysa. In Maes et al. [595], pages 204-213.
This paper describes how the MonaLysa control architecture implements a route-following navigation strategy. Two procedures that allow map building and self-positioning are described, and experimental results are provided that demonstrate that such procedures are robust with respect to noise. This approach is compared to others with similar objectives, and directions for future work are outlined.

[259]
Jean-Yves Donnart. Cognitive Architecture and Adaptive Properties of an Motivationally Autonomous Animat. PhD thesis, Université Pierre et Marie Curie. Paris, France, 1998.
This thesis is centered on MonaLysa, the control architecture of a motivationally autonomous animat. This architecture implements a motivational system that selects the actions and goals of an artificial agent, according to its internal state, the stimuli from the environment, and its evaluation of the long term consequences of its behavioral choices. This architecture is based on an original hierarchical classifier system, that efficiently learns several action plans and builds an internal representation of the animat's environment. The functionalities of MonaLysa are illustrated within the context of the navigation of a simulated animat and of a real robot. In the first part of this work, the animat has to reach a goal and to avoid the obstacles it encounters on its way. We demonstrate that MonaLysa can efficiently learn a general reactive behaviour, notably because it can dynamically change its current goal when the animat encounters an obstacle. Moreover, MonaLysa exploits its interactions with the environment to learn alternative plans and to deduce an optimal path towards its goal; it is also able to modify the organization of its plans so as to adapt to environmental changes. In the second part of this work, the animat has to explore its environment, when various amounts of noise are added to the normal functioning of its odometric sensors. In this context, MonaLysa is able to learn a reliable spatial representation of its environment, while maintaining a correct estimate of its position. This spatial representation is very robust with respect to noise and can adapt to any environment. The generality of the approach presented herein opens the way to many applications, which are outlined at the end of this work.

[260]
Marco Dorigo and Hugues Bersini. A Comparison of Q-Learning and Classifier Systems. In Cliff et al. [202], pages 248-255.
Reinforcement Learning is a class of problems in which an autonomous agent acting in a given environment improves its behavior by progressively maximizing a function calculated just on the basis of a succession of scalar responses received from the environment. Q-learning and classifier systems (CS) are two methods among the most used to solve reinforcement learning problems. Notwithstanding their popularity and their shared goal, they have been in the past often considered as two different models. In this paper we first show that the classifier system, when restricted to a sharp simplification called discounted max very simple classifier system (DMAX-VSCS), boils down to tabular Q-learning. It follows that DMAX-VSCS converges to the optimal policy as proved by Watkins & Dayan (1992), and that it can draw profit from the results of experimental and theoretical works dedicated to improve Q-learning and to facilitate its use in concrete applications. In the second part of the paper, we show that three of the restrictions we need to impose to the CS for deriving its equivalence with Q-learning, that is, no internal states, no don't care symbols, and no structural changes, turn out so essential as to be recently rediscovered and reprogrammed by Q-learning adepts. Eventually, we sketch further similarities among ongoing work within both research contexts. The main contribution of the paper is therefore to make explicit the strong similarities existing between Q-learning and classifier systems, and to show that experience gained with research within one domain can be useful to direct future research in the other one.

[261]
Marco Dorigo and Marco Colombetti. Robot shaping: Developing autonomous agents through learning. Artificial Intelligence, 2:321-370, 1994. ftp://iridia.ulb.ac.be/pub/dorigo/journals/IJ.05-AIJ94.ps.gz.
Learning plays a vital role in the development of situated agents. In this paper, we explore the use of reinforcement learning to shape a robot to perform a predefined target behavior. We connect both simulated and real robots to ALECSYS, a parallel implementation of a learning classifier system with an extended genetic algorithm. After classifying different kinds of Animat-like behaviors, we explore the effects on learning of different types of agent's architecture (monolithic, flat and hierarchical) and of training strategies. In particular, hierarchical architecture requires the agent to learn how to coordinate basic learned responses. We show that the best results are achieved when both the agent's architecture and the training strategy match the structure of the behavior pattern to be learned. We report the results of a number of experiments carried out both in simulated and in real environments, and show that the results of simulations carry smoothly to real robots. While most of our experiments deal with simple reactive behavior, in one of them we demonstrate the use of a simple and general memory mechanism. As a whole, our experimental activity demonstrates that classifier systems with genetic algorithms can be practically employed to develop autonomous agents.

[262]
Marco Dorigo and Marco Colombetti. The Role of the Trainer in Reinforcement Learning. In S. Mahadevan et al., editor, Proceedings of MLC-COLT '94 Workshop on Robot Learning, July 10th, New Brunswick, NJ, pages 37-45, 1994.

[263]
Marco Dorigo and Marco Colombetti. Précis of Robot Shaping: An Experiment in Behavior Engineering. Special Issue on Complete Agent Learning in Complex Environments, Adaptive Behavior, 5(3-4):391-405, 1997.

[264]
Marco Dorigo and Marco Colombetti. Reply to Dario Floreano's ``Engineering Adaptive Behavior''. Special Issue on Complete Agent Learning in Complex Environments, Adaptive Behavior, 5(3-4):417-420, 1997.

[265]
Marco Dorigo and Marco Colombetti. Robot Shaping: An Experiment in Behavior Engineering. MIT Press/Bradford Books, 1998.

[266]
Marco Dorigo and V. Maniezzo. Parallel Genetic Algorithms: Introduction and Overview of Current Research. In J. Stenders, editor, Parallel Genetic Algorithms: Theory and Applications, Amsterdam, 1992. IOS Press.

[267]
Marco Dorigo and U. Schnepf. Organisation of Robot Behaviour Through Genetic Learning Processes. In Proceedings of ICAR'91 -- Fifth IEEE International Conference on Advanced Robotics, Pisa, Italy, pages 1456-1460. IEEE Press, 1991.

[268]
Marco Dorigo and U. Schnepf. Genetics-based Machine Learning and Behaviour Based Robotics: A New Synthesis. IEEE Transactions on Systems, Man and Cybernetics, 23(1):141-154, 1993.
Intelligent robots should be able to use sensor information to learn how to behave in a changing environment. As environmental complexity grows, the learning task becomes more and more difficult. We face this problem using an architecture based on learning classifier systems and on the structural properties of animal behavioural organization, as proposed by ethologists. After a description of the learning technique used and of the organizational structure proposed, we present experiments that show how behaviour acquisition can be achieved. Our simulated robot learns to follow a light and to avoid hot dangerous objects. While these two simple behavioural patterns are independently learnt, coordination is attained by means of a learning coordination mechanism. Again this capacity is demonstrated by performing a number of experiments

[269]
Marco Dorigo and E. Sirtori. A Parallel Environment for Learning Systems. In Proceedings of GAA91 -- Second Italian Workshop on Machine Learning, Bari, Italy, 1991.

[270]
Marco Dorigo and Enrico Sirtori. Alecsys: A Parallel Laboratory for Learning Classifier Systems. In Booker and Belew [74], pages 296-302.
A major problem with learning systems is how to tackle real world problems. A distinctive characteristic of many real world problems is that they present a complexity that cannot be ``user-defined'', and which is generally orders of magnitude higher than in toy systems. The use of more powerful, parallel machines, is a way to attack this problem from two sides: through an increase in the performance of standard algorithms, and by design of a new structural organization of the learning system -- organization that should allow a better control on the environmental complexity. In order to explore these potentialities we have built a tool, ALECSYS, that can be used to implement parallel learning classifier systems in a modular fashion. In ALECSYS parallelism is used both to increase the system performance, by what we call low-level parallelization, and to allow the use of many different learning classifier systems simultaneously, by what we call high-level parallelization. In the paper we first present the system organization and the algorithms used, then we report some simulation results and finally we give some hints for further work.

[271]
Marco Dorigo, V. Maniezzo, and D. Montanari. Classifier-based robot control systems. In IFAC/IFIP/IMACS International Symposium on Artificial Intelligence in Real-Time Control, pages 591-598, Delft, Netherlands, 1992.

[272]
Marco Dorigo, Mukesh J. Patel, and Marco Colombetti. The effect of Sensory Information on Reinforcement Learning by a Robot Arm. In M. Jamshidi et al., editor, Proceedings of ISRAM'94, Fifth International Symposium on Robotics and Manufacturing, August 14-18, Maui, HI, pages 83-88. ASME Press, 1994.

[273]
Marco Dorigo. Message-Based Bucket Brigade: An Algorithm for the Apportionment of Credit Problem. In Y. Kodratoff, editor, Proceedings of European Working Session on Learning '91, Porto, Portugal, number 482 in Lecture notes in Artificial Intelligence, pages 235-244. Springer-Verlag, 1991.

[274]
Marco Dorigo. New perspectives about default hierarchies formation in learning classifier systems. In E. Ardizzone, E. Gaglio, and S. Sorbello, editors, Proceedings of the 2nd Congress of the Italian Association for Artificial Intelligence (AI*IA) on Trends in Artificial Intelligence, volume 549 of LNAI, pages 218-227, Palermo, Italy, October 1991. Springer Verlag.
A major problem with learning systems is how to tackle real world problems. A distinctive characteristic of many real world problems is that they present a complexity that cannot be user-defined, and which is generally orders of magnitude higher than in toy systems. The use of more powerful, parallel machines, is a way to attack this problem from two sides: through an increase in the performance of standard algorithms, and by design of a new structural organization of the learning system - organization that should allow a better control on the environmental complexity. In order to explore these potentialities we have built a tool, ALECSYS, that can be used to implement parallel learning classifier systems in a modular fashion. In ALECSYS parallelism is used both to increase the system performance, by what we call low-level parallelization, and to allow the use of many different learning classifier systems simultaneously, by what we call high-level parallelization. In the paper we first present the system organization and the algorithms used, then we report some simulation results and finally we give some hints for further work.

[275]
Marco Dorigo. Using Transputers to Increase Speed and Flexibility of Genetic-based Machine Learning Systems. Microprocessing and Microprogramming, 34:147-152, 1991.

[276]
Marco Dorigo. Alecsys and the AutonoMouse: Learning to Control a Real Robot by Distributed Classifier Systems. Technical Report 92-011, Politecnico di Milano, 1992.

[277]
Marco Dorigo. Optimization, Learning and Natural Algorithms. PhD thesis, Politecnico di Milano, Italy, 1992. (In Italian).

[278]
Marco Dorigo. Genetic and Non-Genetic Operators in ALECSYS. Evolutionary Computation, 1(2):151-164, 1993. Also Technical Report TR-92-075 International Computer Science Institute.
It is well known that standard learning classifier systems, when applied to many different domains, exhibit a number of problems: payoff oscillation, difficulty in regulating interplay between the reward system and the background genetic algorithm (GA), rule chains' instability, default hierarchies' instability, among others. ALECSYS is a parallel version of a standard learning classifier system (CS) and, as such, suffers from these same problems. In this paper we propose some innovative solutions to some of these problems. We introduce the following original features. Mutespec is a new genetic operator used to specialize potentially useful classifiers. Energy is a quantity introduced to measure global convergence to apply the genetic algorithm only when the system is close to a steady state. Dynamic adjustment of the classifiers set cardinality speeds up the performance phase of the algorithm. We present simulation results of experiments run in a simulated two-dimensional world in which a simple agent learns to follow a light source.

[279]
Marco Dorigo. Gli Algoritmi Genetici, i Sistemi a Classificatori e il Problema dell'Animat. Sistemi Intelligenti, 3(93):401-434, 1993. In Italian.

[280]
Marco Dorigo. Alecsys and the AutonoMouse: Learning to Control a Real Robot by Distributed Classifier Systems. Machine Learning, 19:209-240, 1995.
In this article we investigate the feasibility of using learning classifier systems as a tool for building adaptive control systems for real robots. Their use on real robots imposes efficiency constraints which are addressed by three main tools: parallelism, distributed architecture, and training. Parallelism is useful to speed up computation and to increase the flexibility of the learning system design. Distributed architecture helps in making it possible to decompose the overall task into a set of simpler learning tasks. Finally, training provides guidance to the system while learning, shortening the number of cycles required to learn. These tools and the issues they raise are first studied in simulation, and then the experience gained with simulations is used to implement the learning system on the real robot. Results have shown that with this approach it is possible to let the AutonoMouse, a small real robot, learn to approach a light source under a number of different noise and lesion conditions.

[281]
Marco Dorigo. The Robot Shaping Approach to Behavior Engineering. Thése d'Agrégation de l'Enseignement Supérieur, Faculté des Sciences Appliquées, Université Libre de Bruxelles, pp.176, 1995.

[282]
J. Drugowitsch and A. Barry. XCS with eligibility traces. In H.G. Beyer and U.M. O'Reilly, editors, Genetic and evolutionary computation conference, GECCO 2005, pages 1851-1858. ACM, 2005.

[283]
Jan Drugowitsch and Alwyn Barry. A Formal Framework and Extensions for Function Approximation in Learning Classifier Systems. Machine Learning, 70(1):45-88, 2007.
Learning Classifier Systems (LCS) consist of the three components: function approximation, reinforcement learning, and classifier replacement. In this paper we formalize the function approximation part, by providing a clear problem definition, a formalization of the LCS function approximation architecture, and a definition of the function approximation aim. Additionally, we provide definitions of optimality and what conditions need to be fulfilled for a classifier to be optimal. As a demonstration of the usefulness of the framework, we derive commonly used algorithmic approaches that aim at reaching optimality from first principles, and introduce a new Kalman filter-based method that outperforms all currently implemented methods, in addition to providing further insight into the probabilistic basis of the localized model that a classifier provides. A global function approximation in LCS is achieved by combining the classifier's localized model, for which we provide a simplified approach when compared to current LCS, based on the Maximum Likelihood of a combination of all classifiers. The formalizations in this paper act as the foundation of a currently actively developed formal framework that includes all three LCS components, promising a better formal understanding of current LCS and the development of better LCS algorithms.

[284]
Jan Drugowitsch and Alwyn M. Barry. A principled foundation for lcs. In Jaume Bacardit, Ester Bernadó-Mansilla, Martin Butz, Tim Kovacs, Xavier Llorà, and Keiki Takadama, editors, Learning Classifier Systems. 10th and 11th International Workshops (2006-2007), volume 4998/2008 of Lecture Notes in Computer Science, pages 77-95. Springer, 2008.
In this paper we promote a new methodology for designing LCS that is based on first identifying their underlying model and then using standard machine learning methods to train this model. This leads to a clear identification of the LCS model and makes explicit the assumptions made about the data, as well as promises advances in the theoretical understanding of LCS through transferring the understanding of the applied machine learning methods to LCS. Additionally, it allows us, for the first time, to give a formal and general, that is, representation-independent, definition of the optimal set of classifiers that LCS aim at finding. To demonstrate the feasibility of the proposed methodology we design a Bayesian LCS model by borrowing concepts from the related Mixtures-of-Experts model. The quality of a set of classifiers and consequently also the optimal set of classifiers is defined by the application of Bayesian model selection, which turns finding this set into a principled optimisation task. Using a simple Pittsburgh-style LCS, a set of preliminary experiments demonstrate the feasibility of this approach.

[285]
Jan Drugowitsch. Design and Analysis of Learning Classifier Systems: A Probabilistic Approach. Springer, 2008.
This book provides a comprehensive introduction to the design and analysis of Learning Classifier Systems (LCS) from the perspective of machine learning. LCS are a family of methods for handling unsupervised learning, supervised learning and sequential decision tasks by decomposing larger problem spaces into easy-to-handle subproblems. Contrary to commonly approaching their design and analysis from the viewpoint of evolutionary computation, this book instead promotes a probabilistic model-based approach, based on their defining question "What is an LCS supposed to learn?". Systematically following this approach, it is shown how generic machine learning methods can be applied to design LCS algorithms from the first principles of their underlying probabilistic model, which is in this book -- for illustrative purposes -- closely related to the currently prominent XCS classifier system. The approach is holistic in the sense that the uniform goal-driven design metaphor essentially covers all aspects of LCS and puts them on a solid foundation, in addition to enabling the transfer of the theoretical foundation of the various applied machine learning methods onto LCS. Thus, it not only advances the analysis of existing LCS but also puts forward the design of new LCS within that same framework.

[286]
Barry B. Druhan and Robert C. Mathews. THIYOS: A Classifier System Model of Implicit Knowledge in Artificial Grammars. In Proc. Ann. Cog. Sci. Soc., 1989.

[287]
D. Dumitrescu, B. Lazzerini, L. C. Jain, and A. Dumitrescu. Evolutionary Computation. CRC Press International, 2000.

[288]
Daniel Eckert and Johann Mitlöhner. Modelling individual and endogenous learning in games: the relevance of classifier systems. In Complex Modelling for Socio-Economic Systems, SASA, Vienna, 1997.

[289]
Daniel Eckert, Johann Mitlöhner, and Makus Moschner. Evolutionary stability issues and adaptive learning in classifier systems. In OR'97 Conference on Operations Research, Vienna, 1997.
This work has no abstract.

[290]
Narayanan Edakunni, Tim Kovacs, Gavin Brown, and James Marshall. Modeling UCS as a mixture of experts. In Proceedings of the 2009 Genetic and Evolutionary Computation Conference (GECCO'09). ACM, 2009.
We present a probabilistic formulation of UCS (a sUpervised Classifier System). UCS is shown to be a special case of mixture of experts where the experts are learned independently and later combined during prediction. In this work, we develop the links between the constituent components of UCS and a mixture of experts, thus lending UCS a strong analytical background. We find during our analysis that mixture of experts is a more generic formulation of UCS and possesses more generalization capability and flexibility than UCS, which is also verified using empirical evaluations. This is the first time that a simple probabilistic model has been proposed for UCS and we believe that this work will form a useful tool to analyse Learning Classifier Systems and gain useful insights into their working.

[291]
A. E. Eiben and J. E. Smith. Introduction to evolutionary computation. Natural computing series. Springer-Verlag, 2003.

[292]
Emergent Computation. Proceedings of the Ninth Annual International Conference of the Center for Nonlinear Studies on Self-organizing, Collective, and Cooperative Phenomena in Natural and Artificial Computing Networks. A special issue of Physica D. Stephanie Forrest (Ed.), 1990.

[293]
G. Enee and C. Escazut. Classifier systems evolving multi-agent system with distributed elitism. In Angeline et al. [8], pages 1740-1745.

[294]
G. Enee and C. Escazut. A minimal model of communication for a multi-agent classifier system. In Lanzi et al. [546], pages 32-42.

[295]
Cathy Escazut and Philippe Collard. Learning Disjunctive Normal Forms in a Dual Classifier System. In Nada Lavrac and Stefan Wrobel, editors, Proceedings of the 8th European Conference on Machine Learning, volume 912 of LNAI, pages 271-274. Springer, 1995.

[296]
Cathy Escazut and Terence C. Fogarty. Coevolving Classifier Systems to Control Traffic Signals. In John R. Koza, editor, Late Breaking Papers at the 1997 Genetic Programming Conference, Stanford University, CA, USA, July 1997. Stanford Bookstore.

[297]
Cathy Escazut, Philippe Collard, and Jean-Louis Cavarero. Dynamic Management of the Specificity in Classifier Systems. In Albrecht et al. [7], pages 484-491.
The estimation of the rule usefulness in a classifier system is faced to the credit-apportionment problem. Usually, the apportioning of payoffs process is performed by the bucket brigade algorithm. However, some works have shown that this algorithm presents some difficulties. Generally, the condition part of a rule is defined on an alphabet containing a ``don't care'' symbol. That is why a same rule can be fired in different contexts. In such conditions, it is impossible to use too generalized classifiers because of the incoherence of the strength management. The solution we propose here, can solve the problem: general classifiers belonging to a success-ending sequence are dynamically specialized. In order not to store all the sequence actions, the Bucket Brigade algorithm is applied to the new-created rule specificity. So, the closer a classifier is from the end of the solution sequence, the more specific it is. This new algorithm is presented here and applied to an autonomous moving robot which must learn how to move in an environment with obstacles.

[298]
Larry J. Eshelman, editor. Proceedings of the 6th International Conference on Genetic Algorithms (ICGA95). Morgan Kaufmann Publishers, 1995.

[299]
J. A. Meyer et al., editor. From Animals to Animats 6: Proceedings of the Sixth International Conference on Simulation of Adaptive Behavior, 2000.

[300]
Andrew Fairley and Derek F. Yates. Improving Simple Classifier Systems to alleviate the problems of Duplication, Subsumption and Equivalence of Rules. In Albrecht et al. [7], pages 408-416.
For new, potentially improved rules that is, the search performed by a classifier system's genetic algorithm is guided by the relative strength of the rules in the extant rule base. This paper identifies three general types of rule whose presence in a plan can affect the relative strength of rules in a rule base and thereby provide the potential to compromise the effectiveness of the genetic algorithm. The nature and extent of relative strength distortion is investigated and a method to combat the distortion which involves adaptation of the standard bucket brigade, is proposed.

[301]
Andrew Fairley and Derek F. Yates. Inductive Operators and Rule Repair in a Hybrid Genetic Learning System: Some Initial Results. In Fogarty [324], pages 166-179.
Symbolic knowledge representation schemes have been suggested as one way to improve the performance of classifier systems in the context of complex, real-world problems. The main reason for this is that unlike the traditional binary string representation, high-level languages facilitate the exploitation of problem specific knowledge. However, the two principle genetic operators, crossover and mutation, are, in their basic form, ineffective with regard to discovering useful rules in such representations. Moreover, the operators do not take into account any environmental cues which may benefit the rule discovery process. A further source of inefficiency in classifier systems concerns their capacity for forgetting valuable experience by deleting previously useful rules. In this paper, solutions to both these problems are addressed. First, in respect of the suitability of crossover and mutation, a new set of operators, specifically tailored for a high level language, are proposed. Moreover, to alleviate the problem of forgetfulness, an approach based on the way some enzyme systems facilitate the repair of genes in biological systems, is investigated.

[302]
I. De Falco, A. Iazzetta, E. Tarantino, and A. Della Cioppa. An evolutionary system for automatic explicit rule extraction. In Proceedings of the 2000 Congress on Evolutionary Computation (CEC00) [185], pages 450-457.
The search for novel and useful patterns within large databases, known as data mining, has become of great importance owing to the ever-increasing amounts of data collected by large organizations. In particular the emphasis is devoted to heuristic search methods able to discover patterns that are hard or impossible to detect using standard query mechanisms and classicial statistical techniques. In this paper an evolutionary system capable of extracting explicit classification rules is presented. The results are compared with those obtained by other approaches.

[303]
William Joseph Falke II and Peter Ross. Dynamic strategies in a real-time strategy game. In E. Cantú-Paz, J. A. Foster, K. Deb, D. Davis, R. Roy, U.-M. O'Reilly, H.-G. Beyer, R. Standish, G. Kendall, S. Wilson, M. Harman, J. Wegener, D. Dasgupta, M. A. Potter, A. C. Schultz, K. Dowsland, N. Jonoska, and J. Miller, editors, Genetic and Evolutionary Computation -- GECCO-2003, LNCS, pages 1920-1921. Springer-Verlag, 2003.

[304]
J. Doyne Farmer, N. H. Packard, and A. S. Perelson. The Immune System, Adaptation & Learning. Physica D, 22:187-204, 1986.

[305]
J. Doyne Farmer. A Rosetta Stone for Connectionism. In Special issue of Physica D (Vol. 42) [292], pages 153-187.
The term connectionism is usually applied to neural networks. There are, however, many other models that are mathematically similar, including classifier systems, immune networks, autocatalytic chemical reaction networks, and others. In view of this similarity, it is appropriate to broaden the term connectionism. I define a connectionist model as a dynamical system with two properties: (1) The interactions between the variables at any given time are explicitly constrained to a finite list of connections. (2) The connections are fluid, in that their strength and/or pattern of connectivity can change with time. This paper reviews the four examples listed above and maps them into a common mathematical framework, discussing their similarities and differences. It also suggests new applications of connectionist models, and poses some problems to be addressed in an eventual theory of connectionist systems.

[306]
Francine Federman and Susan Fife Dorchak. Information Theory and NEXTPITCH: A Learning Classifier System. In Bäck [24], pages 442-449.
NEXTPITCH, a learning classifier system using genetic algorithms, inductively learns to predict the next note in a childhood melody. Just as a listener develops expectations of what is to follow based on what has been heard, NEXTPITCH models human music learning by developing the rules that represent actual pitch transitions in the melody. This paper introduces our representation of music and our description of classifier formats. Our results are correlated by analysis of variance statistical routines. Information theory is used to partition melodies into classes so that we may examine the applicability of the results from one set of melodies to another.

[307]
Francine Federman and Susan Fife Dorchak. Representation of Music in a Learning Classifier System. In Rad and Skowron, editors, Foundations of Intelligent Systems: Proceedings 10th International Symposium (ISMIS'97). Springer-Verlag: Heidelberg, 1997.

[308]
Francine Federman and Susan Fife Dorchak. A Study of Classifier Length and Population Size. In Koza et al. [529], pages 629-634.

[309]
Francine Federman, Gayle Sparkman, and Stephanie Watt. Representation of Music in a Learning Classifier System Utilizing Bach Chorales. In Banzhaf et al. [32], page 785. One page poster paper.
This paper addresses the question of the impact of the representation of a domain on the performance in a learning classifier system. NEXTPITCH, a learning classifier system using genetic algorithms, inductively learns to predict the next note of a Bach chorale. This paper presents an analyses of different representations of specific features of Western tonal music. Our results are correlated by analyses of variance statistical routines.

[310]
Francine Federman. NEXTNOTE: A Learning Classifier System. In Annie S. Wu, editor, Proceedings of the Genetic and Evolutionary Computation Conference Workshop Program, pages 136-138, 2000.
NEXTPITCH, a learning classifier system (LCS) using genetic algorithms, inductively learns to predict the next note in a melody. This paper addresses the issues of (1) the impact of the representation of a domain on the performance of an LCS and (2) the classification of the input to an LCS in order to determine performance.

[311]
Thomas Fent. Applications of Learning Classifier Systems for Simulating Learning Organizations, volume 10 of Fortschrittsberichte Simulation. ARGESIM / ASIM-Verlag, Wien, September 2001.
In the field of mathematical modelling of economic phenomena analytical models, assuming the existence of a stable equilibrium, are very popular. It is often expected, that the quantities defining the state of the system gradually approach to the equilibrium, and remain there in the following. Moreover, it is often assumed, that the interacting individuals behave perfectly rational --- i.e they always take those decisions, that maximize their utility. Such approaches, however, have only limited suitability for modelling economic systems subject to a permanent and succesively accelerated change. Moreover, they are only partially capable to describe individuals that do not only possess intelligence but also emotions. Therefore, the theory of complex adaptive systems applies computer simulations whose implementations does not necessaryly depend on the existence of equilibria or perfectly ratinal agents. At the beginning of this thesis the necessity and the practicality of the employment of complex adaptive systems for describing recent economic happenings is discussed. Then methods that are qualified to implement such systems --- in particular classifier systems and genetic algorithms --- are being explained. In the following some examples are provided to illustrate possibilities and also restrictions of the usage of such procedures. Based on this elaborations two comprehensive economic models are formulated. The first model is about the problem of communication within a firm developing a new product. In big enterprises it is often a big challenge to establish an effective and efficient flow of information. Moreover, in distributed decision making processes conflicting objectives may occur. Many different groups of employees are cooperating in the process of designing a new product. The tasks reqired to successfully introduce a new product involve employees from market research, engineering, scheduling, maintenance, and many others. These people may possess different viewpoints and different technical languages. To solve this problem citep *hauser suggested a communication scheme called ``House of Quality''. This thesis introduces a simulation based on the ``House of Quality'', which was implemented in MATLAB. The decision makers are implemented as classifier systems, and apply genetic algorithms to learn a meaningful solution. To evaluate the rules generated by this adaptive learning process, the obtained results are compared with the results gained by full enumeration. It turns out that the genetic algorithms indeed create pretty good decision rules. These rules illustrate how the responsible individuals should react to the encountered situations. The second model examines a heterogenous market of goods that can be substituted for each other. The task of the learning agents is to place there products in the market such that the number of customers who buy their products is maximized. Four classes of different agents occur in this market. Two of these groups contain learning agents. The first group observes the positions (= the needs) of the customers directly. The second group observes the positions of the suppliers and their profits. The decisions for the next planning period are based on these observations. Additionally, there is a group of suppliers placing their products according to a ``random walk'', and another supplier who always takes over the position of the most successful seller of the previous period. To compare these strategies three different behaviour patterns of the demand side are taken into consideration: i) static, ii) cyclic, and iii) random walk. To allow accurate conclusions about the relation between customer behaviour and the success of a certain selling strategy, all the customers exhibit the same behaviour pattern within one particular simulation. It turned out, that in the cases i) and iii) the strategy of imitating the most successful seller is optimal --- under the assumption, that only one supplier follows that strategy. If the customers behave according to ii), on the other hand, the agents observing the customers directly are the winners.

[312]
Rhonda Ficek. Genetic Algorithms. Technical Report NDSU-CS-TR-90-51, North Dakota State University. Computer Science and Operations Research, 1997.
Genetic algorithms are generating a great deal of interest today. They take their inspiration from the ways plants and animals evolve. Developed by John Holland, genetic algorithms use the ideas and language of genetics -- genes, chromosomes, mutations, etc. Holland's unique method for performing adaptive searches has received a great deal of attention, and numerous applications which make use of genetic algorithms have been developed. The algorithm behind evolution solves the problem of producing species able to thrive in particular environments. This same genetic algorithm can solve many other kinds of problems as well. It forms the basis of an emerging field of computer programming that could challenge expert systems. Research is under way on adapting the genetic algorithm to such applications as running pump stations on oil pipelines, eliminating distortion from x-ray images, and designing very-large-scale integrated (VLSI) computer chips. This paper begins with an overview of genetic algorithms, followed by a summary of the history of their development. The next three sections of the paper discuss the major areas of genetic algorithm research: applications of genetic algorithms, theoretical work, and classifier systems. Next, genetic algorithms are compared to other related methods, and their strengths are discussed. The final section discusses the future of genetic algorithms.

[313]
M. V. Fidelis, H. S. Lopes, and A. A. Freitas. Discovering comprehensible classification rules with a genetic algorithm. In Proceedings of the 2000 Congress on Evolutionary Computation (CEC00) [185], pages 805-810.

[314]
Gary William Flake. The Computational Beauty of Nature. MIT Press, 1998. (Contains a chapter on ZCS).

[315]
Peter Fletcher. Simulating the use of `fiat money' in a simple commodity economy. Master's thesis, Schools of Psychology and Computer Science, University of Birmingham, 1996.
This project simulates a simple commodity economy in which artificially intelligent adaptive agents learn to trade with one another. Two versions of the economy are studied in detail. The first consists of three types of agent and three types of commodity, and the second version includes the addition of `fiat money'. The agents make decisions using a classifier system, capable of being rewarded and punished according to the relative success of the economic strategies generated. The economic environment that the agents inhabit is taken from Marimon et al(1989) but a different classifier system is used, to investigate whether an alternative implementation affects the results of the simulation. The results were not fully replicated, and the differences between the two implementations are analysed, giving rise to novel observations about the environment and the original research.

[316]
Terence C. Fogarty and Luis Miramontes Hercog. Social simulation using a multi-agent model based on classifier systems: The emergence of switching agents in the dual pub problem. In Erik D. Goodman, editor, 2001 Genetic and Evolutionary Computation Conference Late Breaking Papers, pages 87-94, 2001.

[317]
Terence C. Fogarty, Brian Carse, and Larry Bull. Classifier Systems -- recent research. AISB Quarterly, 89:48-54, 1994.
We consider; the role of selectionist reinforcement learning in classifier systems; the fuzzy matching and activation of rules, and the evolution of communication within and between classifier systems.

[318]
Terence C. Fogarty, Larry Bull, and Brian Carse. Evolving Multi-Agent Systems. In J. Periaux and G. Winter, editors, Genetic Algorithms in Engineering and Computer Science, pages 3-22. John Wiley & Sons, 1995.

[319]
Terence C. Fogarty, Brian Carse, and Larry Bull. Classifier Systems: selectionist reinforcement learning, fuzzy rules and communication. Presented at the First International Workshop on Biologically Inspired Evolutionary Systems, Tokyo, 1995.

[320]
Terence C. Fogarty, N. S. Ireson, and Larry Bull. Genetic-based Machine Learning -- Applications in Industry and Commerce. In Vic Rayward-Smith, editor, Applications of Modern Heuristic Methods, pages 91-110. Alfred Waller Ltd, 1995.
This paper describes how genetic algorithms are being used to engineer control and classification systems for industrial and commercial applications. Using multiple burner combustion control and credit risk assessment as examples it illustrates how expert human knowledge can be complemented by searching large amounts of data using genetic algorithms in knowledge-based machine learning systems. It goes on to discuss recent research on parallel and distributed genetic algorithms aimed at tackling large complex problems.

[321]
Terence C. Fogarty, Brian Carse, and A. Munro. Artificial evolution of fuzzy rule bases which represent time: A temporal fuzzy classifier system. International Journal of Intelligent Systems, 13(10-11):906-927, 1998.

[322]
T.C. Fogarty. An incremental genetic algorithm for real-time learning. In Proc. Sixth Int. Workshop on Machine Learning, pages 416-419, 1989.

[323]
Terence C. Fogarty. Co-evolving Co-operative Populations of Rules in Learning Control Systems. In Evolutionary Computing, AISB Workshop Selected Papers [324], pages 195-209.
It is shown how co-evolving populations of individual rules can outperform evolving a population of complete sets of rules with the genetic algorithm in learning classifier systems. A rule-based control system is presented which uses only the genetic algorithm for learning individual control rules with immediate reinforcement after the firing of each rule. How this has been used for an industrial control problem is described as an example of its operation. The refinement of the system to deal with delayed reward is presented and its operation on the cart-pole balancing problem described. A comparison is made of the performance of the refined system using only selection and mutation to learn individual rules with that of the genetic algorithm to learn a complete set of rules. A comparison is also made of the performance of the refined system using only selection to learn individual rules with that of the bucket-brigade and other reinforcement algorithms on the same task.

[324]
Terence C. Fogarty, editor. Evolutionary Computing, AISB Workshop Selected Papers, number 865 in Lecture Notes in Computer Science. Springer-Verlag, 1994.

[325]
Terence C. Fogarty. Learning new rules and adapting old ones with the genetic algorithm. In G. Rzevski, editor, Artificial Intelligence in Manufacturing, pages 275-290. Springer-Verlag, 1994.

[326]
Terence C. Fogarty. Optimising Individual Control Rules and Multiple Communicating Rule-based Control Systems with Parallel Distributed Genetic Algorithms. IEE Journal of Control Theory and Applications, 142(3):211-215, 1995.

[327]
Terence C. Fogarty. Genetic algorithms for the optimization of combustion in multiple-burner furnaces and boiler plants. In Thomas Bäck, David B. Fogel, and Zbigniew Michalewicz, editors, Handbook of Evolutionary Computation, pages G3.2:1-G3.2:7. IOP Publishing Ltd and Oxford University Press, 1997.

[328]
David B. Fogel. Evolutionary Computation. The Fossil Record. Selected Readings on the History of Evolutionary Computation, chapter 16: Classifier Systems. IEEE Press, 1998. This is a reprint of (Holland and Reitman, 1978), with an added introduction by Fogel.

[329]
Stephanie Forrest and John H. Miller. Emergent behavior in classifier systems. In Special issue of Physica D (Vol. 42) [292], pages 213-217.
The paper presents examples of emergent behavior in classifier systems, focusing on symbolic reasoning and learning. These behaviors are related to global dynamical properties such as state cycles, basins of attraction, and phase transitions. A mapping is defined between classifier systems and an equivalent dynamical system (Boolean networks). The mapping provides a way to understand and predict emergent classifier system behaviors by observing the dynamical behavior of the Boolean networks. The paper reports initial results and discusses the implications of this approach for classifier systems.

[330]
Stephanie Forrest, Robert E. Smith, and A. Perelson. Maintaining diversity with a genetic algorithm. In Collected Abstracts for the First International Workshop on Learning Classifier System (IWLCS-92) [486]. October 6-8, NASA Johnson Space Center, Houston, Texas.
Paper is an extended abstract

[331]
Stephanie Forrest. A study of parallelism in the classifier system and its application to classification in KL-ONE semantic networks. PhD thesis, University of Michigan, Ann Arbor, MI, 1985.

[332]
Stephanie Forrest. Implementing semantic network structures using the classifier system. In Grefenstette [389], pages 24-44.

[333]
Stephanie Forrest. The Classifier System: A Computational Model that Supports Machine Intelligence. In International Conference on Parallel Processing, pages 711-716, Los Alamitos, Ca., USA, August 1986. IEEE Computer Society Press.

[334]
Stephanie Forrest. Parallelism and Programming in Classifier Systems. Pittman, London, 1991.

[335]
Stephanie Forrest, editor. Proceedings of the 5th International Conference on Genetic Algorithms (ICGA93). Morgan Kaufmann, 1993.

[336]
Richard Forsyth. Machine Learning: Applications in expert systems and information retrival, chapter Evolutionary Learning Strategies, pages 78-95. Ellis Horwood Limited, 1986.

[337]
Peter W. Frey and David J. Slate. Letter Recognition Using Holland-Style Adaptive Classifiers. Machine Learning, 6:161-182, 1991.
Machine rule induction was examined on a difficult categorization problem by applying a Holland-style classifier system to a complex letter recognition task. A set of 20,000 unique letter images was generated by randomly distorting pixel images of the 26 uppercase letters from 20 different commercial fonts. The parent fonts represented a full range of character types including script, italic, serif, and Gothic. The features of each of the 20,000 characters were summarized in terms of 16 primitive numerical attributes. Our research focused on machine induction techniques for generating IF-THEN classifiers in which the IF part was a list of values for each of the 16 attributes and the THEN part was the correct category, i.e., one of the 26 letters of the alphabet. We examined the effects of different procedures for encoding attributes, deriving new rules, and apportioning credit among the rules. Binary and Gray-code attribute encodings that required exact matches for rule activation were compared with integer representations that employed fuzzy matching for rule activation. Random and genetic methods for rule creation were compared with instance-based generalization. The strength/specificity method for credit apportionment was compared with a procedure we call ``accuracy/utility.''

[338]
Chunsheng Fu and Lawrence Davis. A modified classifier system compaction algorithm. In W. B. Langdon, E. Cantú-Paz, K. Mathias, R. Roy, D. Davis, R. Poli, K. Balakrishnan, V. Honavar, G. Rudolph, J. Wegener, L. Bull, M. A. Potter, A. C. Schultz, J. F. Miller, E. Burke, and N. Jonoska, editors, GECCO 2002: Proceedings of the Genetic and Evolutionary Computation Conference, pages 920-925. Morgan Kaufmann Publishers, 2002.

[339]
Chunsheng Fu, Stewart W. Wilson, and Lawrence Davis. Studies of the xcsi classifier system on a data mining problem. In Lee Spector, Erik D. Goodman, Annie Wu, W.B. Langdon, Hans-Michael Voigt, Mitsuo Gen, Sandip Sen, Marco Dorigo, Shahram Pezeshk, Max H. Garzon, and Edmund Burke, editors, Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2001), page 985, San Francisco, California, USA, 7-11 July 2001. Morgan Kaufmann.

[340]
Leeann L. Fu. The XCS Classifier System and Q-learning. In John R. Koza, editor, Late Breaking Papers at the Genetic Programming 1998 Conference, University of Wisconsin, Madison, Wisconsin, USA, 1998. Stanford University Bookstore.

[341]
Leeann L. Fu. What I have come to understand about classifier systems, 1998. Unpublished document. Dept. of Electrical Engineering and Computer Science. University of Michigan.
This document has no abstract.

[342]
Takeshi Furuhashi, Ken Nakaoka, Koji Morikawa, and Yoshiki Uchikawa. Controlling Excessive Fuzziness in a Fuzzy Classifier System. In Forrest [335], pages 635-635.

[343]
T. Furuhashi, K. Nakaoka, and Y. Uchikawa. An efficient finding of fuzzy rules using a new approach to genetic based machine learning. In Proceedings Fourth IEEE International Conference on Fuzzy Systems, pages 715-722. IEEE Computer Press, 1995.

[344]
Takeshi Furuhashi, Ken Nakaoka, and Yoshiki Uchikawa. A Study on Fuzzy Classifier System for Finding Control Knowledge of Multi-Input Systems. In Herrera and Verdegay [423], pages 489-502.
This paper details our attempt to find control knowledge of multi-input systems using a Fuzzy Classifier System (FCS). Simulations are done to show that the FCS can find fuzzy rules for collision avoidance in steering a ship. This paper presents new payoffs and credits for building antecedent parts of fuzzy rules which have truth values larger than zero and for finding fuzzy control rules which achieve the collision avoidance steering. The results show that the FCS can discover fuzzy rules for the multi-input system.

[345]
Takeshi Furuhashi. A Proposal of Hierarchical Fuzzy Classifier Systems. In Forrest [335].

[346]
Michelle Galea and Qiang Shen. Evolutionary approaches to fuzzy rule induction. In Jonathan M. Rossiter and Trevor P. Martin, editors, Proceedings of the 2003 UK Workshop on Computational Intelligence (UKCI-03), pages 205-216, 2003.

[347]
Yang Gao, Joshua Zhexue Huang, Hongqiang Rong, and Da qian Gu. LCSE: Learning Classifier System Ensemble for Incremental Medical Instances. In Tim Kovacs, Xavier LLòra, Keiki Takadama, Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson, editors, Learning Classifier Systems. International Workshops, IWLCS 2003-2005, Revised Selected Papers, volume 4399 of LNCS, pages 93-103. Springer, 2007.

[348]
Santiago Garcia, Fermin Gonzalez, and Luciano Sanchez. Evolving Fuzzy Rule Based Classifiers with GA­P: A Grammatical Approach. In Riccardo Poli, Peter Nordin, William B. Langdon, and Terence C. Fogarty, editors, Genetic Programming, Proceedings of EuroGP'99, volume 1598 of LNCS, pages 203-210, Goteborg, Sweden, May 1999. Springer-Verlag.

[349]
Chris Gathercole. A Classifier System Plays a Simple Board Game. Master's thesis, Department of AI, University of Edinburgh, U.K., 1993.
A description of the problems, successes and failures encountered whilst attempting to encourage a Classifier System to learn to play a simple board game well. Classifier Systems are a kind of free-for-all Production Rule System where the pattern-matching rules compete on the basis of their (modifiable) strength values, and the population of rules is altered by a Genetic Algorithm. They have shown promise in problems where there is very little specific, (i.e. useful) information available from the environment, and the internal adjustments proceed without explicit direction from the environment (or the programmer). In this thesis an attempt is made to `coerce' a variant of Goldberg's Simple Classifier System to `learn' how to play a simple board game (called Dodgems). Various options were tried , among them were: different internal representations, adding more powerful move operators, forcing every move to be valid, and others... The results, whilst not startling, do indicate increased performance with the use of the enhanced move operators over the initial representations. Larger population sizes appear to be beneficial. Also, there is a discussion of the problems involved in choosing the relevant data to study the internal workings of the Classifier System.

[350]
Pierre Gerard and Olivier Sigaud. Combining Anticipation and Dynamic Programming in Classifier Systems. In Proceedings of the International Workshop on Learning Classifier Systems (IWLCS-2000), in the Joint Workshops of SAB 2000 and PPSN 2000 [487]. Extended abstract.
Paper is an extended abstract

[351]
Pierre Gerard and Olivier Sigaud. Adding a generalization mechanism to YACS. In Lee Spector, Erik D. Goodman, Annie Wu, W.B. Langdon, Hans-Michael Voigt, Mitsuo Gen, Sandip Sen, Marco Dorigo, Shahram Pezeshk, Max H. Garzon, and Edmund Burke, editors, Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2001), pages 951-957, San Francisco, California, USA, 7-11 July 2001. Morgan Kaufmann.

[352]
Pierre Gerard and Olivier Sigaud. YACS: Combining dynamic programming with generalization in classifier systems. In Advances in Classifier Systems, volume 1996 of LNAI, pages 52-69. Springer-Verlag, 2001.

[353]
Pierre Gérard and Olivier Sigaud. Designing efficient exploration with MACS: Modules and function approximation. In E. Cantú-Paz, J. A. Foster, K. Deb, D. Davis, R. Roy, U.-M. O'Reilly, H.-G. Beyer, R. Standish, G. Kendall, S. Wilson, M. Harman, J. Wegener, D. Dasgupta, M. A. Potter, A. C. Schultz, K. Dowsland, N. Jonoska, and J. Miller, editors, Genetic and Evolutionary Computation -- GECCO-2003, volume 2724 of LNCS, pages 1882-1893. Springer-Verlag, 2003.
MACS (Modular Anticipatory Classifier System) is a new Anticipatory Classifier System. With respect to its predecessors, ACS, ACS2 and YACS, the latent learning process in MACS is able to take advantage of new regularities. Instead of anticipating all attributes of the perceived situations in the same classifier, MACS only anticipates one attribute per classifier. In this paper we describe how the model of the environment represented by the classifiers can be used to perform active exploration and how this exploration policy is aggregated with the exploitation policy. The architecture is validated experimentally. Then we draw more general principles from the architectural choices giving rise to MACS. We show that building a model of the environment can be seen as a function approximation problem which can be solved with Anticipatory Classifier Systems such as MACS, but also with accuracy-based systems like XCS or XCSF, organized into a Dyna architecture.

[354]
Pierre Gerard, Wolfgang Stolzmann, and Olivier Sigaud. YACS, a new learning classifier system using anticipation. Journal of Soft Computing, 6(3-4):216-228, 2002.
A new and original trend in the learning classifier system (LCS) framework is focussed on latent learning. These new LCSs call upon classifiers with a (condition), an (action) and an (effect) part. In psychology, latent learning is defined as learning without getting any kind of reward. In the LCS framework, this process is in charge of discovering classifiers which are able to anticipate accurately the consequences of actions under some conditions. Accordingly, the latent learning process builds a model of the dynamics of the environment. This model can be used to improve the policy learning process. This paper describes YACS, a new LCS performing latent learning, and compares it with ACS.

[355]
Andreas Geyer-Schulz. Fuzzy Classifier Systems. In Robert Lowen and Marc Roubens, editors, Fuzzy Logic: State of the Art, Series D: System Theory, Knowledge Engineering and Problem Solving, pages 345-354, Dordrecht, 1993. Kluwer Academic Publishers.

[356]
Andreas Geyer-Schulz. Fuzzy Rule-Based Expert Systems and Genetic Machine Learning. Physica Verlag, 1995.

[357]
Andreas Geyer-Schulz. Holland Classifier Systems. In Proceedings of the International Conference on APL (APL'95), volume 25, pages 43-55, New York, NY, USA, June 1995. ACM Press.
A Holland classifier system is an adaptive, general purpose machine learning system which is designed to operate in noisy environments with infrequent and often incomplete feedback. Examples of such environments are financial markets, stock management systems, or chemical processes. In financial markets, a Holland classifier system would develop trading strategies, in a stock management system order heuristics, and in a chemical plant it would perform process control. In this paper we describe a Holland classifier system and present the implementation of its components, namely the production system, the bucket brigade algorithm, the genetic algorithm, and the cover detector, cover effector and triggered chaining operator. Finally, we illustrate the working of a Holland classifier system by learning to find a path with a high payoff in a simple finite state world.

[358]
Andreas Geyer-Schulz. Fuzzy Rule-Based Expert Systems and Genetic Machine Learning. Physica Verlag, 1997.
This book integrates fuzzy rule-languages with genetic algorithms, genetic programming, and classifier systems with the goal of obtaining fuzzy rule-based expert systems with learning capabilities. The main topics are first introduced by solving small problems, then a prototype implementation of the algorithm is explained, and last but not least the theoretical foundations are given. The second edition takes into account the rapid progress in the application of fuzzy genetic algorithms with a survey of recent developments in the field. The chapter on genetic programming has been revised. An exact uniform initialization algorithm replaces the heuristic presented in the first edition. A new method of abstraction, compound derivations, is introduced.

[359]
Antonella Giani, Fabrizio Baiardi, and Antonina Starita. Q-Learning in Evolutionary Rule-Based Systems. In Davidor and Schwefel [228], pages 270-289.
PANIC (Parallelism And Neural networks in Classifier systems), an Evolutionary Rule Based System (ERBS) to evolve behavioral strategies codified by sets of rules, is presented. PANIC assigns credit to the rules through a new mechanism, Q-Credit Assignment (QCA), based on Q-learning. By taking into account the context where a rule is applied, QCA is more accurate than classical methods when a single rule can fire in different situations. QCA is implemented through a multi-layer feed-forward neural network.

[360]
Antonella Giani, Fabrizio Baiardi, and Antonina Starita. PANIC: A parallel evolutionary rule based system. In John R. McDonnell, Robert G. Reynolds, and David B. Fogel, editors, Evolutionary Programming IV. Proceedings of the Fourth Annual Conference on Evolutionary Programming, pages 753-771, 1995.
PANIC (Parallelism And Neural networks In Classifier Systems) is a parallel system to evolve behavioral strategies codified by sets of rules. It integrates several adaptive techniques and computational paradigms, such as genetic algorithms, neural networks, temporal difference methods and classifier systems, to define a powerful and robust learning system. To allocate credit to rules, we propose a new mechanism, Q-Credit Assignment (QCA), based on the temporal difference method Q-learning. To overcome the sharing rule problem, posed by traditional credit assignment strategies in rule based systems, QCA evaluates a rule depending on the context where it is applied. The mechanism is implemented through a multi-layer, feed-forward neural network. To overcome the heavy computational load of this approach, a decentralized and asynchronous parallel model of the genetic algorithm for massive parallel architecture has been devised.

[361]
Antonella Giani, A. Sticca, F. Baiardi, and A. Starita. Q-learning and Redundancy Reduction in Classifier Systems with Internal State. In Claire Nédellec and Céline Rouveirol, editors, Proceedings of the 10th European Conference on Machine Learning (ECML-98), volume 1398 of LNAI, pages 364-369. Springer, 1998.

[362]
Antonella Giani. A Study of Parallel Cooperative Classifier Systems. In John R. Koza, editor, Late Breaking Papers at the Genetic Programming 1998 Conference, University of Wisconsin, Madison, Wisconsin, USA, July 1998. Stanford University Bookstore.

[363]
A. H. Gilbert, Frances Bell, and Christine L. Valenzuela. Adaptive Learning of Process Control and Profit Optimisation using a Classifier System. Evolutionary Computation, 3(2):177-198, 1995.
A Classifier System is used to learn control and profit optimisation of a batch chemical reaction. Ability to learn different market conditions and changes to reaction parameters is demonstrated. The Profit Sharing algorithm is used for Apportionment of Credit. The greater effectiveness of the use of the genetic algorithm over Apportionment of Credit alone or the random replacement of low strength rules is also shown. The Classifier System is unusual in having more than one action per rule.

[364]
Attilio Giordana and G. Lo Bello. Learning classification programs: The genetic algorithm approach. In A. E. Eiben and Z. Michalewicz, editors, Evolutionary Computation, pages 163-177. IOS Press, 1999.
Genetic Algorithms have been proposed by many authors for Machine Learning tasks. In fact, they are appealing for several different reasons, such as the flexibility, the great exploration power, and the possibility of exploiting parallel processing. Nevertheless, it is still controversial whether the genetic approach can really provide effective solutions to learning tasks, in comparison to other algorithms based on classical search strategies. In this paper we try to clarify this point and we overview the work done with respect to the task of learning classification programs from examples. The state of the art emerging from our analysis suggests that the genetic approach can be a valuable alternative to classical approaches, even if further investigation is necessary in order to come to a final conclusion.

[365]
Attilio Giordana and Filippo Neri. Search-Intensive Concept Induction. Evolutionary Computation, 3:375-416, 1995.
This paper describes REGAL, a distributed genetic algorithm-based system, designed for learning First Order Logic concept descriptions from examples. The system is a hybrid between the Pittsburgh and the Michigan approaches, as the population constitutes a redundant set of partial concept descriptions, each evolved separately. In order to increase effectiveness, REGAL is specifically tailored to the concept learning task; hence, REGAL is task-dependent, but, on the other hand, domain-independent. The system proved to be particularly robust with respect to parameter setting across a variety of different application domains. REGAL is based on a selection operator, called Universal Suffrage operator, provably allowing the population to asymptotically converge, in average, to an equilibrium state, in which several species coexist. The system is presented both in a serial and in a parallel version, and a new distributed computational model is proposed and discussed. The system has been tested on a simple artificial domain, for the sake of illustration, and on several complex real-world and artificial domains, in order to show its power, and to analyze its behavior under various conditions. The results obtained so far suggest that genetic search may be a valuable alternative to logic-based approaches to learning concepts, when no (or little) a priori knowledge is available and a very large hypothesis space has to be explored.

[366]
Attilio Giordana and L. Saitta. REGAL: An Integrated System for Learning Relations Using Genetic Algorithms. In Proc. 2nd International Workshop on Multistrategy Learning, pages 234-249, 1993.

[367]
Attilio Giordana and L. Saitta. Learning disjunctive concepts by means of genetic algorithms. In Proc. Int. Conf. on Machine Learning, pages 96-104, 1994.

[368]
R. Giraldez, J. Aguilar-Ruiz, and J. Riquelme. Natural coding: A more efficient representation for evolutionary learning. In Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2003), pages 979-990. Springer-Verlag, 2003.

[369]
R. Giraldez, J. Aguilar-Ruiz, and J. Riquelme. Knowledge-based fast evaluation for evolutionary learning. IEEE Transactions on Systems, Man and Cybernetics, Part C: Applications and Reviews, 35(2):254-261, 2005.
The increasing amount of information available is encouraging the search for efficient techniques to improve the data mining methods, especially those which consume great computational resources, such as evolutionary computation. Efficacy and efficiency are two critical aspects for knowledge-based techniques. The incorporation of knowledge into evolutionary algorithms (EAs) should provide either better solutions (efficacy) or the equivalent solutions in shorter time (efficiency), regarding the same evolutionary algorithm without incorporating such knowledge. In this paper, we categorize and summarize some of the incorporation of knowledge techniques for evolutionary algorithms and present a novel data structure, called efficient evaluation structure (EES), which helps the evolutionary algorithm to provide decision rules using less computational resources. The EES-based EA is tested and compared to another EA system and the experimental results show the quality of our approach, reducing the computational cost about 50%, maintaining the global accuracy of the final set of decision rules.

[370]
David E. Goldberg, Jeffrey Horn, and Kalyanmoy Deb. What Makes a Problem Hard for a Classifier System?. In Collected Abstracts for the First International Workshop on Learning Classifier System (IWLCS-92) [486]. (Also technical report 92007 Illinois Genetic Algorithms Laboratory, University of Illinois at Urbana-Champaign). Available from ENCORE (ftp://ftp.krl.caltech.edu/pub/EC/Welcome.html) in the section on Classifier Systems.
Paper is an extended abstract

[371]
David E. Goldberg. Computer-Aided Gas Pipeline Operation using Genetic Algorithms and Rule Learning. PhD thesis, The University of Michigan, 1983.

[372]
David E. Goldberg. Dynamic System Control using Rule Learning and Genetic Algorithms. In Proceedings of the 9th International Joint Conference on Artificial Intelligence (IJCAI-85), pages 588-592. Morgan Kaufmann, 1985.

[373]
David E. Goldberg. Genetic algorithms and rules learning in dynamic system control. In Grefenstette [389], pages 8-15.

[374]
David E. Goldberg. Genetic Algorithms in Search, Optimization, and Machine Learning. Addison-Wesley, Reading, Mass., 1989.

[375]
David E. Goldberg. Probability Matching, the Magnitude of Reinforcement, and Classifier System Bidding. Machine Learning, 5:407-425, 1990. (Also TCGA tech report 88002, U. of Alabama).
This paper juxtaposes the probability matching paradox of decision theory and the magnitude of reinforcement problem of animal learning theory to show that simple classifier system bidding structures are unable to match the range of behaviors required in the deterministic and probabilistic problems faced by real cognitive systems. The inclusion of a variance-sensitive bidding (VSB) mechanism is suggested, analyzed, and simulated to enable good bidding performance over a wide range of nonstationary probabilistic and deterministic environments.

[376]
David E. Goldberg. Some Reflections on Learning Classifier Systems. Technical Report 2000009, Illinois Genetic Algorithms Laboratory, University of Illinois at Urbana-Champaign, 2000. This appeared as part of Holland2000a.
This work has no abstract

[377]
S. Y. Goldsmith. Steady state analysis of a simple classifier system. PhD thesis, University of New Mexico, Albuquerque, USA, 1989.

[378]
M. Goodloe and S. J. Graves. Improving performance of an electric power expert system with genetic algorithms. In Proceedings of the 1st International Conference on the Applications of Artificial Intelligence and Expert Systems (IEA/AIE-88), pages 298-305. ACM Press, 1988.
Nickel cadmium batteries are an important source of power for aerospace applications. One such application is being developed at the Marshall Space Flight Center (MSFC) for use with the Hubble Space Telescope. A battery testbed has been built at MSFC to aid in that development. In addition, the Nickel Cadmium Battery Expert System (NICBES) was developed by Martin Marietta Corporation to assist NASA engineers in battery management. This paper describes an extension to NICBES which will make it more effective as a battery management tool. The extension involves the incorporation of classifier system machine learning techniques into a subsystem of NICBES. The principal reason for suggesting this extension is the nature of battery management itself. There is still much which is unknown about these batteries and the factors affecting their performance [2]. Hence, battery management might be said to be as much an art as a science and relies heavily on the expertise of the battery manager. NICBES is an attempt to incorporate that battery expertise into an expert system. One difficulty, however, is that battery behavior is likely to change over time in unforseen ways. This detracts from the usefulness of the expert system. Consequently, the battery manager who is using NICBES as a tool would be required to make changes to the expert system in order to accomodate the changed parameters of battery behavior. This should be the function of the knowledge engineer, however, not the battery expert. This is an example of the familiar problem of knowledge acquisition in knowledge engineering. The solution presented here is to use machine learning techniques to help overcome the knowledge acquisition problem. The expert system then interacts at a high level with the battery manager and undertakes adaptation on itself in order to determine new rules conforming to the changed parameters of the power system. The basic principles of learning classifier systems based on genetic algorithms will be presented first. Next, a brief description of NICBES will be given, particularly the advice subsystem to which the learning component will be added. A discussion of specific techniques by which machine learning can be incorporated into this particular rule-based expert system will follow. This discussion will come under the headings of the bit-string representation of rules, the initial rule population, an evaluation function for this system, and the genetic operators. Finally, some comments will be made concerning the implementation of a user interface for a system such as this.

[379]
E. G. Goodman, V. L. Uskov, and W. F. Punch, editors. Proceedings of the First International Conference on Evolutionary Algorithms and their Application EVCA'96, Moscow, 1996. The Presidium of the Russian Academy of Sciences.

[380]
D.P. Greene and S.F. Smith. A genetic system for learning models of consumer choice. In Proceedings of the Second International Conference on Genetic Algorithms and their Applications, pages 217-223. Morgan Kaufmann, 1987.

[381]
David Perry Greene and Stephen F. Smith. Competition-based induction of decision models from examples. Machine Learning, 13:229-257, 1993.
Symbolic induction is a promising approach to constructing decision models by extracting regularities from a data set of examples. The predominant type of model is a classification rule (or set of rules) that maps a set of relevant environmental features into specific categories or values. Classifying loan risk based on borrower profiles, consumer choice from purchase data, or supply levels based on operating conditions are all examples of this type of model-building task. Although current inductive approaches, such as ID3 and CN2, perform well on certain problems, their potential is limited by the incremental nature of their search. Genetic algorithms (GA) have shown great promise on complex search domains, and hence suggest a means for overcoming these limitations. However, effective use of genetic search in this context requires a framework that promotes the funamental model-building objectives of predictive accuracy and model simplicity. In this article we describe COGIN, a GA-based inductive system that exploits the conventions of induction from examples to provide this framework. The novelty of COGIN lies in its use of training set coverage to simultaneously promote competition in various classification niches within the model and constrain overall model complexity. Experimental comparisons with NewID and CN2 provide evidence of the effectiveness of the COGIN framework and the viability of the GA approach.

[382]
David Perry Greene and Stephen F. Smith. Using Coverage as a Model Building Constraint in Learning Classifier Systems. Evolutionary Computation, 2(1):67-91, 1994.
Promoting and maintaining diversity is a critical requirement of search in learning classifier systems (LCSs). What is required of the genetic algorithm (GA) in an LCS context is not convergence to a single global maximum, as in the standard optimization framework, but instead the generation of individuals (i.e., rules) that collectively cover the overall problem space. COGIN (COverage-based Genetic INduction) is a system designed to exploit genetic recombination for the purpose of constructing rule-based classification models from examples. The distinguishing characteristic of COGIN is its use of coverage of training set examples as an explicit constraint on the search, which acts to promote appropriate diversity in the population of rules over time. By treating training examples as limited resources, COGIN creates an ecological model that simultaneously accommodates a dynamic range of niches while encouraging superior individuals within a niche, leading to concise and accurate decision models. Previous experimental studies with COGIN have demonstrated its performance advantages over several well-known symbolic induction approaches. In this paper, we examine the effects of two modifications to the original system configuration, each designed to inject additional diversity into the search: increasing the carrying capacity of training set examples (i.e., increasing coverage redundancy) and increasing the level of disruption in the recombination operator used to generate new rules. Experimental results are given that show both types of modifications to yield substantial improvements to previously published results.

[383]
D.P. Greene. Automated knowledge acquisition: Overcoming the expert systems bottleneck. In Proceedings of the Seventh International Conference on Information Systems, pages 107-117. Lawrence Erlbaum, 1987.

[384]
D.P. Greene. Inductive knowledge acquisition using genetic adaptive search. PhD thesis, The Graduate School of Industrial Administration, Carnegie Mellon University, 1992.

[385]
A. Greenyer. The use of a learning classifier system JXCS. In P. van der Putten and M. van Someren, editors, CoIL Challenge 2000: The Insurance Company Case. Leiden Institute of Advanced Computer Science, June 2000. Technical report 2000-09.

[386]
John J. Grefenstette and H. G. Cobb. User's guide for SAMUEL, Version 1.3. Technical Report NRL Memorandum Report 6820, Naval Research Laboratory, 1991.

[387]
John J. Grefenstette and Alan C. Schultz. An evolutionary approach to learning in robots. In Machine Learning Workshop on Robot Learning, New Brunswick, NJ, 1994. http://www.ib3.gmu.edu/gref/.
Evolutionary learning methods have been found to be useful in several areas in the development of intelligent robots. In the approach described here, evolutionary algorithms are used to explore alternative robot behaviors within a simulation model as a way of reducing the overall knowledge engineering effort. This paper presents some initial results of applying the SAMUEL genetic learning system to a collision avoidance and navigation task for mobile robots.

[388]
John J. Grefenstette, C. L. Ramsey, and Alan C. Schultz. Learning Sequential Decision Rules using Simulation Models and Competition. Machine Learning, 5(4):355-381, 1990. http://www.ib3.gmu.edu/gref/publications.html.
The problem of learning decision rules for sequential tasks is addressed, focusing on the problem of learning tactical decision rules from a simple flight simulator. The learning method relies on the notion of competition and employs genetic algorithms to search the space of decision policies. Several experiments are presented that address issues arising from differences between the simulation model on which learning occurs and the target environment on which the decision rules are ultimately tested.

[389]
John J. Grefenstette, editor. Proceedings of the 1st International Conference on Genetic Algorithms and their Applications (ICGA85). Lawrence Erlbaum Associates: Pittsburgh, PA, July 1985.

[390]
John J. Grefenstette. Multilevel Credit Assignment in a Genetic Learning System. In Proceedings of the 2nd International Conference on Genetic Algorithms (ICGA87) [391], pages 202-207.
Genetic algorithms assign credit to building blocks based on the performance of the knowledge structures in which they occur. If the knowledge structures are rules sets, then the bucket brigade algorithm provides a means of performing additional credit assignment at the level of individual rules. This paper explores one possibility for using the fine-grained feedback provided by the bucket brigade in genetic learning systems that manipulate sets of rules.

[391]
John J. Grefenstette, editor. Proceedings of the 2nd International Conference on Genetic Algorithms (ICGA87), Cambridge, MA, July 1987. Lawrence Erlbaum Associates.

[392]
John J. Grefenstette. Credit Assignment in Rule Discovery Systems Based on Genetic Algorithms. Machine Learning, 3:225-245, 1988.
In rule discovery systems, learning often proceeds by first assessing the quality of the system's current rules and then modifying rules based on that assessment. This paper addresses the credit assignment problem that arises when long sequences of rules fire between successive external rewards. The focus is on the kinds of rule assessment schemes which have been proposed for rule discovery systems that use genetic algorithms as the primary rule modification strategy. Two distinct approaches to rule learning with genetic algorithms have been previously reported, each approach offering a useful solution to a different level of the credit assignment problem. We describe a system, called RUDI, that exploits both approaches. We present analytic and experimental results that support the hypothesis that multiple levels of credit assignment can improve the performance of rule learning systems based on genetic algorithms.

[393]
John J. Grefenstette. A System for Learning Control Strategies with Genetic Algorithms. In Schaffer [718], pages 183-190.
A system called SAMUEL is described for learning rules to control a process, given only a weak model of the process consisting of a set of sensors, a set of control variables, and feedback mechanism that provides intermittent performance measurements. This paper focuses on features that distinguish SAMUEL from previous systems for learning rules with genetic algorithms. In particular, a restricted high level rule language is used, and genetic operators suitable for the language are presented. An empirical study shows that SAMUEL can learn rules to control a challenging dynamic process.

[394]
John J. Grefenstette. Lamarckian Learning in Multi-Agent Environments. In Booker and Belew [74], pages 303-310. http://www.ib3.gmu.edu/gref/publications.html.
Genetic algorithms gain much of their power from mechanisms derived from the field of population genetics. However, it is possible, and in some cases desirable, to augment the standard mechanisms with additional features not available in biological systems. In this paper, we examine the use of Lamarckian learning operators in the SAMUEL architecture. The use of the operators is illustrated on three tasks in multi-agent environments.

[395]
John J. Grefenstette. Learning decision strategies with genetic algorithms. In Proc. Intl. Workshop on Analogical and Inductive Inference, volume 642 of Lecture Notes in Artificial Intelligence, pages 35-50. Springer-Verlag, 1992. http://www.ib3.gmu.edu/gref/.
Machine learning offers the possibility of designing intelligent systems that refine and improve their initial knowledge through their own experience. This article focuses on the problem of learning sequential decision rules for multi-agent environments. We describe the SAMUEL learning system that uses genetic algorithms and other competition based techniques to learn decision strategies for autonomous agents. One of the main themes in this research is that the learning system should be able to take advantage of existing knowledge where available. This article describes some of the mechanisms for expressing existing knowledge in SAMUEL, and explores some of the issues in selecting constraints for the learning system.

[396]
John J. Grefenstette. The Evolution of Strategies for Multi-agent Environments. Adaptive Behavior, 1:65-89, 1992. http://www.ib3.gmu.edu/gref/.
SAMUEL is an experimental learning system that uses genetic algorithms and other learning methods to evolve reactive decision rules from simulations of multi-agent environments. The basic approach is to explore a range of behavior within a simulation model, using feedback to adapt its decision strategies over time. One of the main themes in this research is that the learning system should be able to take advantage of existing knowledge where available. This has led to the adoption of rule representations that ease the expression existing knowledge. A second theme is that adaptation can be driven by competition among knowledge structures. Competition is applied at two levels in SAMUEL. Within a strategy composed of decision rules, rules compete with one another to influence the behavior of the system. At a higher level of granularity, entire strategies compete with one another, driven by a genetic algorithm. This article focuses on recent elaborations of the agent model of SAMUEL that are specifically designed to respond to multiple external agents. Experimental results are presented that illustrate the behavior of SAMUEL on two multi-agent predator-prey tasks.

[397]
John J. Grefenstette. Using a genetic algorithm to learn behaviors for autonomous vehicles. In Proceedings American Institute of Aeronautics and Astronautics Guidance, Navigation and Control Conference, pages 739-749. AIAA, 1992. http://www.ib3.gmu.edu/gref/.
Truly autonomous vehicles will require both projective planning and reactive components in order to perform robustly. Projective components are needed for long-term planning and replanning where explicit reasoning about future states is required. Reactive components allow the system to always have some action available in real-time, and themselves can exhibit robust behavior, but lack the ability to explicitly reason about future states over a long time period. This work addresses the problem of creating reactive components for autonomous vehicles. Creating reactive behaviors (stimulus-response rules) is generally difficult, requiring the acquisition of much knowledge from domain experts, a problem referred to as the knowledge acquisition bottleneck. SAMUEL is a system that learns reactive behaviors for autonomous agents. SAMUEL learns these behaviors under simulation, automating the process of creating stimulus-response rules and therefore reducing the bottleneck. The learning algorithm was designed to learn useful behaviors from simulations of limited fidelity. Current work is investigating how well behaviors learned under simulation environments work in real world environments. In this paper, we describe SAMUEL, and describe behaviors that have been learned for simulated autonomous aircraft, autonomous underwater vehicles, and robots. These behaviors include dog fighting, missile evasion, tracking, navigation, and obstacle avoidance.

[398]
John J. Grefenstette. Evolutionary Algorithms in Robotics. In M. Jamshedi and C. Nguyen, editors, Robotics and Manufacturing: Recent Trends in Research, Education and Applications, v5. Proc. Fifth Intl. Symposium on Robotics and Manufacturing, ISRAM 94, pages 65-72. ASME Press: New York, 1994. http://www.ib3.gmu.edu/gref/.
Evolutionary algorithms incorporate principles from biological population genetics to perform search, optimization, and learning. This article discusses issues arising in the application of evolutionary algorithms to problems in robotics.

[399]
T. Nakashima H. Ishibuchi and T. Kuroda. A fuzzy genetics-based machine learning method for designing linguistic classification systems with high comprehensibility. In Proceedings 6th Int. Conf. on Neural Information Processing, volume 2, pages 597-602, 1999.

[400]
T. Nakashima H. Ishibuchi and T. Kuroda. A hybrid fuzzy gbml algorithm for designing compact fuzzy rule-based classification systems. In Proc. 9th IEEE Int. Conf. on Fuzzy Systems (FUZZ IEEE 2000), volume 2, pages 706-711, 2000.

[401]
T. Nakashima H. Ishibuchi and T. Murata. Genetic-algorithm-based approaches to the design of fuzzy systems for multi-dimensional pattern classification problems. In Proc. 1996 IEEE Int. conf on Evolutionary Computation, pages 229-234, 1996.

[402]
T. Nakashima H. Ishibuchi and T. Murata. Performance evaluation of fuzzy classifier systems for multidimensional pattern classification problems. IEEE Transactions on Systems, Man and Cybernetics, Part B, 29(5):601-618, 1999.

[403]
Chris Lokan Hai H. Dam, Hussein A. Abbass. Bcs: Bayesain learning classifier system. Technical Report TR-ALAR-200604005, Artificial Life and Adaptive Robotics Laboratory, UNSW, 2006.

[404]
Ali Hamzeh and Adel Rahmani. A fuzzy system to control exploration rate in xcs. In Tim Kovacs, Xavier LLòra, Keiki Takadama, Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson, editors, Learning Classifier Systems. International Workshops, IWLCS 2003-2005, Revised Selected Papers, volume 4399 of LNCS, pages 115-127. Springer, 2007.

[405]
Hisashi Handa, Takashi Noda, Tadataka Konishi, Osamu Katai, and Mitsuru Baba. Coevolutionary fuzzy classifier system for autonomous mobile robots. In Takadama [815].

[406]
Greg Harrison. Genetically programmed learning classifier system for complex adaptive system processing with agent-based architecture. US Patent 6,741,974, 2004.

[407]
Adrian Hartley. Genetics Based Machine Learning as a Model of Perceptual Category Learning in Humans. Master's thesis, University of Birmingham, 1998. ftp://ftp.cs.bham.ac.uk/pub/authors/T.Kovacs/index.html.
The ability to recognize quickly and accurately what they encounter is a fundamental ability of normal intelligent human behaviour. How people perform the task of learning the categories that objects in the world fit into is still an unanswered question. However, in this thesis I follow up an idea that Holland et al. (1986) proposed and present genetic based machine learning as a model of perceptual category learning. Genetics based machine learning has a certain amount of popularity within computer science yet the ideas have been slow to cross the boundaries of disciplines. The main drive of this research is to adopt the theoretical position of Holland et al. (1986) and see if realistically genetic based machine learning can be used as a model of learning in humans. Within Psychology the domain of category learning has grown as an area of interest, for categorization is considered basic to all our intellectual abilities (Estes 1994). Categorization ``is the process of assigning objects (of whatever kind) to categories (which are collections of objects which are grouped together for some purpose)'' (Lamberts 1997). There is a bench mark set of results within perceptual category learning from Shepard et al (1961). These tasks have a well defined difficulty ordering and serve as a starting point for any model of perceptual category learning. In psychology neural networks are often used for cognitive modelling. However Sen (1996) used a simple classifier system (Newboole) to model the classic Shepard et al. (1961) tasks. I will begin with a replication of this study and then move on to use another more advanced classifier system (XCS: Wilson 1995) to see if this provides a better model of perceptual category learning. Also, I introduce a category switch task that has not previously been evaluated with these systems. Results show that although simple classifier systems can capture qualitatively results from humans, they fail to show elegant solutions to problems and are limited in the tasks they can model. XCS can solve a greater variety of problems as it can handle multi step problems as well as single step, also it models both the Shepard and the switch problems. XCS develops a covering map of knowledge giving the system much more complete knowledge of the classification problem compared to simpler systems. XCS learns not only what is a correct classification but also what is an incorrect classification. Also due to the derivation of a covering map of knowledge XCS finds simple elegant solutions to problems. Some classifier systems (XCS) may offer a realistic alternative to neural networks in cognitive modelling.

[408]
Adrian Hartley. Accuracy-based fitness allows similar performance to humans in static and dynamic classification environments. In Banzhaf et al. [32], pages 266-273.
Traditionally within classifier systems the ability of a classifier to obtain reward (as measured by its strength) indicates the fitness of the classifier within the rule population. However, Wilson (1995) proposed a new approach to fitness in terms of a classifier's prediction accuracy. This paper presents experiments with two different classifier systems: Newboole (Bonelli et al. 1990) and XCS (Wilson 1995). Both systems demonstrate qualitative matches to data from perceptual category learning in humans. However, the different methods of fitness evaluation of classifiers alter the knowledge the systems learn and maintain. When fitness is based upon strength (Newboole) the system acquires knowledge to solve the classification problem. But when fitness is based on accuracy (XCS) the system acquires a more complete knowledge of the problem space. Further experiments show that the optimal covering map (Kovacs 1997) of knowledge that emerges in XCS allows the system to compensate rapidly in a dynamic classification environment. This is also more similar to human performance on comparable tasks.

[409]
U. Hartmann. Efficient Parallel Learning in Classifier Systems. In Albrecht et al. [7], pages 515-521.
Classifier systems are simple production systems working on binary messages of fixed length. Genetic algorithms are employed in classifier systems in order to discover new classifiers. We use methods of the computational complexity theory in order to analyse the inherent difficulty of learning in classifier systems. Hence our results do not depend on special (possibly genetic) learning algorithms. The paper formalises this rule discovery or learning problem for classifier systems which has been proved to be hard in general. It will be proved that restrictions on two distinct learning problems lead to problems in NC, i.e. problems which are efficiently solvable in parallel.

[410]
U. Hartmann. On the Complexity of Learning in Classifier Systems. In Davidor and Schwefel [228], pages 280-289. Republished in: ECAI 94. 11th European Conference on Artificial Intelligence. A Cohn (Ed.), pp.438-442, 1994. John Wiley and Sons.
Genetic algorithms are employed in classifier systems in order to discover new classifiers. The paper formalises this rule discovery or learning problem for classifier systems and uses methods of computational complexity theory to analyse its inherent difficulty. It is proved that two distinct learning problems are NP-complete, i.e. not likely to be solvable efficiently. The practical relevance of these theoretical results is briefly discussed.

[411]
Marianne Haslev. A Classifier System for the Production by Computer of Past Tense Verb-Forms. Presented at a Genetic Algorithms Workshop at the Rowland Institute, Cambridge MA, Nov 1986, 1986.

[412]
Mozart Hasse and Aurora R. Pozo. Using Phenotypic Sharing in a Classifier Tool. In Whitely et al. [878], page 392. One page poster paper.
This paper describes a classifier tool that uses a genetic algorithm to make rule induction. The genetic algorithm uses the Michigan approach, is domain independent and is able to process continuous and discrete attributes. Some optimizations include the use of phenotypic sharing (with linear complexity) to direct the search. The results of accuracy are compared with other 33 algorithms in 32 datasets. The difference of accuracy is not statistically significant at the 10% level when compared with the best of the other 33 algorithms. The implementation allows the configuration of many parameters, and intends to be improved with the inclusion of new operators.

[413]
Akira Hayashi and Nobuo Suematsu. Viewing Classifier Systems as Model Free Learning in POMDPs. In Advances in Neural Information Processing Systems (NIPS) 11, pages 989-995, 1999.
Classifier systems are now viewed disappointing because of their problems such as the rule strength vs rule set performance problem and the credit assignment problem. In order to solve the problems, we have developed a hybrid classifier system: GLS (Generalization Learning System). In designing GLS, we view CSs as model free learning in POMDPs and take a hybrid approach to finding the best generalization, given the total number of rules. GLS uses the policy improvement procedure by Jaakkola et al. for an locally optimal stochastic policy when a set of rule conditions is given. GLS uses GA to search for the best set of rule conditions.

[414]
Jörg Heitkötter and David Beasley. The Hitch-Hiker's Guide to Evolutionary Computation (FAQ for comp.ai.genetic). Accessed 28/2/09. http://www.aip.de/~ast/EvolCompFAQ/, 2001.

[415]
J. Hekanaho. Symbiosis in multimodal concept learning. In Proc. 1995 Int. Conf. on Machine Learning (ML'95), pages 278-285, 1995.

[416]
Luis Miramontes Hercog and Terence C. Fogarty. XCS-based inductive intelligent multi-agent system. In Late Breaking Papers at the 2000 Genetic and Evolutionary Computation Conference (GECCO-2000), pages 125-132, 2000.
Induction is tested in a population of XCS-based agents, tested in the frame of the ``El Farol'' bar problem. Two reward schemes are used, selfish and co-operative, being the latter the best for the purposes of the experiment.

[417]
Luis Miramontes Hercog and Terence C. Fogarty. XCS-based Inductive Multi-Agent System. In Proceedings of the International Workshop on Learning Classifier Systems (IWLCS-2000), in the Joint Workshops of SAB 2000 and PPSN 2000 [487]. Extended abstract.
Paper is an extended abstract

[418]
Luis Miramontes Hercog and Terence C. Fogarty. Analysis of inductive intelligence in xcs-based multi-agent system (maxcs). In J.Periaux, P. Joly, and E. Onate, editors, Innovative Tools for Scientific Computation in Aeronautical Engineering, pages 351-366. CIMNE, Barcelona, 2001. ISBN: 84-90025-78-X.

[419]
Luis Miramontes Hercog and Terence C. Fogarty. Co-evolutionary classifier systems for multi-agent simulation. In David B. Fogel, Mohamed A. El-Sharkawi, Xin Yao, Garry Greenwood, Hitoshi Iba, Paul Marrow, and Mark Shackleton, editors, Proceedings of the 2002 Congress on Evolutionary Computation CEC2002, pages 1798-1803. IEEE Press, 2002.

[420]
Luis Miramontes Hercog and Terence C. Fogarty. Social simulation using a Multi-Agent Model based on Classifier Systems: The Emergence of Vacillating Behaviour in the ``El Farol'' Bar Problem. In Lanzi et al. [546], pages 88-111.

[421]
Luis Miramontes Hercog. Hand-eye coordination: An evolutionary approach. Master's thesis, Department of Artificial Intelligence. University of Edinburgh, 1998.
The coordination between the sensor and motor systems is an essential feature in autonomous intelligent systems. This thesis investigates the evolutionary approach to sensorimotor control using learning classifier systems. A simple classifier system is used to solve the problem of coordinating a pair of eyes and an arm in order to catch an object. First, an analysis of the previous approaches based on neural networks is done. Then a review of what a simple classifier system is, as well as the specific implementation of the simple classifier system to solve this problem. Afterwards an analysis of the results is presented. Finally a review of the advantages and disadvantages of this approach comparing it with previous ones is offered. Results have shown that classifier systems are a promising tool, which solve this sensorimotor coordination problem, further work needs to be done to determine the limitations of this approach.

[422]
Luis Miramontes Hercog. Evolutionary algorithms and collective learning for traffic demand policy management. Technical report, South Bank University, 2003.

[423]
F. Herrera and J. L. Verdegay, editors. Genetic Algorithms and Soft Computing, (Studies in Fuzziness, 8). Physica-Verlag, Berlin, 1996.

[424]
E. Herrera-Viedma. Sistemas Clasificadores de Aprendizaje. Aproximaciones Difusas. Technical Report DECSAI-95132, Dept. of Computer Science and A.I., University of Granada, 1995.
La Inteligencia Artificial tiene planteado el reto de lograr crear sistemas computacionales que desarrollen una conducta similar a la conducta humana. Esto no es tarea facil y es el quebradero de cabeza de muchos investigadores. En el presente trabajo hemos intentado estudiar algunas de las aportaciones que se estan haciendo desde el campo de los Algoritmos Geneticos y su integracion con tecnicas basadas en Conjuntos Fuzzy. Concretamente, analizaremos uno de los paradigmas de aprendizaje de los AGs, Los Sistemas Clasificadores, y algunas de sus aproximaciones fuzzy, varios modelos de Sistemas Clasificadores Fuzzy. Es importante destacar que estos sistemas son de especial interes por la implicaciones que pueden tener en el desarrollo de sistema de control de aprendizaje autonomos y adaptativos que pueden actuar de forma ``inteligente'', de modo similar a los seres humanos.

[425]
Tetsuya Higuchi and Bernard Manderick. Hardware realizations of evolutionary algorithms. In Thomas Bäck, David B. Fogel, and Zbigniew Michalewicz, editors, Handbook of Evolutionary Computation, pages E2.3:1-G2.3:8. IOP Publishing Ltd and Oxford University Press, 1997.

[426]
M. R. Hilliard, G. E. Liepins, Mark Palmer, Michael Morrow, and Jon Richardson. A classifier based system for discovering scheduling heuristics. In Grefenstette [391], pages 231-235.

[427]
M. R. Hilliard, G. E. Liepins, and M. Palmer. Machine learning applications to job shop scheduling. In Proc. AAAI-SIGMAN Workshop on Production Planning and Scheduling, 1988.

[428]
John H. Holland and Arthur W. Burks. Adaptive Computing System Capable of Learning and Discovery. Patent 4697242 United States 29 Sept., 1987.

[429]
John H. Holland and J. S. Reitman. Cognitive systems based on adaptive algorithms. In D. A. Waterman and F. Hayes-Roth, editors, Pattern-directed Inference Systems. New York: Academic Press, 1978. Reprinted in: Evolutionary Computation. The Fossil Record. David B. Fogel (Ed.) IEEE Press, 1998. ISBN: 0-7803-3481-7.
The type of cognitive system (CS) studied here has four basic parts: (1) a set of interacting elementary productions, called classifiers, (2) a performance algorithm that directs the action of the system in the environment, (3) a simple learning algorithm that keeps a record of each classifier's success in bringing about rewards, and (4) a more complex learning algorithm, called the genetic algorithm, that modifies the set of classifiers so that variants of good classifiers persist and new, potentially better ones are created in a provably efficient manner. Two ``proof-of-principle'' experiments are reported. One experiment shows CS's performance in a maze when it has only the ability to adjust the predictions about ensuing rewards of classifiers (similar to adjusting the ``weight'' of a classifier) vs. when the power of the genetic algorithm is added. Criterion was achieved an order of magnitude more rapidly when the genetic algorithm was operative. A second experiment examines transfer of learning. Placed in a more difficult maze, CS with experience in the simpler maze reaches criterion an order of magnitude more rapidly than CS without prior experience.

[430]
John H. Holland, Keith J. Holyoak, Richard E. Nisbett, and P. R. Thagard. Induction: Processes of Inference, Learning, and Discovery. MIT Press, Cambridge, 1986.

[431]
John H. Holland, Keith J. Holyoak, Richard E. Nisbett, and Paul R. Thagard. Classifier Systems, Q-Morphisms, and Induction. In Davis [234], pages 116-128.

[432]
John H. Holland, Lashon B. Booker, Marco Colombetti, Marco Dorigo, David E. Goldberg, Stephanie Forrest, Rick L. Riolo, Robert E. Smith, Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson. What is a Learning Classifier System? In Lanzi et al. [544], pages 3-32.
We asked ``What is a Learning Classifier System'' to some of the best-known researchers in the field. These are their answers.

[433]
John H. Holland. Processing and processors for schemata. In E. L. Jacks, editor, Associative Information Processing, pages 127-146. New York: American Elsevier, 1971.

[434]
John H. Holland. Adaptation in Natural and Artificial Systems. University of Michigan Press, Ann Arbor, 1975. Republished by the MIT press, 1992.

[435]
John H. Holland. Adaptation. In R. Rosen and F. M. Snell, editors, Progress in Theoretical Biology. New York: Plenum, 1976.

[436]
John H. Holland. Adaptive algorithms for discovering and using general patterns in growing knowledge bases. International Journal of Policy Analysis and Information Systems, 4(3):245-268, 1980.

[437]
John H. Holland. Genetic Algorithms and Adaptation. Technical Report 34, University of Michigan. Department of Computer and Communication Sciences, Ann Arbor, 1981.

[438]
John H. Holland. Escaping brittleness. In Proceedings Second International Workshop on Machine Learning, pages 92-95, 1983.
Expert systems are powerful when working within the domain-specific boundaries of the initial design. It is widely agreed that the usefulness of such systems would be greatly enhanced if they could be made more versatile or less brittle - tolerant of changes in the domain and underlying model. This paper suggests 7 criteria for escaping brittleness (combination, parallelism, declarative and procedural information, categorization, synchronic and diachronic pointing, gracefulness and confirmation) and gives an example of a class of general purpose systems, classifier systems, that attempt to meet these criteria.

[439]
John H. Holland. Properties of the bucket brigade. In Grefenstette [389], pages 1-7.
This paper does not have an absract.

[440]
John H. Holland. A Mathematical Framework for Studying Learning in a Classifier System. In Doyne Farmer, Alan Lapedes, Norman Packard, and Burton Wendroff, editors, Evolution, Games and Learning: Models for Adaptation in Machines and Nature, pages 307-317, Amsterdam, 1986. North-Holland.

[441]
John H. Holland. A Mathematical Framework for Studying Learning in Classifier Systems. Physica D, 22:307-317, 1986.
Massively parallel, rule-based systems offer both a practical and a theoretical tool for understanding systems that act usefully in complex environments [see, for example, refs 1-4]. However, these systems pose a number of problems of a high order of difficulty -- problems that can be broadly characterized as problems in nonlinear dynamics. The difficulties stem from the fact that the systems are designed to act in environments with complex transition functions -- environments that, in all circumstances of interest, are far from equilibrium. Interactions with the environment thus face the systems with perpetual novelty, and the usual simplifications involving fixed points, limit cycles, etc,. just do not apply. Learning procedures (adaptive algorithms) offer a way of combating these difficulties, but an understanding of the possibilities is not a simple matter. The key question is easy enough to state informally: What kinds of environmental regularity can be exploited by learning? However, if answers that are both useful and widely applicable are to be forthcoming, the question must be reformulated in a way that gives it precision without losing generality. The usual tool for this task is a mathematical framework that suitably encompasses the subject. It is the purpose of this paper to explore a framework that gives a precise definition to the notion of an environmental regularity and then treats learning procedures as procedures for revising rules in response to detected environmental regularities. In this context, procedures for revising rules become more than a convenience, they take a central place in the design. Whether carried out by a human or a machine, rule revision requires the solution of two broad problems. First, one must rate rules as to their usefulness to the system as a whole -- the apportionment of credit problem. Then one must devise new rules that serve the system better than the least useful of the rules already in place -- the rule discovery problem. Though these two problems are sometimes treated separately, they are closely interrelated. A machine learning approach is used here to illustrate the interaction of apportionment of credit and rule discovery algorithms, and then the overall system is abstracted and translated to the mathematical framework. To give the framework a concrete subject matter, section 1 introduces a particular class of highly parallel, rule-based systems called classifier systems. The next section....

[442]
John H. Holland. Escaping Brittleness: The Possibilities of General-Purpose Learning Algorithms Applied to Parallel Rule-Based Systems. In Mitchell, Michalski, and Carbonell, editors, Machine Learning, an Artificial Intelligence Approach. Volume II, chapter 20, pages 593-623. Morgan Kaufmann, 1986.
Message-passing, rule-based production systems in which many rules are active simultaneously offer attractive possibilities for the exploitation of general-purpose machine learning algorithms. In such systems each rule can be looked upon as a tentative hypothesis about some aspect of the task environment, competing against other plausible hypotheses being entertained at the same time. In this context there are two major tasks for machine learning algorithms: (1) apportionment of credit and (2) rule discovery. The apportionment-of-credit algorithm(s) must assign ``strength'' to rules on the basis of their observed usefulness to the system. The problem is complicated by the difficulty of determining which of a cluster of rules active in an early, ``stage-setting'' capacity has contributed to a later useful outcome (e.g., rules controlling early moves that make possible later a triple jump in checkers). If strengths can be assigned appropriately, then they can be used to determine a rule's ability to win against competing rules, and they can be used to determine the rule's likelihood of being used as a ``parent'' for new rules. Surprisingly, for credit apportionment algorithms of the bucket-brigade variety, one can prove fixed-point theorems that provide some guarantees of an appropriate apportionment. The task of rule discovery depends critically upon the discovery of good ``building blocks'' for generating plausible rules (hypotheses). A parallel system designed with machine learning in mind must permit a constant flux of new rules to be tested and exploited or discarded. Moreover this flux must not disturb the system's behavior in task environments for which it has well-practiced, effective procedures. Genetic algorithms, using the strengths as ``fitnesses'', offer subtle ways of discovering good building blocks, and there are new versions of theorems from mathematical genetics that enable us to understand this discovery process.

[443]
John H. Holland. Genetic Algorithms and Classifier Systems: Foundations and Future Directions. In Grefenstette [391], pages 82-89.
Theoretical questions about classifier systems, with rare exceptions, apply equally to other adaptive nonlinear networks (ANNs) such as the connectionist models of cognitive psychology, the immune system, economic systems, ecologies and genetic systems. This paper discusses pervasive properties of ANNs and the kinds of mathematics relevant to questions about these properties. It discusses relevant functional extensions of the basic classifier system and extensions of the extant mathematical theory. An appendix briefly reviews some of the key theorems about classifier systems.

[444]
John H. Holland. Concerning the Emergence of Tag-Mediated Lookahead in Classifier Systems. In Special issue of Physica D (Vol. 42) [292], pages 188-201.
This paper, after a general introduction to the area, discusses the architecture and learning algorithms that permit automatic parallel, distributed lookahead to emerge in classifier systems. Simple additions to a ``standard'' classifier system suffice, principally a new register called the virtual strength register, and a provision to use the bucket brigade credit-assignment algorithm in ``virtual'' mode to modify values in this register. With these additions, current actions are decided on the basis of expected values associated with the ``lookahead cones'' of possible alternatives.

[445]
John H. Holmes, Dennis R. Durbin, and Flaura K. Winston. A New Bootstrapping Method to Improve Classification Performance in Learning Classifier Systems. In Proceedings of Parallel Problem Solving from Nature (PPSN VI), 2000.
A new technique for improving the classification performance of learning classifier systems (LCS) was developed and applied to a real-world data mining problem. EpiCS, a stimulus-response LCS, was adapted to perform prevalence-based bootstrapping, wherein data from training and testing sets were sampled according to the prevalence of the individual classes, rather than randomly using the class distribution inherent in the data. Prevalence-based bootstrapping was shown to improve classification performance significantly on training and testing (p<0.0001). Furthermore, this procedure was shown to enhance EpiCS's classification performance on testing compared to C4.5 when similar bootstrapping procedures were applied to the latter.

[446]
John H. Holmes, Dennis R. Durbin, and Flaura K. Winston. The learning classifier system: an evolutionary computation approach to knowledge discovery in epidemiologic surveillance. Artificial Intelligence In Medicine, 19(1):53-74, 2000.
The learning classifier system (LCS) integrates a rule-based system with reinforcement learning and genetic algorithm-based rule discovery. This investigation reports on the design, implementation, and evaluation of EpiCS, a LCS adapted for knowledge discovery in epidemiologic surveillance. Using data from a large, national child automobile passenger protection program, EpiCS was compared with C4.5 and logistic regression to evaluate its ability to induce rules from data that could be used to classify cases and to derive estimates of outcome risk, respectively. The rules induced by EpiCS were less parsimonious than those induced by C4.5, but were potentially more useful to investigators in hypothesis generation. Classification performance of C4.5 was superior to that of EpiCS (P<0.05). However, risk estimates derived by EpiCS were significantly more accurate than those derived by logistic regression (P<0.05).

[447]
John H. Holmes, Jennifer A. Sager, and Warren B. Bilker. Three methods for covering missing input data in xcs. In Learning Classifier Systems, volume 4399 of LNAI, pages 181-192. Springer, 2007.

[448]
John H. Holmes, Jennifer A. Sager, and Warren B. Bilker. Three methods for covering missing input data in xcs. In Tim Kovacs, Xavier LLòra, Keiki Takadama, Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson, editors, Learning Classifier Systems. International Workshops, IWLCS 2003-2005, Revised Selected Papers, volume 4399 of LNCS, pages 181-192. Springer, 2007.
Missing data pose a potential threat to learning and classification in that they may compromise the ability of a system to develop robust, generalized models of the environment in which they operate. This investigation reports on the effects of three approaches to covering these data using an XCS-style learning classifier system. Using fabricated datasets representing a wide range of missing value densities, it was found that missing data do not appear to adversely affect LCS learning and classification performance. Furthermore, three types of missing value covering were found to exhibit similar efficiency on these data, with respect to convergence rate and classification accuracy.

[449]
John H. Holmes. Evolution-Assisted Discovery of Sentinel Features in Epidemiologic Surveillance. PhD thesis, Drexel University, 1996. http://cceb.med.upenn.edu/holmes/disstxt.ps.gz.
The use of a genetics-based classifier system (CS) in generating epidemiologic hypotheses was investigated. In addition, epidemiologic analytical techniques were used to evaluate the performance of a CS in this problem domain. Five component studies were implemented, using epidemiologic surveillance data over a range of prevalences. The evaluation study investigated the use of the area under the receiver operating characteristic curve (*) as an alternative to crude accuracy (CA) during the training period. The classification study examined the ability of the CS to classify unencountered patients. The reproducibility study demonstrated the stochastic processes underlying CS performance during training and testing. The payoff-penalty parameterization study investigated the effects of differential penalty for false negative and false positive decisions on learning rate and classification ability. The risk assessment study examined the ability of a CS to derive estimates of risk for purposes of classification. At 50% prevalence, * was identical to CA over the entire training period; with decreasing prevalence, CA increasingly overestimated the learning rate, while * provided more accurate depictions of this measure. Across all four prevalences investigated, the CS was able to classify unseen patients well, with *s ranging from 0.95 at 50% prevalence to 0.78 at 10%. The classifier populations after training indicated considerable generalization; decision rules were discernible on visual examination. When trained and tested using 1,000 different data sets drawn from the same pool, the CS was fairly consistent in terms of learning rate and classification ability, although with sufficient variation to warrant investigating the use of bootstrapping techniques. Biasing the ratio of false positive to false negative (FP:FN) decisions affected the learning rate relative to prevalence. Learning rate was most enhanced at 25% and 10% prevalence by a FP:FN ratio of 4:1 and 10:1, respectively. Across all four prevalences, the CS was able to produce risk estimates that consistently outperformed decision rules derived using logistic regression. The CS was shown to be a useful adjunct to hypothesis generation during epidemiologic surveillance.

[450]
John H. Holmes. A genetics-based machine learning approach to knowledge discovery in clinical data. Journal of the American Medical Informatics Association Supplement, 1996.

[451]
John H. Holmes. Discovering Risk of Disease with a Learning Classifier System. In Bäck [24]. http://cceb.med.upenn.edu/holmes/icga97.ps.gz.
A learning classifier system, EpiCS, was used to derive a continuous measure of disease risk in a series of 250 individuals. Using the area under the receiver-operating characteristic curve, this measure was compared with the risk estimate derived for the same individuals by logistic regression. Over 20 training-testing trials, risk estimates derived by EpiCS were consistently more accurate (mean area=0.97, SD=0.01) than that derived by logistic regression (mean area=0.89, SD=0.02). The areas for the trials with minimum and maximum classification performance on testing were significantly greater (p=0.019 and p<0.001, respectively) than the area for the logistic regression curve. This investigation demonstrated the ability of a learning classifier system to produce output that is clinically meaningful in diagnostic classification.

[452]
John H. Holmes. Differential negative reinforcement improves classifier system learning rate in two-class problems with unequal base rates. In Koza et al. [529], pages 635-642. http://cceb.med.upenn.edu/holmes/gp98.ps.gz.
The effect of biasing negative reinforcement levels on learning rate and classification accuracy in a learning classifier system (LCS) was investigated. Simulation data at five prevalences (base rates) were used to train and test the LCS. Erroneous decisions made by the LCS during training were punished differentially according to type: false positive (FP) or false negative (FN), across a range of four FP:FN ratios. Training performance was assessed by learning rate, determined from the number of iterations required to reach 95% of the maximum area under the receiver operating characteristic (ROC) curve obtained during learning. Learning rates were compared across the three biased ratios with those obtained at the unbiased ratio. Classification performance of the LCS at testing was evaluated by means of the area under the ROC curve. During learning, differences were found between the biased and unbiased penalty schemes, but only at unequal base rates. A linear relationship between bias level and base rate was suggested. With unequal base rates, biasing the FP:FN ratio improved the learning rate. Little effect was observed on testing the LCS with novel cases.

[453]
John H. Holmes. Evaluating Learning Classifier System Performance In Two-Choice Decision Tasks: An LCS Metric Toolkit. In Banzhaf et al. [32], page 789. One page poster paper.
A ``metric toolkit'' to evaluate learning classifier system performance is proposed. The metrics are shown to be superior to crude accuracy in evaluating classification performance, especially for data with unequal numbers of positive and negative cases. In addition, these metrics provide information to the researcher that is not available from crude accuracy. When used appropriately, these metrics provide accurate depictions of learning classifier system performance during training and testing in supervised learning environments.

[454]
John H. Holmes. Quantitative Methods for Evaluating Learning Classifier System Performance in Forced Two-Choice Decision Tasks. In Wu [923], pages 250-257.
Applying a learning classifier system to two-class decision problems requires a special approach to performance evaluation. This paper presents a suite of quantitative tools that addresses the evaluation requirements of two-class problems. These metrics, borrowed from the domain of medical decision making, are proposed as adjuncts to commonly used evaluation methods such as crude accuracy (``percent correct''). They include sensitivity, specificity, area under the receiver operating characteristic curve, and predictive value. These metrics are shown to be superior to crude accuracy in evaluating learning classifier system performance, especially when applied to data with unequal numbers of positive and negative cases. In addition, these metrics provide information to the researcher that is not available from crude accuracy. When used appropriately, these metrics provide accurate depictions of learning classifier system performance during training and testing in supervised learning environments.

[455]
John H. Holmes. Applying a Learning Classifier System to Mining Explanatory and Predictive Models from a Large Database. In Proceedings of the International Workshop on Learning Classifier Systems (IWLCS-2000), in the Joint Workshops of SAB 2000 and PPSN 2000 [487]. Extended abstract.
Paper is an extended abstract

[456]
John H. Holmes. Learning Classifier Systems Applied to Knowledge Discovery in Clinical Research Databases. In Lanzi et al. [544], pages 243-261.
A stimulus-response learning classifier system (LCS), EpiCS, was developed from the BOOLE and NEWBOOLE models to address the needs of knowledge discovery in databases used in clinical research. Two specific needs were investigated: the derivation of accurate estimates of disease risk, and the ability to deal with rare clinical outcomes. EpiCS was shown to have excellent classification accuracy, compared to logistic regression, when using risk estimates as the primary means for classification. This was especially true in data with low disease prevalence. EpiCS was designed to accommodate differential negative reinforcement when false positive or false negative decisions were made by the system. This feature was investigated to determine its effect on learning rate and classification accuracy. Tested across a range of disease prevalences, the learning rate improved when erroneous decisions were differentially negatively reinforced. However, classification accuracy was not affected by differential negative reinforcement.

[457]
John H. Holmes. A representation for accuracy-based assessment of classifier system prediction performance. In Lanzi et al. [546], pages 43-56.

[458]
Keith J. Holyoak, K. Koh, and Richard E. Nisbett. A Theory of Conditioning: Inductive Learning within Rule-Based Default Hierarchies. Psych. Review, 96:315-340, 1990.

[459]
Jeffrey Horn and David E. Goldberg. Natural Niching for Cooperative Learning in Classifier Systems. In Koza et al. [527], pages 553-564.

[460]
Jeffrey Horn and David E. Goldberg. A Timing Analysis of Convergence to Fitness Sharing Equilibrium. In Parallel Problem Solving from Nature (PPSN), 1998.
Fitness sharing has been shown to be an effective niching mechanism in genetic algorithms (GAs). Sharing allows GAs to maintain multiple, cooperating ``species'' in a single population for many generations under severe selective pressure. While recent studies have shown that the maintenance time for niching equilibrium is long, it has never been shown that the time it takes to reach equilibrium is sufficiently fast. While experiments indicate that selection under fitness sharing drives the population to equilibrium just as fast and as effectively as selection alone drives the simple GA to a uniform population, we can now show analytically that this is the case.

[461]
Jeffrey Horn and David E. Goldberg. Towards a Control Map for Niching. In Foundations of Genetic Algorithms (FOGA), pages 287-310, 1998.
Niching can allow a diverse population to cooperatively represent a single, distributed solution to the problem at hand. Successful niching mechanisms must promote both cooperation (i.e., co- existence of separate ``species'' for each desired niche), and competition (i.e., intensive search for the best species for each niche, and for the best niches). In this paper we seek the competitive- cooperative boundary in the space of possible niche relationships, that will allow us to successfully predict which pairs of interacting niches will survive under GA selection and which niche pairs will be resolved to yield a single winner. By combining extant models of niching equilibrium, niche maintenance, and convergence, we define the regions of cooperation and competition on a map of niching scenarios varying along the dimensions of niche overlap and relative niche fitness. We verify this predictive map of niching failure/success, and discuss its utility in allowing us to control for the competitive evolution of desired types of cooperation. Although our models are specific to the niching mechanism we call resource sharing, we believe the development of competitive-cooperative control maps is important for niching theory in general.

[462]
Jeffrey Horn, David E. Goldberg, and Kalyanmoy Deb. Implicit Niching in a Learning Classifier System: Nature's Way. Evolutionary Computation, 2(1):37-66, 1994. Also IlliGAL Report No 94001, 1994.
We approach the difficult task of analyzing the complex behavior of even the simplest learning classifier system (LCS) by isolating one crucial subfunction in the LCS learning algorithm: covering through niching. The LCS must maintain a population of diverse rules that together solve a problem (e.g., classify examples). To maintain a diverse population while applying the GA's selection operator, the LCS must incorporate some kind of niching mechanism. The natural way to accomplish niching in an LCS is to force competing rules to share resources (i.e., rewards). This implicit LCS fitness sharing is similar to the explicit fitness sharing used in many niched GAs. Indeed, the LCS implicit sharing algorithm can he mapped onto explicit fitness sharing with a one-to-one correspondence between algorithm components. This mapping is important because several studies of explicit fitness sharing, and of niching in GAs generally, have produced key insights and analytical tools for understanding the interaction of the niching and selection forces. We can now bring those results to bear in understanding the fundamental type of cooperation (a.k.a. weak cooperation) that an LCS must promote.

[463]
Jeffrey Horn. The Nature of Niching: Genetic Algorithms and the Evolution of Optimal, Cooperative Populations. PhD thesis, University of Illinois at Urbana-Champaign (UMI Dissertation Service No. 9812622, 1997.

[464]
D. Howard and L. Bull. On the effects of node duplication and connection-orientated constructivism in neural XCSF. In M. Keijzer et al., editor, GECCO-2008: Proceedings of the Genetic and Evolutionary Computation Conference, pages 1977-1984. ACM, 2008.

[465]
D. Howard, L. Bull, and P.L. Lanzi. Self-Adaptive Constructivism in Neural XCS and XCSF. In M. Keijzer et al., editor, GECCO-2008: Proceedings of the Genetic and Evolutionary Computation Conference, pages 1389-1396. ACM, 2008.

[466]
Dijia Huang. A framework for the credit-apportionment process in rule-based systems. IEEE Transactions on Systems, Man and Cybernetics, 1989.

[467]
Dijia Huang. Credit Apportionment in Rule-Based Systems: Problem Analysis and Algorithm Synthesis. PhD thesis, University of Michigan, 1989.

[468]
Dijia Huang. The Context-Array Bucket-Brigade Algorithm: An Enhanced Approach to Credit-Apportionment in Classifier Systems. In Schaffer [718], pages 311-316.
The credit-apportionment problem in the classifier system refers to the problem of assigning credit or blame to each rule involved in achieving a certain goal. This paper presents an enhanced credit-apportionment algorithm: the context-array bucket-brigade algorithm for solving the credit apportionment problems in the classifier system. This algorithm separates the contexts (the circumstances in which rules can fire) into context subsets. Then it employs array-valued bids (strengths) to estimate rule usefulness at the context subset level. The essence of this algorithm is that (i) it improves credit-apportionment and (ii) it provides explicit information for rule discovery. Favorable results have been observed in the tests of two versions of the algorithm.

[469]
Jacob Hurst and Larry Bull. A Self-Adaptive Classifier System. In Proceedings of the International Workshop on Learning Classifier Systems (IWLCS-2000), in the Joint Workshops of SAB 2000 and PPSN 2000 [487]. Extended abstract.
Paper is an extended abstract

[470]
Jacob Hurst and Larry Bull. A Self-Adaptive XCS. In Lanzi et al. [546], pages 57-73.

[471]
J. Hurst and L. Bull. Self-adaptation in classifier system controllers. Artificial Life and Robotics, 5(2):109-119, 2003.

[472]
J. Hurst and L. Bull. A self-adaptive neural learning classifier system with constructivism for mobile robot control. In X. Yao et al., editor, Parallel problem solving from nature (PPSN VIII), volume 3242 of LNCS, pages 942-951. Springer, 2004.

[473]
Francesc Xavier Llorà i Fàbrega and Josep Maria Garrell i Guiu. GENIFER: A Nearest Neighbour based Classifier System using GA. In Banzhaf et al. [32], page 797. One page poster paper appeared at GECCO. The full version is available at http://www.salleurl.edu/~xevil/Work/index.html.
This work has no abstract

[474]
Francesc Xavier Llorà i Fàbrega, Josep Maria Garrell i Guiu, and Ester Bernadó i Mansilla. A Classifier System based on Genetic Algorithm under the Pittsburgh approach for problems with real valued attributes. In Viceng Torra, editor, Proceedings of Artificial Intelligence Catalan Workshop (CCIA98), volume 14-15, pages 85-93. ACIA Press, 1998. In Catalan http://www.salleurl.edu/~xevil/Work/index.html.

[475]
Francesc Xavier Llorà i Fàbrega. Automatic Classification using genetic algorithms under a Pittsburgh approach. Master's thesis, Enginyeria La Salle - Ramon Llull University, 1998. http://www.salleurl.edu/~xevil/Work/index.html.
This master thesis explores the application of genetic algorithms as a supervised machine learning technique. It studies different classifications approaches (Michigan and Pittsburgh), and generates different individual codification and evaluation function alternatives. The classification centers its scope in problems described using real value attributes. All these approaches are modelled under a Pittsburgh philosophy. To validated them, and to analize they performance, they are applied to solve a real world problem: ``The automatic classification of mammary biopsy images''.

[476]
Josep Maria Garrell i Guiu, Elisabet Golobardes i Ribé, Ester Bernadó i Mansilla, and Francesc Xavier Llorà i Fàbrega. Automatic Classification of mammary biopsy images with machine learning techniques. In E. Alpaydin, editor, Proceedings of Engineering of Intelligent Systems (EIS'98), volume 3, pages 411-418. ICSC Academic Press, 1998. http://www.salleurl.edu/~xevil/Work/index.html.
This paper describes the application of Machine Learning (ML) techniques to a real world problem: the Automatic Diagnosis (classification) of Mammary Biopsy Images. The starting point consists of a set of data (solved cases) provided by the Signal Theory Research Group of our University [9]. The techniques applied are Genetic Algorithms (GA) and Case-Based Reasoning (CBR). The paper compares our results with previous ones obtained using Neural Networks (NN) [10]. The main goals are: to efficiently solve classification problems of such a type, to compare different alternatives for ML and to study hybrid systems. The paper also introduces the systems we developed for solving this kind of classification problems: GeB-CS (Genetic Based Classifier System) for a GA approach, and CaB-CS (Case-Based Classifier System) for a CBR approach.

[477]
Josep Maria Garrell i Guiu, Elisabet Golobardes i Ribé, Ester Bernadó i Mansilla, and Francesc Xavier Llorà i Fàbrega. Automatic Diagnosis with Genetic Algorithms and Case-Based Reasoning. Artificial Intelligence in Engineering, 13(4):367-372, 1999. (This is an expanded version of Guiu98a).

[478]
H. Iba, H. de Garis, and T. Higuchi. Evolutionary Learning of Predatory Behaviors Based on Structured Classifiers. In Roitblat and Wilson [700], pages 356-363.

[479]
H. Inoue, K. Takadama, M. Okada, K. Shimohara, , and O. Katai. Agent architecture based on self-reflection learning classifier system. In The 5th International Symposium on Artificial Life and Robotics (AROB'2000), pages 454-457, 2000.

[480]
H. Inoue, K. Takadama, and K. Shimohara. Inference of user's internal states and its agent's architecture. In The 20th System Engineering Meeting of SICE (The Society of Instrument and Control Engineers), pages 55-60, 2000.

[481]
Charalambos Ioannides and Will Browne. Investigating scaling of an abstracted lcs utilising ternary and s-expression alphabets. In Jaume Bacardit, Ester Bernadó-Mansilla, Martin Butz, Tim Kovacs, Xavier Llorà, and Keiki Takadama, editors, Learning Classifier Systems. 10th and 11th International Workshops (2006-2007), volume 4998/2008 of Lecture Notes in Computer Science, pages 46-56. Springer, 2008.
Utilising the expressive power of S-Expressions in Learning Classifier Systems often prohibitively increases the search space due to increased flexibility of the encoding. This work shows that selection of appropriate S-Expression functions through domain knowledge improves scaling in problems, as expected. It is also known that simple alphabets perform well on relatively small sized problems in a domain, e.g. ternary alphabet in the 6, 11 and 20 bit MUX domain. Once fit ternary rules have been formed it was investigated whether higher order learning was possible and whether this staged learning facilitated selection of appropriate functions in complex alphabets, e.g. selection of S-Expression functions. This novel methodology is shown to provide compact results (135-MUX) and exhibits potential for scaling well (1034-MUX), but is only a small step towards introducing abstraction to LCS.

[482]
N. Ireson, Y. J. Cao, L. Bull, and R. Miles. A Communication Architecture for Multi-Agent Learning Systems. In Proceedings of the EvoNet Workshops - EvoTel 2000, pages 255-266, 2000.

[483]
Hisao Ishibuchi and Tomoharu Nakashima. Linguistic Rule Extraction by Genetics-Based Machine Learning. In Whitely et al. [878], pages 195-202.
This paper shows how linguistic classification knowledge can be extracted fro numerical data for pattern classification problems with many continuous attributes by genetic algorithms. Classification knowledge is extracted in the form of linguistic if-then rules. In this paper, emphasis is placed on the simplicity of the extracted knowledge. The simplicity is measured by two criteria: the number of extracted linguistic rules and the length of each rule (i.e, the number of antecedent conditions involved in each rule). The classification ability of extracted linguistic rules, which is measured by the classification rate on given training patterns, is also considered. Thus our task is formulated as a linguistic rule extraction problem with three objectives: to maximize the classification rate, to minimize the number of extracted rules, and to minimize the length of each rule. For tackling this problem, we propose a multi-objective genetics-based machine learning (GBML) algorithm, which is a hybrid algorithm of Michigan approach and Pittsburgh approach. Our hybrid algorithm is basically a Pittsburgh-style algorithm with variable string length. A Michigan-style algorithm is combined as a kind of mutation for partially modifying each string.

[484]
Hisao Ishibuchi and Takashi Yamamoto. Fuzzy rule selection by data mining criteria and genetic algorithms. In W. B. Langdon, E. Cantú-Paz, K. Mathias, R. Roy, D. Davis, R. Poli, K. Balakrishnan, V. Honavar, G. Rudolph, J. Wegener, L. Bull, M. A. Potter, A. C. Schultz, J. F. Miller, E. Burke, and N. Jonoska, editors, GECCO 2002: Proceedings of the Genetic and Evolutionary Computation Conference, pages 399-406, New York, 9-13 July 2002. Morgan Kaufmann Publishers.

[485]
Yasushi Ishikawa and Takao Terano. Co-evolution of multiagents via organizational-learning classifier system and its application to marketing simulation. In Proc. 4th Pacific-Asia Conf. on Information Systems ( PACIS-2000), pages 1114-1127, 2000.

[486]
Collected Abstracts for the First International Workshop on Learning Classifier System (IWLCS92), 1992. October 6-8, NASA Johnson Space Center, Houston, Texas.

[487]
Proceedings of the International Workshop on Learning Classifier Systems (IWLCS-2000), in the Joint Workshops of SAB 2000 and PPSN 2000, 2000. Pier Luca Lanzi, Wolfgang Stolzmann and Stewart W. Wilson (workshop organisers).

[488]
Jacob Hurst, Larry Bull and Chris Melhuish. TCS learning classifier system controller on a real robot. In H.-P. Schwefel J.-J. Merelo Guervós, P. Adamidis, H.-G. Beyer, J.-L. Fernández-Villacañas, editor, Parallel Problem Solving from Nature - PPSN VII, 7th International Conference, Granada, Spain, September 7-11, 2002. Proceedings, number 2439 in Lecture Notes in Computer Science, LNCS, page 588 ff. Springer-Verlag, 2002.

[489]
C.Z. Janikow. Indictive learning of decision rules in attribute-based examples: a knowledge-intensive genetic algorithm approach. PhD thesis, University of North Carolina, 1991.

[490]
C.Z. Janikow. A knowledge-intensive genetic algorithm for supervised learning. Machine Learning, 13:189-228, 1993.

[491]
Cezary Z. Janikow. Genetic information learning. In Thomas Bäck, David B. Fogel, and Zbigniew Michalewicz, editors, Handbook of Evolutionary Computation, pages G2.3:1-G2.3:10. IOP Publishing Ltd and Oxford University Press, 1997.

[492]
Kenneth A. De Jong and William M. Spears. Learning Concept Classification Rules using Genetic Algorithms. In Proceedings of the Twelfth International Conference on Artificial Intelligence IJCAI-91, volume 2, 1991.

[493]
K. Takadama, T. Terano, K. Shimohara, K. Hori and S. Nakasuka. Towards a multiagent design principle - analyzing an organizational-learning oriented classifier system. In V. Loia and S. Sessa, editors, Soft Computing Agents: New Trends for Designing Autonomous Systems, Series of Studies in Fuzziness and Soft Computing. Springer-Verlag, 2001.

[494]
Daisuke Katagami and Seiji Yamada. Real robot learning with human teaching. In Takadama [815].

[495]
Hiroharu Kawanaka, Tomohiro Yoshikawa, and Shinji Tsuruoka. A Study of Parallel GA Using DNA Coding Method for Acquisition of Fuzzy Control Rules. In Late Breaking Papers at the 2000 Genetic and Evolutionary Computation Conference (GECCO-2000), pages 431-436, 2000.
Fuzzy controls have been widely used in industry for its high degree of performance in human-computer interactions. DNA coding method, which is one of the coding methods in Genetic Algorithms, is based on biological DNA and a mechanism of development from the artificial DNA. This method has redundancy and overlapping of genes, and it is suitable for knowledge representation. In this paper, we propose the Parallel Genetic Algorithm using the DNA coding method. This paper applies this method to acquisition of fuzzy control rules with multiple input/output system for a mobile robot. This method can select input variables from many candidates and tune membership functions. The result of simulation shows that the robot can reach the goal quickly and efficiently. Effective fuzzy rules for the mobile robot are acquired by using this method while the length of the chromosomes in the population is automatically adjusted.

[496]
Yeong-Joon Kim and Christoph F. Eick. Multi-rule-set decision-making schemes for a genetic algorithm learning environment for classification tasks. In John R. McDonnell, Robert G. Reynolds, and David B. Fogel, editors, Evolutionary Programming IV. Proceedings of the Fourth Annual Conference on Evolutionary Programming, pages 773-788, 1995.
Over the last three years, we developed an inductive learning environment called DELVAUX for classification tasks that learns PROSPECTOR-style, Bayesian rules from ests of examples, using a genetic algorithm to evolve a population consists of rule-sets. Several problems comlicate the search for the best rule-set. First, the search space that is explored by DELVAUX is enormously large, which makes it difficult to predict if a particular solution is a good solution. The second problem is the problem of convergence with outliers that perform well in training but not in testing. This paper describes efforts to alleviate these two problems centering on multi-rule-set learning techniques that learn multiple rule-sets and proposes several decision-making schemes that are employed by the multi-rule-set learning environment to derive a decision. Empirical results are presented that compare the single rule-set learning environment of DELVAUX with several multi-rule-set learning environments that use different decision-making schemes. Moreover, a more sophisticated fitness function for the multi-rule-set learning approach is introduced, and a genetic algorithm approach that finds the `best' multi-rule-set for a given set of rule-sets is discussed.

[497]
Hiroaki Kitano, Stephen F. Smith, and Tetsuya Higuchi. GA-1: A Parallel Associative Memory Processor for Rule Learning with Genetic Algorithms. In Booker and Belew [74], pages 311-317.
In this paper, we discuss the underlying hardware and supportable learning paradigms provided by the GA-1 system. GA-1 is a system currently under development which offers unique opportunities for research into large-scale rule learning with genetic algorithms (GAs). The base hardware is the IXM2 parallel associative memory machine which enables high performance processing by using 64 T800 transputers and associative memories providing 256K parallelism. Various population/subpopulation models, mating strategies, and generation models can be implemented to investigate architectures for high performance GA-based systems. Regardless of these options, however, GA-based rule learning takes maximum advantage of the hardware through extensive use of associative memory for bit-vector matching. Preliminary experiments indicate that GA-1 exhibits high execution speeds for such an approach.

[498]
Leslie Knight and Sandip Sen. PLEASE: A Prototype Learning System using Genetic Algorithms. In Eshelman [298], pages 429-435.

[499]
Gabriella Kókai, Zoltán Tóth, and Szilvia Zvada. An experimental comparison of genetic and classical concept learning methods. In W. B. Langdon, E. Cantú-Paz, K. Mathias, R. Roy, D. Davis, R. Poli, K. Balakrishnan, V. Honavar, G. Rudolph, J. Wegener, L. Bull, M. A. Potter, A. C. Schultz, J. F. Miller, E. Burke, and N. Jonoska, editors, GECCO 2002: Proceedings of the Genetic and Evolutionary Computation Conference, page 952. Morgan Kaufmann Publishers, 2002.

[500]
Kostyantyn Korovkin and Robert Richards. Visual Auction: A Classifier System Pedagogical and Researcher Tool. In Scott Brave and Annie S. Wu, editors, Late Breaking Papers at the 1999 Genetic and Evolutionary Computation Conference (GECCO-99), pages 159-163, 1999.
This paper surveys the Visual Auction. The Visual Auction is a pedagogical and research tool, which provides a dynamic visual representation of the matching and auction process in a classifier system. The tool allows for the visual representation of both the matching process and the process of determining the winner of the auction. The tool can be used pedagogically, for it quickly demonstrates how antecedent matching occurs in a classifier system and shows how the matched classifiers compete to win the auction. As a research tool, it assists the researcher in implementing a classifier system by providing more in-depth knowledge of the auction process, thus providing insight into how parameter settings or bid equations could be modified to generate more efficient learning.

[501]
Tim Kovacs and Manfred Kerber. Some dimensions of problem complexity for XCS. In Annie S. Wu, editor, Proceedings of the 2000 Genetic and Evolutionary Computation Conference Workshop Program, pages 289-292, 2000.
Despite two decades of work, learning classifier systems researchers have had relatively little to say on the subject of what makes a problem difficult for a classifier system. One focus of our work has been the issue of what makes a problem difficult for XCS -- Wilson's recent accuracy-based classifier system. This document outlines the approach taken, provides some initial results and outlines possible directions for future work.

[502]
Tim Kovacs and Manfred Kerber. What makes a problem hard for XCS? In Proceedings of the International Workshop on Learning Classifier Systems (IWLCS-2000), in the Joint Workshops of SAB 2000 and PPSN 2000 [487]. Extended abstract.
Paper is an extended abstract

[503]
Tim Kovacs and Manfred Kerber. What makes a problem hard for XCS? In Lanzi et al. [545], pages 80-99.
Despite two decades of work learning classifier systems researchers have had relatively little to say on the subject of what makes a problem difficult for a classifier system. Wilson's accuracy-based XCS, a promising and increasingly popular classifier system, is, we feel, the natural first choice of classifier system with which to address this issue. To make the task more tractable we limit our considerations to a restricted, but very important, class of problems. Most significantly, we consider only single step reinforcement learning problems and the use of the standard binary/ternary classifier systems language. In addition to distinguishing several dimensions of problem complexity for XCS, we consider their interactions, identify bounding cases of difficulty, and consider complexity metrics for XCS. Based on these results we suggest a simple template for ternary single step test suites to more comprehensively evaluate classifier systems.

[504]
Tim Kovacs and Manfred Kerber. High classification accuracy does not imply effective genetic search. In K. Deb et al., editor, Proceedings of the 2004 Genetic and Evolutionary Computation Conference (GECCO), volume 3102 of LNCS, pages 785-796. Springer, 2004.

[505]
Tim Kovacs and Manfred Kerber. A study of structural and parametric learning in XCS. Evolutionary Computation, 14(1):1-19, 2006.

[506]
Tim Kovacs and Pier Luca Lanzi. A Learning Classifier Systems Bibliography. Technical Report 99.52, Dipartimento di Elettronica e Informazione, Politecnico di Milano, 1999.
We present a bibliography of all works we could find on Learning Classifier Systems (LCS) -- the genetics-based machine learning systems introduced by John Holland. With over 400 entries, this is at present the largest bibliography on classifier systems in existence. We include a list of LCS resources on the world wide web.

[507]
Tim Kovacs and Pier Luca Lanzi. A Learning Classifier Systems Bibliography. In Lanzi et al. [544], pages 321-347.
We present a bibliography of all works we could find on Learning Classifier Systems (LCS) -- the genetics-based machine learning systems introduced by John Holland. With over 400 entries, this is at present the largest bibliography on classifier systems in existence. We include a list of LCS resources on the world wide web.

[508]
Tim Kovacs and Pier Luca Lanzi. A Bigger Learning Classifier Systems Bibliography. In Lanzi et al. [545], pages 213-249.
With over 600 entries, this is by far the most comprehensive bibliography of the machine learning systems introduced by John Holland.

[509]
Tim Kovacs. Evolving Optimal Populations with XCS Classifier Systems. Master's thesis, School of Computer Science, University of Birmingham, Birmingham, U.K., 1996. Also technical report CSR-96-17 and CSRP-96-17 ftp://ftp.cs.bham.ac.uk/pub/tech-reports/1996/CSRP-96-17.ps.gz.
This work investigates some uses of self-monitoring in classifier systems (CS) using Wilson's recent XCS system as a framework. XCS is a significant advance in classifier systems technology which shifts the basis of fitness evaluation for the Genetic Algorithm (GA) from the strength of payoff prediction to the accuracy of payoff prediction. Initial work consisted of implementing an XCS system in Pop-11 and replicating published XCS multiplexer experiments from (Wilson 1995, 1996a). In subsequent original work, the XCS Optimality Hypothesis, which suggests that under certain conditions XCS systems can reliably evolve optimal populations (solutions), is proposed. An optimal population is one which accurately maps inputs to actions to reward predictions using the smallest possible set of classifiers. An optimal XCS population forms a complete mapping of the payoff environment in the reinforcement learning tradition, in contrast to traditional classifier systems which only seek to maximise classifier payoff (reward). The more complete payoff map allows XCS to deal with payoff landscapes with more than 1 niche (i.e. those with more than 2 payoff levels) which traditional payoff-maximising CS find very difficult. This makes XCS much more suitable as the foundation of animat control systems than traditional CS. In support of the Optimality Hypothesis, techniques were developed which allow the system to highly reliably evolve optimal populations for logical multiplexer functions. A technique for auto-termination of learning was also developed to allow the system to recognise when an optimal population has been evolved. The self-monitoring mechanisms involved in this work are discussed in terms of the design space of adaptive systems.

[510]
Tim Kovacs. Steady State Deletion Techniques in a Classifier System. Unpublished document -- partially subsumed by Kovacs1999a `Deletion Schemes for Classifier Systems', 1997.
In a standard genetic algorithm a chromosome can be fully evaluated (assigned a fitness) immediately. In classifier systems, however, a chromosome can only be fully evaluated after many interactions with the environment, since a chromosome may generalise over many environmental states. In this work it is suggested that evolutionary systems which cannot fully evaluate candidate solutions immediately will benefit from protecting them from deletion until they have been well evaluated. A new technique which protects poorly evaluated chromosomes outperforms both techniques from (Wilson, 1995) on two types of boolean function and a delayed reward problem. Next a weeding operator which deletes low fitness rules is introduced and found to improve performance on one boolean function. Results indicate the XCS classifier system is able to learn boolean functions for which no (or few) useful generalisations can be made over the input string, despite its drive towards accurate generalisation.

[511]
Tim Kovacs. XCS Classifier System Reliably Evolves Accurate, Complete, and Minimal Representations for Boolean Functions.. In Roy, Chawdhry, and Pant, editors, Soft Computing in Engineering Design and Manufacturing, pages 59-68. Springer-Verlag, London, 1997. ftp://ftp.cs.bham.ac.uk/pub/authors/T.Kovacs/index.html.
Wilson's recent XCS classifier system forms complete mappings of the payoff environment in the reinforcement learning tradition thanks to its accuracy based fitness. According to Wilson's Generalization Hypothesis, XCS has a tendency towards generalization. With the XCS Optimality Hypothesis, I suggest that XCS systems can evolve optimal populations (representations); populations which accurately map all input/action pairs to payoff predictions using the smallest possible set of non-overlapping classifiers. The ability of XCS to evolve optimal populations for boolean multiplexer problems is demonstrated using condensation, a technique in which evolutionary search is suspended by setting the crossover and mutation rates to zero. Condensation is automatically triggered by self-monitoring of performance statistics, and the entire learning process is terminated by autotermination. Combined, these techniques allow a classifier system to evolve optimal representations of boolean functions without any form of supervision. A more complex but more robust and efficient technique for obtaining optimal populations called subset extraction is also presented and compared to condensation.

[512]
Tim Kovacs. XCS Classifier System Reliably Evolves Accurate, Complete, and Minimal Representations for Boolean Functions. Technical Report Version.. Technical Report CSRP-97-19, School of Computer Science, University of Birmingham, Birmingham, U.K., 1997. http://www.cs.bham.ac.uk/system/tech-reports/tr.html.
This paper extends the work presented in (Kovacs, 1996) on evolving optimal solutions to boolean reinforcement learning problems using Wilson's recent XCS classifier system. XCS forms complete mappings of the payoff environment in the reinforcement learning tradition thanks to its accuracy based fitness, which, according to Wilson's Generalization Hypothesis, also gives XCS a tendency towards accurate generalization. (Kovacs, 1996) introduced the XCS Optimality Hypothesis which suggests that XCS systems can evolve optimal populations (representations); populations which accurately map all input/action pairs to payoff predictions using the smallest possible set of non-overlapping classifiers. The ability of XCS to evolve optimal populations for boolean multiplexer problems was demonstrated in (Kovacs, 1996) using condensation, a technique in which evolutionary search is suspended by setting the crossover and mutation rates to zero. Condensation is automatically triggered by self-monitoring of performance statistics, and the entire learning process is terminated by autotermination. Combined, these techniques allow a classifier system to evolve optimal representations of boolean functions without any form of supervision. The present work shows how condensation can be greatly accelerated by truncating each action set around its most numerous member. Following this, a more complex but more robust and efficient technique for obtaining optimal populations called subset extraction is presented and compared to condensation.

[513]
Tim Kovacs. Deletion schemes for classifier systems. In Banzhaf et al. [32], pages 329-336. Also technical report CSRP-99-08, School of Computer Science, University of Birmingham.
The issue of deletion schemes for classifier systems has received little attention. In a standard genetic algorithm a chromosome can be evaluated (assigned a reasonable fitness) immediately. In classifier systems, however, a chromosome can only be fully evaluated after many interactions with the environment, since a chromosome may generalise over many environmental states. A new technique which protects poorly evaluated chromosomes outperforms both techniques from (Wilson, 1995) in two very different single step problems. Results indicate the XCS classifier system is able to learn single step problems for which no (or few) useful generalisations can be made over the input string, despite its drive towards accurate generalisation.

[514]
Tim Kovacs. Strength or accuracy? A comparison of two approaches to fitness calculation in learning classifier systems. In Wu [923], pages 258-265.
Wilson's XCS is a clear departure from earlier classifier systems in the way it calculates the fitness of classifiers for use in the genetic algorithm. Despite the growing body of work on XCS and the advantages claimed for it, there has been no detailed comparison of XCS and traditional strength-based systems. We distinguish different definitions of overgenerality for strength and accuracy-based fitness and analyse some implications of the use of accuracy, including an advantage in exploration. We analyse the formation of strong overgenerals, a major problem for strength-based systems, and show that they require biased reward functions. We also show that all non-trivial multi step environments have biased reward functions and thus suffer from strong overgenerals. We conclude that strength-based systems are not suitable for multi step environments or indeed many single step environments.

[515]
Tim Kovacs. Strength or Accuracy? Fitness calculation in learning classifier systems. In Lanzi et al. [544], pages 143-160.
Wilson's XCS is a clear departure from earlier classifier systems in terms of the way it calculates the fitness of classifiers for use in the genetic algorithm. Despite the growing body of work on XCS and the advantages claimed for it there has been no detailed comparison of XCS and traditional strength-based systems. This work takes a step towards rectifying this situation by surveying a number of issues related to the change in fitness. I distinguish different definitions of overgenerality for strength and accuracy-based fitness and analyse some implications of the use of accuracy, including an apparent advantage in addressing the explore/exploit problem. I analyse the formation of strong overgenerals, a major problem for strength-based systems, and illustrate their dependence on biased reward functions. I consider motivations for biasing reward functions in single step environments, and show that non-trivial multi step environments have biased Q-functions. I conclude that XCS's accuracy-based fitness appears to have a number of significant advantages over traditional strength-based fitness.

[516]
Tim Kovacs. Towards a theory of strong overgeneral classifiers. In Worthy Martin and William M. Spears, editors, Foundations of Genetic Algorithms (FOGA) Volume 6, pages 165-184. Morgan Kaufmann, 2000. Also technical report CSRP-00-20, School of Computer Science, University of Birmingham.
We analyse the concept of strong overgeneral rules, the Achilles' heel of traditional Michigan-style learning classifier systems, using both the traditional strength-based and newer accuracy-based approaches to rule fitness. We argue that different definitions of overgenerality are needed to match the goals of the two approaches, present minimal conditions and environments which will support strong overgeneral rules, demonstrate their dependence on the reward function, and give some indication of what kind of reward functions will avoid them. Finally, we distinguish fit overgeneral rules, show how strength and accuracy-based fitness differ in their response to fit overgenerals and conclude by considering possible extensions to this work.

[517]
Tim Kovacs. Trends in learning classifier systems publication. Technical Report CSRP-00-21, School of Computer Science, University of Birmingham, 2000.
Using data from the world's most comprehensive Learning Classifier Systems (LCS) bibliography, we examine trends in LCS publication and attempt to account for them. We find support for the notions of a classic period and an ongoing LCS renaissance, and find that Wilson's XCS has become a major focus of LCS research.

[518]
Tim Kovacs. What should a classifier system learn? In Proceedings of the 2001 Congress on Evolutionary Computation (CEC01) [186], pages 775-782.

[519]
Tim Kovacs. A Comparison and Strength and Accuracy-based Fitness in Learning Classifier Systems. PhD thesis, University of Birmingham, 2002.

[520]
Tim Kovacs. Learning Classifier Systems Resources. Journal of Soft Computing, 6(3-4):240-243, 2002.
This article lists currently available sources of information on classifier systems and classifier systems research, both on-line and in print. The need for new resources, and improvements to certain existing ones, are suggested.

[521]
Tim Kovacs. Performance and population state metrics for rule-based learning systems. In David B. Fogel, Mohamed A. El-Sharkawi, Xin Yao, Garry Greenwood, Hitoshi Iba, Paul Marrow, and Mark Shackleton, editors, Proceedings of the 2002 Congress on Evolutionary Computation CEC2002, pages 1781-1786. IEEE Press, 2002.

[522]
Tim Kovacs. Two views of classifier systems. In Lanzi et al. [546], pages 74-87.
This work suggests two ways of looking at Michigan classifier systems; as Genetic Algorithm-based systems, and as Reinforcement Learning-based systems, and argues that the former is more suitable for traditional strength-based systems while the latter is more suitable for accuracy-based XCS. The dissociation of the Genetic Algorithm from policy determination in XCS is noted, and the two types of Michigan classifier system are contrasted with Pittsburgh systems.

[523]
Tim Kovacs. What should a classifier system learn and how should we measure it? Journal of Soft Computing, 6(3-4):171-182, 2002.
We consider the issues of how a classifier system should learn to represent a Boolean function, and how we should measure its progress in doing so. We identify four properties which may be desirable of a representation; that it be complete, accurate, minimal and non-overlapping. We distinguish two categories of learning metric, introduce new metrics and evaluate them. We demonstrate the superiority of population state metrics over performance metrics in two situations, and in the process find evidence of XCS's strong bias against overlapping rules.

[524]
Tim Kovacs. Strength or Accuracy: Credit Assignment in Learning Classifier Systems. Springer, 2004.

[525]
Tim Kovacs. Genetics-based machine learning. In Grzegorz Rozenberg, Thomas Bäck, and Joost Kok, editors, Handbook of Natural Computing: Theory, Experiments, and Applications. Springer Verlag, 2010.

[526]
Yuhsuke Koyama. The emergence of the cooperative behaviors in a small group. In Takadama [815].

[527]
John R. Koza, David E. Goldberg, David B. Fogel, and Rick L. Riolo, editors. Genetic Programming 1996: Proceedings of the First Annual Conference, Stanford University, CA, USA, 1996. MIT Press.

[528]
John R. Koza, Kalyanmoy Deb, Marco Dorigo, David B. Fogel, Max H. Garzon, Hitoshi Iba, and Rick Riolo, editors. Genetic Programming 1997: Proceedings of the Second Annual Conference. Morgan Kaufmann, 1997.

[529]
John R. Koza, Wolfgang Banzhaf, Kumar Chellapilla, Kalyanmoy Deb, Marco Dorigo, David B. Fogel, Max H. Garzon, David E. Goldberg, Hitoshi Iba, and Rick Riolo, editors. Genetic Programming 1998: Proceedings of the Third Annual Conference. Morgan Kaufmann, 1998.

[530]
Setsuya Kurahashi and Takao Terano. Technology extraction of expert operator skills from process time series data. In Jaume Bacardit, Ester Bernadó-Mansilla, Martin Butz, Tim Kovacs, Xavier Llorà, and Keiki Takadama, editors, Learning Classifier Systems. 10th and 11th International Workshops (2006-2007), volume 4998/2008 of Lecture Notes in Computer Science, pages 269-285. Springer, 2008.
Continuation processes in chemical and/or biotechnical plants always generate a large amount of time series data. However, since conventional process models are described as a set of control models, it is difficult to explain complicated and active plant behaviors. To uncover complex plant behaviors, this paper proposes a new method of developing a process response model from continuous time-series data. The method consists of the following phases: (1) Reciprocal correlation analysis; (2) Process response model; (3) Extraction of control rules; (4) Extraction of a workflow; and (5) Detection of outliers. The main contribution of the research is to establish a method to mine a set of meaningful control rules from a Learning Classifier System using the Minimum Description Length criteria and Tabu search method. The proposed method has been applied to an actual process of a biochemical plant and has shown its validity and effectiveness.

[531]
Samuel Landau, Sébastien Picault, Oliver Sigaud, and Pierre Gérard. A comparison between ATNoSFERES and XCSM. In W. B. Langdon, E. Cantú-Paz, K. Mathias, R. Roy, D. Davis, R. Poli, K. Balakrishnan, V. Honavar, G. Rudolph, J. Wegener, L. Bull, M. A. Potter, A. C. Schultz, J. F. Miller, E. Burke, and N. Jonoska, editors, GECCO 2002: Proceedings of the Genetic and Evolutionary Computation Conference, pages 926-933. Morgan Kaufmann Publishers, 2002.

[532]
Samuel Landau, Sébastien Picault, Olivier Sigaud, and Pierre Gérard. Further Comparison between ATNoSFERES and XCSM. In Wolfgang Stolzmann, Pier Luca Lanzi, and Stewart W. Wilson, editors, IWLCS-02. Proceedings of the Fifth International Workshop on Learning Classifier Systems, LNAI. Springer, 2002.

[533]
Samuel Landau, Olivier Sigaud, Sébastien Picault, and Pierre Gérard. An Experimental Comparison between ATNoSFERES and ACS. In Wolfgang Stolzmann, Pier Luca Lanzi, and Stewart W. Wilson, editors, IWLCS-03. Proceedings of the Sixth International Workshop on Learning Classifier Systems, LNAI. Springer, 2003.

[534]
Samuel Landau, Olivier Sigaud, and Marc Schoenauer. ATNoSFERES revisited. In Proceedings of the Genetic and Evolutionary Computation Conference GECCO-2005, pages 1867-1874. ACM, 2005.
ATNoSFERES is a Pittsburgh style Learning Classifier System (LCS) in which the rules are represented as edges of an Augmented Transition Network. Genotypes are strings of tokens of a stack-based language, whose execution builds the labeled graph. The original ATNoSFERES, using a bitstring to represent the language tokens, has been favorably compared in previous work to several Michigan style LCSs architectures in the context of Non Markov problems. Several modifications of ATNoSFERES are proposed here: the most important one conceptually being a representational change: each token is now represented by an integer, hence the genotype is a string of integers; several other modifications of the underlying grammar language are also proposed. The resulting ATNoSFERES-II is validated on several standard animat Non Markov problems, on which it outperforms all previously published results in the LCS literature. The reasons for these improvement are carefully analyzed, and some assumptions are proposed on the underlying mechanisms in order to explain these good results.

[535]
Samuel Landau, Olivier Sigaud, Sébastien Picault, and Pierre Gérard. An experimental comparison between atnosferes and acs. In Tim Kovacs, Xavier LLòra, Keiki Takadama, Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson, editors, Learning Classifier Systems. International Workshops, IWLCS 2003-2005, Revised Selected Papers, volume 4399 of LNCS, pages 144-160. Springer, 2007.

[536]
Pier Luca Lanzi and Marco Colombetti. An Extension of XCS to Stochastic Environments. Technical Report 98.85, Dipartimento di Elettronica e Informazione - Politecnico di Milano, 1998.

[537]
Pier Luca Lanzi and Marco Colombetti. An Extension to the XCS Classifier System for Stochastic Environments. In Banzhaf et al. [32], pages 353-360.
We analyze XCS learning capabilities in stochastic environments where the result of agent actions can be uncertain. We show that XCS can deal when degree of environmental uncertainty is limited. We analyze our experimental results and propose an extension to XCS, called XCS_mu, which can learn optimal solutions for higher degrees of uncertainty. We test XCS_mu when the uncertainty affects agent actions in the whole environment and when the uncertainty is limited to some areas. Finally, we show that XCS_mu is a proper extension of XCS in that it coincides with it when it is applied to deterministic environments.

[538]
P.L. Lanzi and D. Loiacono. Standard and averaging reinforcement learning in XCS. In M. Cattolico, editor, GECCO 2006: Proceedings of the 8th annual conference on genetic and evolutionary computation, pages 1480-1496. ACM, 2006.

[539]
P.L. Lanzi and D. Loiacono. Classifier systems that compute action mappings. In H. Lipson, editor, Genetic and Evolutionary Computation Conference, GECCO 2007, Proceedings, pages 1822-1829. ACM, 2007.

[540]
Pier Luca Lanzi and Rick L. Riolo. A Roadmap to the Last Decade of Learning Classifier System Research (from 1989 to 1999). In Lanzi et al. [544], pages 33-62.
In 1989 Wilson and Goldberg presented a critical review of the first ten years of learning classifier system research. With this paper we review the subsequent ten years of learning classifier systems research, discussing the main achievements and the major research directions pursued in those years.

[541]
Pier Luca Lanzi and Stewart W. Wilson. Optimal classifier system performance in non-Markov environments. Technical Report 99.36, Dipartimento di Elettronica e Informazione - Politecnico di Milano, 1999. Also IlliGAL technical report 99022, University of Illinois.
Wilson's (1994) bit-register memory scheme was incorporated into the XCS classifier system and investigated in a series of non-Markov environments. Two extensions to the scheme proved important for reaching optimal performance in the harder environments. The first was an exploration strategy in which exploration of external actions was probabilistic as in Markov environments, but internal ``actions'' (register settings) were selected deterministically. The second was use of a register having more bit-positions than were strictly necessary to resolve environmental aliasing. The origins and effects of the two extensions are discussed.

[542]
Pier Luca Lanzi and Stewart W. Wilson. Toward Optimal Classifier System Performance in Non-Markov Environments. Evolutionary Computation, 8(4):393-418, 2000.
Wilson's (1994) bit-register memory scheme was incorporated into the XCS classifier system and investigated in a series of non-Markov environments. Two extensions to the scheme were important in obtaining near-optimal performance in the harder environments. The first was an exploration strategy in which exploration of external actions was probabilistic as in Markov environments, but internal ``actions'' (register settings) were selected deterministically. The second was use of a register having more bit-positions than were strictly necessary to resolve environmental aliasing. The origins and effects of the two extensions are discussed.

[543]
P.L. Lanzi and S.W. Wilson. Using convex hulls to represent classifier conditions. In M. Cattolico, editor, Proc. genetic and evolutionary computation conference (GECCO 2006), pages 1481-1488. ACM, 2006.

[544]
Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson, editors. Learning Classifier Systems. From Foundations to Applications, volume 1813 of LNAI. Springer-Verlag, Berlin, 2000.

[545]
Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson, editors. Advances in Learning Classifier Systems, volume 1996 of LNAI. Springer-Verlag, Berlin, 2001.

[546]
Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson, editors. Advances in Learning Classifier Systems, volume 2321 of LNAI. Springer-Verlag, Berlin, 2002.

[547]
Pier Luca Lanzi, Daniele Loiacono, Stewart W. Wilson, and David E. Goldberg. Classifier prediction based on tile coding. In Genetic and Evolutionary Computation -- GECCO-2006, pages 1497-1504. ACM, 2006.
This paper introduces XCSF extended with tile coding prediction: each classifier implements a tile coding approximator; the genetic algorithm is used to adapt both classifier conditions (i.e., to partition the problem) and the parameters of each approximator; thus XCSF evolves an ensemble of tile coding approximators instead of the typical monolithic approximator used in reinforcement learning. The paper reports a comparison between (i) XCSF with tile coding prediction and (ii) plain tile coding. The results show that XCSF with tile coding always reaches optimal performance, it usually learns as fast as the best parametrized tile coding, and it can be faster than the typical tile coding setting. In addition, the analysis of the evolved tile coding ensembles shows that XCSF actually adapts local approximators following what is currently considered the best strategy to adapt the tile coding parameters in a given problem.

[548]
Pier Luca Lanzi, Daniele Loiacono, Stewart W. Wilson, and David E. Goldberg. Prediction update algorithms for xcsf: Rls, kalman filter and gain adaptation. In Genetic and Evolutionary Computation -- GECCO-2006, pages 1505-1512. ACM, 2006.
We study how different prediction update algorithms influence the performance of XCSF. We consider three classical parameter estimation algorithms (NLMS, RLS, and Kalman filter) and four gain adaptation algorithms (K1, K2, IDBD, and IDD). The latter have been shown to perform comparably to the best algorithms (RLS and Kalman), but they have a lower complexity. We apply these algorithms to update classifier prediction in XCSF and compare the performances of the seven versions of XCSF on a set of real functions. Our results show that the best known algorithms still perform best: XCSF with RLS and XCSF with Kalman perform significantly better than the others. In contrast, when added to XCSF, gain adaptation algorithms perform comparably to NLMS, the simplest estimation algorithm, the same used in the original XCSF. Nevertheless, algorithms that perform similarly generalize differently. For instance: XCSF with Kalman filter evolves more compact solutions than XCSF with RLS and gain adaptation algorithms allow better generalization than NLMS.

[549]
P.L. Lanzi, M.V. Butz, and D.E. Goldberg. Empirical analysis of generalization and learning in XCS with gradient descent. In H. Lipson, editor, Genetic and Evolutionary Computation Conference, GECCO 2007, Proceedings, volume 2, pages 1814-1821. ACM, 2007.

[550]
P.L. Lanzi, D. Loiacono, S.W. Wilson, and D.E. Goldberg. Generalization in the XCSF classifier system: analysis, improvement, and extension. Evolutionary Computation, 15(2):133-168, 2007.

[551]
Pier Luca Lanzi, Daniele Loiacono, and Matteo Zanini. Evolving classifiers ensembles with heterogeneous predictors. In Jaume Bacardit, Ester Bernadó-Mansilla, Martin Butz, Tim Kovacs, Xavier Llorà, and Keiki Takadama, editors, Learning Classifier Systems. 10th and 11th International Workshops (2006-2007), volume 4998/2008 of Lecture Notes in Computer Science, pages 218-234. Springer, 2008.
XCS with computed prediction, namely XCSF, extends XCS by replacing the classifier prediction with a parametrized prediction function. Although several types of prediction functions have been introduced, so far XCSF models are still limited to evolving classifiers with the same prediction function. In this paper, we introduce XCSF with heterogeneous predictors, XCSFHP, which allows the evolution of classifiers with different types of prediction function within the same population. We compared XCSFHP to XCSF on several problems. Our results suggest that XCSFHP generally performs as XCSF with the most appropriate prediction function for the given problem. In particular, XCSFHP seems able to evolve, in each problem subspace, the most adequate type of prediction function.

[552]
Pier Luca Lanzi, Stefano Rocca, Kumara Sastry, and Stefania Solari. Analysis of population evolution in classifier systems using symbolic representations. In Jaume Bacardit, Ester Bernadó-Mansilla, Martin Butz, Tim Kovacs, Xavier Llorà, and Keiki Takadama, editors, Learning Classifier Systems. 10th and 11th International Workshops (2006-2007), volume 4998/2008 of Lecture Notes in Computer Science, pages 22-45. Springer, 2008.
This paper presents an approach to analyze population evolution in classifier systems using a symbolic representation. Given a sequence of populations, representing the evolution of a solution, the method simplifies the classifiers in the populations by reducing them to their “canonical form”. Then, it extracts all the subexpressions that appear in all the classifier conditions and, for each subexpression, it computes the number of occurrences in each population. Finally, it computes the trend of all the subexpressions considered. The expressions which show an increasing trend through the course of evolution are viewed as building blocks that the system has used to construct the solution.

[553]
Pier Luca Lanzi. A Model of the Environment to Avoid Local Learning (An Analysis of the Generalization Mechanism of XCS). Technical Report 97.46, Politecnico di Milano. Department of Electronic Engineering and Information Sciences, 1997. http://ftp.elet.polimi.it/people/lanzi/report46.ps.gz.
We analyze generalization with the XCS classifier system when the system is applied to animat problems in grid-worlds. The aim of the paper is to give an unified view of generalization with XCS, in order to explain some of the phenomena reported in the literature. Initially, we extend the results previously presented in the literature applying XCS to three environments. Our results confirm what already reported in the literature, showing that the generalization mechanism of XCS may prevent the system from converging to optimal performance. Accordingly, we study XCS generalization mechanism analyzing Wilson's generalization hypothesis and the conditions under which it may fail. We draw a hypothesis in order to explain the results we report. We hypothesize that XCS fails to learn an optimal solution when, due to the environment structure and to the exploration strategy employed, the system is not able to explore all the environmental niches frequently. We validate our hypothesis introducing a new exploration strategy which guarantees a frequent exploration of all the areas of the environment, we call it teletransportation. Teletransportation is not introduced as a real solution to the problems we evidenced, since it is not feasible for real applications; rather, we exploit teletransportation as a tool to validate our hypothesis. Nevertheless, as we subsequently show, the ideas which support teletransportation can be implemented integrating XCS with a model of the environment, learned by experience, in a dyna architecture. We end of the paper discussing another important aspect of generalization in XCS: the conditions under which XCS may fail to produce a compact representation of a learned task. We show this is likely to happen in environments where there is no direct relation between the number of don't care symbols a classifier condition has, and the number of environmental conditions the classifier matches. Accordingly, we discuss the role of subsumption deletion for the problem of evolving a compact representation of the learned task.

[554]
Pier Luca Lanzi. A Study of the Generalization Capabilities of XCS. In Bäck [24], pages 418-425. http://ftp.elet.polimi.it/people/lanzi/icga97.ps.gz.
We analyze the generalization behavior of the XCS classifier system in environments in which only a few generalizations can be done. Experimental results presented in the paper evidence that the generalization mechanism of XCS can prevent it from learning even simple tasks in such environments. We present a new operator, named Specify, which contributes to the solution of this problem. XCS with the Specify operator, named XCSS, is compared to XCS in terms of performance and generalization capabilities in different types of environments. Experimental results show that XCSS can deal with a greater variety of environments and that it is more robust than XCS with respect to population size.

[555]
Pier Luca Lanzi. Solving Problems in Partially Observable Environments with Classifier Systems (Experiments on Adding Memory to XCS). Technical Report 97.45, Politecnico di Milano. Department of Electronic Engineering and Information Sciences, 1997. http://ftp.elet.polimi.it/people/lanzi/report45.ps.gz.
XCS is a classifier system recently introduced by Wilson that differs from Holland's framework in that classifier fitness is based on the accuracy of the prediction instead of the prediction itself. According to the original proposal, XCS has no internal message list as traditional classifier systems does; hence XCS learns only reactive input/output mappings that are optimal in Markovian environments. When the environment is partially observable, i.e. non-Markovian, XCS evolves suboptimal solutions; in order to evolve an optimal policy in such environments the system needs some sort of internal memory mechanism. In this paper, we add internal memory mechanism to the XCS classifier system. We then test XCS with internal memory, named XCSM, in non-Markovian environments of increasing difficulty. Experimental results, we present, show that XCSM is able to evolve optimal solutions in simple environments, while in more complex problems the system needs special operators or special exploration strategies. We show also that the performance of XCSM is very stable with respect to the size of the internal memory involved in learning. Accordingly, when complex non-Markovian environments are faced XCSM performance results to be more stable when more bits than necessary are employed. Finally, we extend some of the results presented in the literature for classifier system in non-Markovian problems, applying XCSM to environments which require the agent to perform sequences of actions in the internal memory. The results presented suggest that the exploration strategies currently employed in the study of XCS are too simple to be employed with XCSM; accordingly, other exploration strategies should be investigated in order to develop better classifier systems

[556]
Pier Luca Lanzi. Adding Memory to XCS. In Proceedings of the IEEE Conference on Evolutionary Computation (ICEC98). IEEE Press, 1998. http://ftp.elet.polimi.it/people/lanzi/icec98.ps.gz.
We add internal memory to the XCS classifier system. We then test XCS with internal memory, named XCSM, in non-Markovian environments with two and four aliasing states. Experimental results show that XCSM can easily converge to optimal solutions in simple environments; moreover, XCSM's performance is very stable with respect to the size of the internal memory involved in learning. However, the results we present evidence that in more complex non-Markovian environments, XCSM may fail to evolve an optimal solution. Our results suggest that this happens because, the exploration strategies current employed with XCS, are not adequate to guarantee the convergence to an optimal policy with XCSM, in complex non-Markovian environments.

[557]
Pier Luca Lanzi. An analysis of the memory mechanism of XCSM. In Koza et al. [529], pages 643-651. http://ftp.elet.polimi.it/people/lanzi/gp98.ps.gz.
We analyze the memory mechanism of XCSM, the extension of XCS with internal memory. Our aim is to explain some of the results reported in the literature, which show that XCSM fails to learn an optimal policy in complex partially observable environments. The analysis we present reveals that the XCSM's memory management strategy cannot guarantee the convergence to an optimal solution. We thus extend XCSM introducing a novel hierarchical exploration technique and modifying the technique used for updating internal memory. We apply the novel version of XCSM, called XCSMH, to a set of partially observable environments of different complexity. Our results show that XCSMH is able to learn an optimal policy in all the environments, outperforming XCSMH in more difficult problems.

[558]
Pier Luca Lanzi. Generalization in Wilson's XCS. In A. E. Eiben, T. Bäck, M. Shoenauer, and H.-P. Schwefel, editors, Proceedings of the Fifth International Conference on Parallel Problem Solving From Nature -- PPSN V, number 1498 in LNCS. Springer Verlag, 1998.
We analyze generalization with the XCS classifier system when the system is applied to animat problems in grid-worlds. The aim of the paper is to give an unified view of generalization with XCS, in order to explain some of the phenomena reported in the literature. First, we extend the results previously presented in the literature applying XCS to two environments. Our results confirm what already reported in the literature, showing that the generalization mechanism of XCS may prevent the system from converging to optimal performance. Accordingly, we study XCS generalization mechanism analyzing the conditions under which it may fail to evolve an optimal solution. We draw a hypothesis in order to explain the results reported so far. Our hypothesis suggest that XCS fails to learn an optimal solution when, due to the environment structure and to the exploration strategy employed, the system is not able to explore all the environmental niches frequently. We test our hypothesis introducing a new exploration strategy which guarantees a frequent exploration of all the areas of the environment, we call it teletransportation. We then apply XCS with teletransportation the environments previously introduced in order to validate our hypothesis experimentally.

[559]
Pier Luca Lanzi. Reinforcement Learning by Learning Classifier Systems. PhD thesis, Politecnico di Milano, 1998.
This thesis investigates a learning paradigm which merge together the concept of learning by interactions, which comes from the research field of reinforcement learning, with the concept of learning through evolution, which was introduced by John H. Holland, the father of genetic algorithms, in 1978 who called this paradigm learning classifier systems. This work is devoted to the study of a particular model of learning classifier system called XCS, introduced by Stewart W. Wilson in 1995, in which the most interesting ideas of Holland's paradigm co-exist with more recent ideas borrowed from reinforcement learning. The aim of this thesis is to take XCS beyond the very first results presented by Wilson in his original paper and to investigate the peculiarities of this new model of learning classifier systems with respect to other learning paradigms that come from reinforcement learning and evolutionary computation. In doing this, I propose a number of extensions to the original Wilson's framework that improves its performance and its applicability to a larger number of problems. Initially, I take XCS beyond its very first environments and parameter settings, and I show that in certain difficult sequential environments, XCS performance can fail dramatically. I propose an extension to XCS that improve its performance in such types of applications. I present an experimental analysis of generalization in XCS to understand the failures in XCS performance I observed, and to justify the improvements that the extension I propose introduces. Then, I introduce an extension to XCS for dealing with the possible incomplete perceptions that the system may experience in certain applications. This is done in two steps. First, following an idea suggested by Wilson, I implement a simple extension to XCS for dealing with incomplete perception. I show that this is capable to learn optimal behaviors in particular cases but not in general ones. I analyze this result and present an experimental study which points out the reasons underlying the poor performance of this extension. I suggest how the system should be extended in order to evolve optimal solutions in a wider set of applications. I implement this new extension of XCS and present experimental results showing that indeed the new system can learn in more difficult applications. Finally, I analyze XCS behavior in those applications where the results of the agent actions can be affected by uncertainty. I show that Wilson's XCS can deal with these types of applications when the uncertainty on agent action is limited, other-wise the system may fail to converge to an optimal performance. I study this phenomenon and develop an explanation which results in an extension of XCS that is able to deal with higher degree of uncertainty. I end the thesis presenting some initial results of the research directions I am currently pursuing.

[560]
Pier Luca Lanzi. An Analysis of Generalization in the XCS Classifier System. Evolutionary Computation, 7(2):125-149, 1999.
The XCS classifier system represents a major advance in classifier systems research because (1) it has a sound and accurate generalization mechanism, and (2) its learning mechanism is based on Q-learning, a recognized learning technique. In taking XCS beyond its very first environments and parameter settings, we show that, in certain difficult sequential (``animat'') environments, performance is poor. We suggest that this occurs because in the chosen environments, some conditions for proper functioning of the generalization mechanism do not hold, resulting in overly general classifiers that cause reduced performance. We hypothesize that one such condition is a lack of sufficiently wide exploration of the environment during learning. We show that if XCS is forced to explore its environment more completely, performance improves dramatically. We propose a technique based on Sutton's Dyna concept, through which wider exploration would occur naturally. Separately, we demonstrate that the compactness of the representation evolved by XCS is limited by the number of instances of each generalization actually present in the environment. The paper shows that XCS's generalization mechanism is effective, but that the conditions under which it works must be clearly understood.

[561]
Pier Luca Lanzi. Extending the Representation of Classifier Conditions Part I: From Binary to Messy Coding. In Banzhaf et al. [32], pages 337-344.
We present an initial study on the alternative representations of classifier conditions. Instead of considering the representations that have been suggested in the literature, S-expressions or real-coding, we focus on messy representation of classifier conditions which has the interesting characteristic of being independent from the sensors position coding scheme. We develop an extension of the XCS classifier system in which variable-length messy chromosomes replace original bitstring representation. With a series of experiments we show that to reach optimal performance covering, matching, and mutation must be adequately defined in order to avoid overgeneralization due to underspecification of classifier conditions. These conclusions appear to be very general since they are independent of the messy representation scheme. Accordingly, they can be used as guidelines for general variable-length representations like S-Expressions.

[562]
Pier Luca Lanzi. Extending the Representation of Classifier Conditions Part II: From Messy Coding to S-Expressions. In Banzhaf et al. [32], pages 345-352.
In this paper we present the results of the second part of our research which is aimed to the study of alternative representations of classifier conditions. In particular we introduce an extension of the XCS classifier system in which bitstring representation is replaced by S-expressions. We show that XCS with S-expressions, SXCS, can reach optimal performance in different types of applications with different complexity. The results we present also suggest that great care must be taken in choosing the representation language; in particular we show that in certain cases the use of ``or'' clauses may lead to an unstable performance. Overall, our initial results show that this is a promising approach for a future development of a general purpose representation of classifier conditions and that there are still many issues which shall be investigated.

[563]
Pier Luca Lanzi. Adaptive Agents with Reinforcement Learning and Internal Memory. In Sixth International Conference on the Simulation of Adaptive Behavior (SAB2000), pages 333-342, 2000.
Perceptual aliasing is a serious problem for adaptive agents. Internal memory is a promising approach to extend reinforcement learning algorithms to problems involving perceptual aliasing. In this paper we investigate the effectiveness of internal memory for tackling perceptual aliasing problems with adaptive agents and reinforcement learning. Specifically, we try to give a unified view of some interesting results that have been presented in different frameworks, i.e.: tabular reinforcement learning and learning classifier systems.

[564]
Pier Luca Lanzi. Learning Classifier Systems from a Reinforcement Learning Perspective . Technical Report 00-03, Dipartimento di Elettronica e Informazione, Politecnico di Milano, 2000.
In this paper we approach the problem of defining learning classifier systems from the perspective of reinforcement learning. The peculiarity of this approach is that it does not assume any knowledge on learning classifier systems but try to develop classifier systems ``from scratch'', i.e., starting from one of the most known reinforcement learning technique: Q-learning. We begin considering some basic elements of reinforcement learning: a problem modelled as a Markov Decision Process and tabular Q-learning. We introduce a formal framework to define a general purpose rule-based representation which we use to implement tabular Q-learning. We analyze different methods to add generalization capabilities to the rule-based representation. We argue that genetic algorithms are probably the most general method for adding generalization to classifiers, although they might be not the only solution.

[565]
Pier Luca Lanzi. Mining interesting knowledge from data with the xcs classifier system. In Lee Spector, Erik D. Goodman, Annie Wu, W.B. Langdon, Hans-Michael Voigt, Mitsuo Gen, Sandip Sen, Marco Dorigo, Shahram Pezeshk, Max H. Garzon, and Edmund Burke, editors, Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2001), pages 958-965, San Francisco, California, USA, 7-11 July 2001. Morgan Kaufmann.

[566]
Pier Luca Lanzi. Learning classifier systems from a reinforcement learning perspective. Journal of Soft Computing, 6(3-4):162-170, 2002.

[567]
Pier Luca Lanzi. Estimating classifier generalization and action's effect: A minimalist approach. In E. Cantú-Paz, J. A. Foster, K. Deb, D. Davis, R. Roy, U.-M. O'Reilly, H.-G. Beyer, R. Standish, G. Kendall, S. Wilson, M. Harman, J. Wegener, D. Dasgupta, M. A. Potter, A. C. Schultz, K. Dowsland, N. Jonoska, and J. Miller, editors, Genetic and Evolutionary Computation -- GECCO-2003, volume 2724 of LNCS, pages 1894-1905. Springer-Verlag, 2003.

[568]
Pier Luca Lanzi. Using raw accuracy to estimate classifier fitness in XCS. In E. Cantú-Paz, J. A. Foster, K. Deb, D. Davis, R. Roy, U.-M. O'Reilly, H.-G. Beyer, R. Standish, G. Kendall, S. Wilson, M. Harman, J. Wegener, D. Dasgupta, M. A. Potter, A. C. Schultz, K. Dowsland, N. Jonoska, and J. Miller, editors, Genetic and Evolutionary Computation -- GECCO-2003, volume 2724 of LNCS, pages 1922-1923. Springer-Verlag, 2003.

[569]
Pier Luca Lanzi. Learning classifier systems: then and now. Evolutionary Intelligence, 1(1):63-82, 2008.

[570]
Larry Bull, David Wyatt and Ian Parmee. Initial modifications to XCS for use in interactive evolutionary design. In H.-P. Schwefel J.-J. Merelo Guervós, P. Adamidis, H.-G. Beyer, J.-L. Fernández-Villacañas, editor, Parallel Problem Solving from Nature - PPSN VII, 7th International Conference, Granada, Spain, September 7-11, 2002. Proceedings, number 2439 in Lecture Notes in Computer Science, LNCS, page 568 ff. Springer-Verlag, 2002.

[571]
Claude Lattaud. Non-Homogeneous Classifier Systems in a Macro-Evolution Process. In Wu [923], pages 266-271.
The synthesis of artifacts reproducing behaviors and properties of living beings is one of the main goals of Artificial Life. These artificial entities often evolve according to algorithms based on models of modern genetics. Evolutionary algorithms generally produce micro-evolution in these entities, by mutation and crossover applied on their genome. The aim of this paper is to present Non-Homogeneous Classifier Systems, NHCS, integrating the process of macroevolution. A NHCS is a derived type of classical classifier systems, CS. In a standard CS, all classifiers are built on the same structure and own the same properties. With a NHCS, the behavior of an artificial creature is defined by the co-evolution between several differently structured classifiers making its organism. These agents, moving in a 2D discreet environment with obstacles and resources, must adapt themselves and breed to build viable populations. Finally, ecological niches and particular behaviors, individual and collective, appear according to initial parameters of agents and environment.

[572]
Claude Lattaud. Non-Homogeneous Classifier Systems in a Macro-Evolution Process. In Lanzi et al. [544], pages 161-174.
The synthesis of artifacts reproducing behaviors and properties of living beings is one of the main goals of Artificial Life. These artificial entities often evolve according to algorithms based on models of modern genetics. Evolutionary algorithms generally produce micro-evolution in these entities, by applying mutation and crossover on their genotype. The aim of this paper is to present Non-Homogenous Classifier Systems, NHCS, integrating the process of macro-evolution. A NHCS is a derived type of classical Classifier Systems, CS. In a CS, all classifiers are built on the same structure and own the same properties. With a NHCS, the behavior of artificial creatures is defined by the co-evolution between several differently structured classifiers. These agents, moving in a 2D environment with obstacles and resources, must adapt themselves and breed to build viable populations. Finally, ecological niches and specific behaviors, individual and collective, appear according to initial parameters of agents and environment.

[573]
Blake Lebaron, W. Brian Arthur, and R. Palmer. The Time Series Properties of an Artificial Stock Market. Journal of Economic Dynamics and Control, 23:1487-1516, 1999.

[574]
Martin Lettau and Harald Uhlig. Rules of Thumb and Dynamic Programming. Technical report, Department of Economics, Princeton University, 1994.

[575]
Martin Lettau and Harald Uhlig. Rules of thumb versus dynamic programming. American Economic Review, 89:148-174, 1999.
This paper studies decision-making with rules of thumb in the context of dynamic decision problems and compares it to dynamic programming. A rule is a fixed mapping from a subset of states into actions. Rules are compared by averaging over past experiences. This can lead to favoring rules which are only applicable in good states. Correcting this good state bias requires solving the dynamic program. We provide a general framework and characterize the asymptotic properties. We apply it to provide a candidate explanation for the sensitivity of consumption to transitory income.

[576]
Pen-Yang Liao and Jiah-Shing Chen. Dynamic trading strategey learning model using learning classifier systems. In Proceedings of the 2001 Congress on Evolutionary Computation (CEC01) [186], pages 783-789.
Current trading strategy learning models often proceed in three separate phases, i.e., training, validation, and application (testing). After a specific time span of application, a new learning process is started to adapt the trading strategy to the new environment states. The time span of application is usually fixed and determined according to experiences. This may result ni earning losses as compared to the perfect trading strategy which trades at each turning point of the stock price movement. Some learning methods, such as neural networks, are hard to explain intuitively and unstable in some dynamic environment states. Other learning models like simple genetic algorithms result in a single trading rule which is applied for a specific time span without being adapted even when the environment has changed. This paper adopts learning classifier systems (LCSs) technique to provide a dynamic trading strategy learning model (DTSLM), which makes continuous and instant learning while executing real prediction and prodcues a trading rule set to deal with different environment states. The simulation results show that this model could get a remarkable profit.

[577]
Gunar E. Liepins and Lori A. Wang. Classifier System Learning of Boolean Concepts. In Booker and Belew [74], pages 318-323.
We investigate classifier system learning of Boolean concepts. We introduce a symmetric reward-penalty mechanism, speciation, generality thresholds and rule evaluation by queries. These enable the classifier system to learn the twenty multiplexor significantly faster than previously reported for classifier systems. Conversely, we provide theoretical analyses that suggest that classifier systems are not competitive with the best known learning algorithms for stationary deterministic Boolean problems. We suggest instead that they are particularly well suited to non-stationary problems for which the target concept evolves over time.

[578]
Gunar E. Liepins, Michael R. Hilliard, Mark Palmer, and Gita Rangarajan. Alternatives for Classifier System Credit Assignment. In Proceedings of the Eleventh International Joint Conference on Artificial Intelligence (IJCAI-89), pages 756-761, 1989.

[579]
Gunar E. Liepins, M. R. Hillard, M. Palmer, and G. Rangarajan. Credit Assignment and Discovery in Classifier Systems. International Journal of Intelligent Systems, 6:55-69, 1991.

[580]
Derek A. Linkens and H. Okola Nyongesah. Genetic Algorithms for fuzzy control - Part II: Off-line system development and application. Technical Report CTA/94/2387/1st MS, Department of Automatic Control and System Engineering, University of Sheffield, U.K., 1994.
Although fuzzy logic controllers and expert systems have been successfully applied in many complex industrial processes, they experience a deficiency in knowledge acquisition and rely to a great extent on empirical and heuristic knowledge, which in many cases cannot be objectively elicited. Among the problems to be resolved in fuzzy controller design are the determination of the linguistic state space, definition of the membership functions of each linguistic term and the derivation of the control rules. Some of these problems can be solved by application of machine learning. First, it is desirable to simplify and automate the specification of linguistic rules. Secondly, it is also desirable that modification of control rules be possible in order to cope with previously unknown, or changes in process dynamics. Machine learning methods have in recent years emerged from the use of learning algorithms modelled on natural and biological systems. These methods attempt to abstract the advanced mechanisms of learning exhibited by such systems, which can, consequently, be applied to intelligent control. One of these new algorithms is the genetic algorithm which is modelled on the processes of natural evolution. This paper develops the application of genetic algorithm techniques for fuzzy controller design. Genetic algorithms are used to automate and introduce objective criteria in defining fuzzy controller parameters.

[581]
Juliet Juan Liu and James Tin-Yau Kwok. An extended genetic rule induction algorithm. In Proceedings of the 2000 Congress on Evolutionary Computation (CEC00) [185], pages 458-463.
This paper describes an extension of a GA-based, separate-and-conquer propositional rule incuction algorithm called SIA. While the original algorithm is computationally attractive and is also able to handle both nominal and continuous attributes efficiently, our algorithm further improves it by taking into account of the recent advances in the rule induction and evolutionary computation communities. The refined system has been compared to other GA-based and non GA-based rule learning algorithms on a number of benchmark datasets from the UCI machine learning Repository. Results show that the proposed system can achieve higher performance while still produces a smaller number of rules.

[582]
Xavier Llorà and Josep M. Garrell. Automatic Classification and Artificial Life Models. In Proceedings of the International Worshop on Learning (Learning00), 2000.
ALE is a classication-oriented model based on cooperative agent aggregates spread over a two dimensional board. The model solves classi cation problems descrived by continuous-valued imputs. This paper describes ALE focussing on a key point, resources allocation. The main apportation is that using an accurate resources allocation we can clearly improve convergence speed at the same time that agents aggregate complexity reduces

[583]
Xavier Llorà and Josep M. Garrell. Evolution of Decision Trees. In Forth Catalan Conference on Artificial Intelligence (CCIA'2001), page to appear. ACIA Press, 2001.

[584]
Xavier Llorà and Josep M. Garrell. Evolving Partially-Defined Instances with Evolutionary Algorithms. In Proceedings of the 18th International Conference on Machine Learning (ICML'2001), pages 337-344. Morgan Kaufmann Publishers, 2001.

[585]
Xavier Llorà and Josep M. Garrell. Knowledge-Independent Data Mining with Fine-Grained Parallel Evolutionary Algorithms. In Lee Spector, Erik D. Goodman, Annie Wu, W.B. Langdon, Hans-Michael Voigt, Mitsuo Gen, Sandip Sen, Marco Dorigo, Shahram Pezeshk, Max H. Garzon, and Edmund Burke, editors, Proceedings of the Genetic and Evolutionary Computation Conference (GECCO'2001), pages 461-468. Morgan Kaufmann Publishers, 2001.

[586]
Xavier Llorà and Josep M. Garrell. Coevolving different knowledge representations with fine-grained parallel learning classifier systems. In W. B. Langdon, E. Cantú-Paz, K. Mathias, R. Roy, D. Davis, R. Poli, K. Balakrishnan, V. Honavar, G. Rudolph, J. Wegener, L. Bull, M. A. Potter, A. C. Schultz, J. F. Miller, E. Burke, and N. Jonoska, editors, GECCO 2002: Proceedings of the Genetic and Evolutionary Computation Conference, pages 934-941. Morgan Kaufmann Publishers, 2002.

[587]
Xavier Llorà and Stewart W. Wilson. Mixed Decision Trees: Minimizing Knowledge Representation Bias in LCS. In Kalyanmoy Deb et al., editor, Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2004), volume 3103 of Lecture Notes in Computer Science, pages 797-809. Springer, 2004.
Learning classifier systems tend to inherit -a priori- a given knowledge representation language for expressing the concepts to learn. Hence, even before getting started, this choice biases what can be learned, becoming critical for some real-world applications like data mining. However, such bias may be minimized by hybridizing different knowledge representations via evolutionary mixing. This paper presents a first attempt to produce an evolutionary framework that evolves mixed decision trees of heterogeneous knowledge representations.

[588]
Xavier Llorà, K. Sastry, and D.E. Goldberg. Binary rule encoding schemes: a study using the compact classifier system. In F. Rothlauf, editor, GECCO '05: Proceedings of the 2005 conference on genetic and evolutionary computation, workshop proceedings, pages 88-89. ACM Press, 2005.

[589]
Xavier Llorà, K. Sastry, and D.E. Goldberg. The compact classifier system: scalability analysis and first results. In F. Rothlauf, editor, Proceedings of the IEEE congress on evolutionary computation, CEC 2005, pages 596-603. IEEE, 2005.

[590]
Xavier Llorà, Kumara Sastry, Cláudio F. Lima, Fernando G. Lobo, and David E. Goldberg. Linkage learning, rule representation, and the χ-ary extended compact classifier system. In Jaume Bacardit, Ester Bernadó-Mansilla, Martin Butz, Tim Kovacs, Xavier Llorà, and Keiki Takadama, editors, Learning Classifier Systems. 10th and 11th International Workshops (2006-2007), volume 4998/2008 of Lecture Notes in Computer Science, pages 189-205. Springer, 2008.
This paper reviews a competent Pittsburgh LCS that automatically mines important substructures of the underlying problems and takes problems that were intractable with first-generation Pittsburgh LCS and renders them tractable. Specifically, we propose a χ-ary extended compact classifier system (χeCCS) which uses (1) a competent genetic algorithm (GA) in the form of χ-ary extended compact genetic algorithm, and (2) a niching method in the form restricted tournament replacement, to evolve a set of maximally accurate and maximally general rules. Besides showing that linkage exist on the multiplexer problem, and that χeCCS scales exponentially with the number of address bits (building block size) and quadratically with the problem size, this paper also explores non-traditional rule encodings. Gene expression encodings, such as the Karva language, can also be used to build χeCCS probabilistic models. However, results show that the traditional ternary encoding 0,1,# presents a better scalability than the gene expression inspired ones for problems requiring binary conditions

[591]
Xavier Llorà. Genetic Based Machine Learning using Fine-grained Parallelism for Data Mining. PhD thesis, Enginyeria i Arquitectura La Salle. Ramon Llull University, 2002.

[592]
D. Loiacono, A. Marelli, and P.L. Lanzi. Support vector regression for classifier prediction. In GECCO '07: Proceedings of the 9th annual conference on Genetic and evolutionary computation, pages 1806-1813. ACM, 2007.
In this paper we introduce XCSF with support vector prediction:the problem of learning the prediction function is solved as a support vector regression problem and each classifier exploits a Support Vector Machine to compute the prediction. In XCSF with support vector prediction, XCSFsvm, the genetic algorithm adapts classifier conditions, classifier actions, and the SVM kernel parameters.We compare XCSF with support vector prediction to XCSF with linear prediction on the approximation of four test functions.Our results suggest that XCSF with support vector prediction compared to XCSF with linear prediction (i) is able to evolve accurate approximations of more difficult functions, (ii) has better generalization capabilities and (iii) learns faster.

[593]
Daniele Loiacono, Jan Drugowitsch, Alwyn Barry, and Pier Luca Lanzi. Analysis and improvements of the classifier error estimate in xcsf. In Jaume Bacardit, Ester Bernadó-Mansilla, Martin Butz, Tim Kovacs, Xavier Llorà, and Keiki Takadama, editors, Learning Classifier Systems. 10th and 11th International Workshops (2006-2007), volume 4998/2008 of Lecture Notes in Computer Science, pages 117-135. Springer, 2008.
The estimation of the classifier error plays a key role in accuracy-based learning classifier systems. In this paper we study the current definition of the classifier error in XCSF and discuss the limitations of the algorithm that is currently used to compute the classifier error estimate from online experience. Subsequently, we introduce a new definition for the classifier error and apply the Bayes Linear Analysis framework to find a more accurate and reliable error estimate. This results in two incremental error estimate update algorithms that we compare empirically to the performance of the currently applied approach. Our results suggest that the new estimation algorithms can improve the generalization capabilities of XCSF, especially when the action-set subsumption operator is used.

[594]
Sean Luke. Essentials of Metaheuristics. 2009.

[595]
Pattie Maes, Maja J. Mataric, Jean-Arcady Meyer, Jordan Pollack, and Stewart W. Wilson, editors. From Animals to Animats 4. Proceedings of the Fourth International Conference on Simulation of Adaptive Behavior (SAB96). A Bradford Book. MIT Press, 1996.

[596]
Chikara Maezawa and Masayasu Atsumi. Collaborative Learning Agents with Structural Classifier Systems. In Banzhaf et al. [32], page 777. One page poster paper.
We propose a new learning agent architecture for collaborative learning. To learn any complicated task in multi-agent environment, simple reinforcement architectures have limitations on learning. Therefore, we propose splitting learning mechanism into three separate layers to learn required behavior, which are respectively organized by the Classifier System [Holland86]. It can learn to communicate with other agents, to make plans, and to select actions based on the plans and other agents' behavior. We show that these agents can select cooperative actions as a collaborative group.

[597]
Bernard Manderick. Selectionist Categorization. In Schwefel and Männer [728], pages 326-330.
A selectionist recognition system is presented that categorizes inputs generation by pattern generators.

[598]
Ester Bernadó I Mansilla and Josep Maria Garrell i Guiu. MOLeCS: A MultiObjective Learning Classifier System. In Whitely et al. [878], page 390. One page poster paper.
This work has no abstract.

[599]
Ramon Marimon, Ellen McGrattan, and Thomas J. Sargent. Money as a Medium of Exchange in an Economy with Artificially Intelligent Agents. Journal of Economic Dynamics and Control, 14:329-373, 1990. Also Technical Report 89-004, Santa Fe Institute, 1989.

[600]
James A. R. Marshall and Tim Kovacs. A representational ecology for learning classifier systems. In Maarten Keijzer et al., editor, Proceedings of the 2006 Genetic and Evolutionary Computation Conference (GECCO 2006), pages 1529-1536. ACM, 2006.

[601]
James A. R. Marshall, Gavin Brown, and Tim Kovacs. Bayesian Estimation of Rule Accuracy in UCS. In Tina Yu, editor, To appear in the proceedings of the 2007 workshops on genetic and evolutionary computation. ACM, 2007.

[602]
Javier G. Marín-Blázquez and Sonia Schulenburg. A hyper-heuristic framework with xcs: Learning to create novel problem-solving algorithms constructed from simpler algorithmic ingredients. In Tim Kovacs, Xavier LLòra, Keiki Takadama, Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson, editors, Learning Classifier Systems. International Workshops, IWLCS 2003-2005, Revised Selected Papers, volume 4399 of LNCS, pages 193-218. Springer, 2007.

[603]
Maja J. Mataric. A comparative analysis of reinforcement learning methods. A.I. Memo No. 1322, Massachusetts Institute of Technology, 1991.
This paper analyzes the suitability of reinforcement learning for both programming and adapting situated agents. In the the first part of the paper we discuss two specific reinforcement learning algorithms: Q-learning and the Bucket Brigade. We introduce a special case of the Bucket Brigade, and analyze and compare its performance to Q- learning in a number of experiments. The second part of the paper discusses the key problems of reinforcement learning: time and space complexity, input generalization, sensitivity to parameter values, and selection of the reinforcement function. We address the tradeoff between the amount of built in and learned knowledge in the context of the number of training examples required by a learning algorithm. Finally, we suggest directions for future research.

[604]
Alaster D. McAulay and Jae Chan Oh. Image Learning Classifier System Using Genetic Algorithms. In Proceedings IEEE NAECON '89, 1989.

[605]
John R. McDonnell. Control. In Thomas Bäck, David B. Fogel, and Zbigniew Michalewicz, editors, Handbook of Evolutionary Computation, pages F1.3:1-F1.3:7. IOP Publishing Ltd and Oxford University Press, 1997.

[606]
Chris Melhuish and Terence C. Fogarty. Applying A Restricted Mating Policy To Determine State Space Niches Using Immediate and Delayed Reinforcement. In Fogarty [324], pages 224-237.
Approaches for rule based control often rely heavily on the pre-classification of the state space for their success. In the pre-determined regions individual or groups of rules may be learned. Clearly, the success of such strategies depends on the quality of the partitioning of the state space. When no such a priori partitioning is available it is a significantly more difficult task to learn an appropriate division of the state space as well as the associated rules. Yet another layer of potential difficulty is the nature of the reinforcement applied to the rules since it is not always possible to generate an immediate reinforcement signal to supply judgement on the efficacy of activated rules. One approach to combine the joint goals of determining partitioning of the state space and discovery of associated rules is to use a genetic algorithm employing a restricted mating policy to generate rule clusters which dominate regions of the state space thereby effecting the required partitioning. Such rule clusters are termed niches. A niching algorithm, which includes a `creche' facility to protect `inexperienced' classifiers, and the results of determining a simple 2D state space using an immediate reward scheme are presented. Details of how the algorithm may modified to incorporate a delayed reinforcement scheme on a real-world beam balancing control problem are reported.

[607]
Drew Mellor. A first order logic classifier system. In F. Rothlauf, editor, GECCO '05: Proceedings of the 2005 conference on genetic and evolutionary computation, pages 1819-1826. ACM Press, 2005.

[608]
Drew Mellor. Policy transfer with a relational learning classifier system. In GECCO Workshops 2005, pages 82-84. ACM Press, 2005.

[609]
Drew Mellor. A learning classifier system approach to relational reinforcement learning. In Jaume Bacardit, Ester Bernadó-Mansilla, Martin Butz, Tim Kovacs, Xavier Llorà, and Keiki Takadama, editors, Learning Classifier Systems. 10th and 11th International Workshops (2006-2007), volume 4998/2008 of Lecture Notes in Computer Science, pages 169-188. Springer, 2008.
This article describes a learning classifier system (LCS) approach to relational reinforcement learning (RRL). The system, Foxcs-2, is a derivative of XCS that learns rules expressed as definite clauses over first-order logic. By adopting the LCS approach, Foxcs-2, unlike many RRL systems, is a general, model-free and “tabula rasa” system. The change in representation from bit-strings in XCS to first-order logic in Foxcs-2 necessitates modifications, described within, to support matching, covering, mutation and several other functions. Evaluation on inductive logic programming (ILP) and RRL tasks shows that the performance of Foxcs-2 is comparable to other systems. Further evaluation on RRL tasks highlights a significant advantage of Foxcs-2’s rule language: in some environments it is able to represent policies that are genuinely scalable; that is, policies that are independent of the size of the environment.

[610]
Drew Mellor. A learning classifier system approach to relational reinforcement learning. PhD thesis, University of Newcastle, 2008.
Machine learning methods usually represent knowledge and hypotheses using attribute-value languages, principally because of their simplicity and demonstrated utility over a broad variety of problems. However, attribute-value languages have limited expressive power and for some problems the target function can only be expressed as an exhaustive conjunction of specific cases. Such problems are handled better with inductive logic programming (ILP) or relational reinforcement learning (RRL), which employ more expressive languages, typically languages over first-order logic. Methods developed within these fields generally extend upon attribute-value algorithms; however, many attribute-value algorithms that are potentially viable for RRL, the younger of the two fields, remain to be extended. This thesis investigates an approach to RRL derived from the learning classifier system XCS. In brief, the new system, FOXCS, generates, evaluates, and evolves a population of ``condition-action'' rules that are definite clauses over first-order logic. The rules are typically comprehensible enough to be understood by humans and can be inspected to determine the acquired principles. Key properties of FOXCS, which are inherited from XCS, are that it is general (applies to arbitrary Markov decision processes), model-free (rewards and state transitions are ``black box'' functions), and ``tabula rasa'' (the initial policy can be unspecified). Furthermore, in contrast to decision tree learning, its rule-based approach is ideal for incrementally learning expressions over first-order logic, a valuable characteristic for an RRL system. Perhaps the most novel aspect of FOXCS is its inductive component, which synthesizes evolutionary computation and first-order logic refinement for incremental learning. New evolutionary operators were developed because previous combinations of evolutionary computation and first-order logic were non-incremental. The effectiveness of the inductive component was empirically demonstrated by benchmarking on ILP tasks, which found that FOXCS produced hypotheses of comparable accuracy to several well-known ILP algorithms. Further benchmarking on RRL tasks found that the optimality of the policies learnt were at least comparable to those of existing RRL systems. Finally, a significant advantage of its use of variables in rules was demonstrated: unlike RRL systems that did not use variables, FOXCS, with appropriate extensions, learnt scalable policies that were genuinely independent of the dimensionality of the task environment.

[611]
Drew Mellor. Reinforcement Learning, Logic and Evolutionary Computation: A Learning Classifier System Approach to Relational Reinforcement Learning. Lambert Academic Publishing, 2009.
Reinforcement learning (RL) consists of methods that automatically adjust behaviour based on numerical rewards and penalties. While use of the attribute-value framework is widespread in RL, it has limited expressive power. Logic languages, such as first-order logic, provide a more expressive framework, and their use in RL has led to the field of relational RL. This thesis develops a system for relational RL based on learning classifier systems (LCS). In brief, the system generates, evolves, and evaluates a population of condition-action rules, which take the form of definite clauses over first-order logic. Adopting the LCS approach allows the resulting system to integrate several desirable qualities: model-free and "tabula rasa" learning; a Markov Decision Process problem model; and importantly, support for variables as a principal mechanism for generalisation. The utility of variables is demonstrated by the system's ability to learn genuinely scalable behaviour - behaviour learnt in small environments that translates to arbitrarily large versions of the environment without the need for retraining.

[612]
Marc Metivier and Claud Lattaud. Imitation guided learning in learning classifier systems. Natural Computing, 8(1):29-56, 2009.
In this paper, we study the means of developing an imitation process allowing to improve learning in the framework of learning classifier systems. We present three different approaches in the way a behavior observed may be taken into account through a guidance interaction: two approaches using a model of this behavior, and one without modelling. Those approaches are evaluated and compared in different environments when they are applied to three major classifier systems: ZCS, XCS and ACS. Results are analyzed and discussed. They highlight the importance of using a model of the observed behavior to enable an efficient imitation. Moreover, they show the advantages of taking this model into account by a specialized internal action. Finally, they bring new results of comparison between ZCS, XCS and ACS.

[613]
J. A. Meyer and S. W. Wilson, editors. From Animals to Animats 1. Proceedings of the First International Conference on Simulation of Adaptive Behavior (SAB90). A Bradford Book. MIT Press, 1990.

[614]
Zbigniew Michalewicz. Genetic Algorithms + Data Structures = Evolution Programs. Springer-Verlag, 1996. Contains introductory chapter on LCS.

[615]
John H. Miller and Stephanie Forrest. The dynamical behavior of classifier systems. In Schaffer [718], pages 304-310.
A methodology is described for studying the dynamical behavior of classifier systems. The methodology is useful because of the current lack of analytical results describing interactions among the various components of classifier systems. A mapping is defined between classifier systems and an equivalent dynamical system (Boolean networks). The mapping provides a way to understand and predict classifier system behaviors by observing the dynamical behavior of the Boolean networks. The paper reports initial results produced by the methodology and discusses the implications of this approach for classifier systems.

[616]
M. Mitchell and S. Forrest. Genetic Algorithms and Artificial Life. Technical Report 93-11-072, Santa Fe Institute, 1993. Contains a 2 page review of work on LCS.
Genetic algorithms are computational models of evolution that play a central role in many artificial-life models. We review the history and current scope of research on genetic algorithms in artificial life, using illustrative examples in which the genetic algorithm is used to study how learning and evolution interact, and to model ecosystems, immune system, cognitive systems, and social systems. We also outline a number of open questions and future directions for genetic algorithms in artificial-life research.

[617]
Johann Mitlöhner. Classifier systems and economic modelling. In APL '96. Proceedings of the APL 96 Conference on Designing the Future, volume 26 (4), pages 77-86, 1996.
Human economic decisions are characterized by a number of factors which make them difficult to model with standard mathematical tools. Decisions can be more easily described by a set of rules, and some of them may be rules of thumb. Economic behaviour is adaptive, in that people are able to adjust to a changing environment. It is argued in this paper that the classifier system framework is a suitable means of modelling human economic decisions. A case of a simple economic decision of finding an optimal price is discussed, which is later made more complex by introducing an input variable that effects the optimal price. It is shown that classifier systems can be used in both tasks, and their performance is compared to human decisions in the same set of circumstances.

[618]
Chilukuri K. Mohan. Expert Systems: A Modern Overview. Kluwer, 2000. Contains an introductory survey chapter on LCS.

[619]
Renan C. Moioli, Patricia A. Vargas, and Fernando J. Von Zuben. Analysing learning classifier systems in reactive and non-reactive robotic tasks. In Jaume Bacardit, Ester Bernadó-Mansilla, Martin Butz, Tim Kovacs, Xavier Llorà, and Keiki Takadama, editors, Learning Classifier Systems. 10th and 11th International Workshops (2006-2007), volume 4998/2008 of Lecture Notes in Computer Science, pages 286-305. Springer, 2008.
There are few contributions to robot autonomous navigation applying Learning Classifier Systems (LCS) to date. The primary objective of this work is to analyse the performance of the strength-based LCS and the accuracy-based LCS, named EXtended Learning Classifier System (XCS), when applied to two distinct robotic tasks. The first task is purely reactive, which means that the action to be performed can rely only on the current status of the sensors. The second one is non-reactive, which means that the robot might use some kind of memory to be able to deal with aliasing states. This work presents a rule evolution analysis, giving examples of evolved populations and their peculiarities for both systems. A review of LCS derivatives in robotics is provided together with a discussion of the main findings and an outline of future investigations.

[620]
D. Montanari. Classifier systems with a constant-profile bucket brigade. In Collected Abstracts for the First International Workshop on Learning Classifier System (IWLCS-92) [486]. October 6-8, NASA Johnson Space Center, Houston, Texas.
Paper is an extended abstract

[621]
David E. Moriarty, Alan C. Schultz, and John J. Grefenstette. Evolutionary Algorithms for Reinforcement Learning. Journal of Artificial Intelligence Research, 11:199-229, 1999. http://www.ib3.gmu.edu/gref/papers/moriarty-jair99.html.
There are two distinct approaches to solving reinforcement learning problems, namely, searching in value function space and searching in policy space. Temporal difference methods and evolutionary algorithms are well-known examples of these approaches. Kaelbling, Littman and Moore recently provided an informative survey of temporal difference methods. This article focuses on the application of evolutionary algorithms to the reinforcement learning problem, emphasizing alternative policy representations, credit assignment methods, and problem-specific genetic operators. Strengths and weaknesses of the evolutionary approach to reinforcement learning are presented, along with a survey of representative applications.

[622]
Rémi Munos and Jocelyn Patinel. Reinforcement learning with dynamic covering of state-action space: Partitioning Q-learning. In Cliff et al. [202], pages 354-363.
This paper presents a reinforcement learning algorithm: ``Partitioning Q-learning'', designed for generating an adaptive behavior of a reactive system with local perception in a complex and changing environment. This algorithm includes two dynamics: the learning algorithm based on the Q-learning and Bucket Brigade algorithms, and the structural dynamics (the partitioning of regions of the state-action space) that modelizes the acquisition of expert knowledge. The combination of these two dynamics intends to solve the problem of the combinatory explosion of the number of qualities to be estimated (the generalization problem), by dividing the state-action space into a minimal number of homogeneous regions using the formalism of Classifier Systems. This algorithm is applied to the simulation of a reactive robot which tries to cut weeds and to avoid plants in a cultivated field.

[623]
Tadahiko Murata, Shuhei Kawakami, Hiroyuki Nozawa, Mitsuo Gen, and Hisao Ishibuchi. Three-objective genetic algorithms for designing compact fuzzy rule-based systems for pattern classification problems. In Lee Spector, Erik D. Goodman, Annie Wu, W.B. Langdon, Hans-Michael Voigt, Mitsuo Gen, Sandip Sen, Marco Dorigo, Shahram Pezeshk, Max H. Garzon, and Edmund Burke, editors, Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2001), pages 485-492, San Francisco, California, USA, 7-11 July 2001. Morgan Kaufmann.

[624]
Jorge Muruzábal and A. Muñoz. Diffuse pattern learning with Fuzzy ARTMAP and PASS. In Davidor and Schwefel [228], pages 376-385.

[625]
Jorge Muruzábal. Fuzzy and Probabilistic Reasoning in Simple Learning Classifier Systems. In Proceedings of the 2nd IEEE International Conference on Evolutionary Computation, volume 1, pages 262-266. IEEE Press, 1995.

[626]
Jorge Muruzábal. Mining the space of generality with uncertainty-concerned cooperative classifiers. In Banzhaf et al. [32], pages 449-457.
Generality is a recurrent theme in automated inductive systems. Induction of general patterns/rules is of course complicated by several factors. For example, higher levels of uncertainty and error are naturally introduced by generality. Moreover, it is not clear what sort of trade-off should be sought between increasing generality and decreasing predictive power. As a result, specific criteria to guide the search for useful general rules do not abound. In this paper, I reconsider these issues in the context of the generalized, fuzzy-like classifier system first proposed by Frey and Slate (1991) and later equipped with a Bayesian learning component by Muruzabal (1998). A crucial feature of this approach is that uncertainty is probabilistically measured at each classifier in the population. A new reinforcement policy exploiting this probabilistic structure and priming cooperation among general classifiers is introduced and shown to promote the stability of niches of reasonably high predictive power. The underlying genetic algorithm contributes effectively to learning although it somehow counteracts the built-in bias towards generality.

[627]
Ichiro Nagasaka and Toshiharu Taura. 3D Geometric Representation for Shape Generation using Classifier System. In Koza et al. [528], pages 515-520.
It is difficult to represent free-form shape features in a solid geometric model which is capable of holding and manipulating them after synthesis. In this paper, we propose a new model of representation of free-form shape features for this task. The key concept of this study is the Shape Feature Generation Process model (SFGP model) which consists of numerous sets of rules using a classifier system (CS). The idea of developmental biology is applied to a development of a computational model of the representation called the cell division model. In this model, the rules are evolved through interaction between cells and its growing free-form shapes. Finally, a computer program is developed for evaluation of the model by combination of two existing shapes for verification that the shape features are preserved in the combined shapes. It is demonstrated that the model can produce a variety of combined shapes with original, often exaggerated, features and the rules which hold features can be specified.

[628]
Filippo Neri and Attilio Giordana. A distributed genetic algorithm for concept learning. In Eshelman [298], pages 436-443.

[629]
Filippo Neri and L. Saitta. Exploring the power of genetic search in learning symbolic classifiers. IEEE Trans. on Pattern Analysis and Machine Intelligence, PAMI-18:1135-1142, 1996.
Abstract¿In this paper we show, in a constructive way, that there are problems for which the use of genetic algorithm based learning systems can be at least as effective as traditional symbolic or connectionist approaches. To this aim, the system REGAL* is briefly described, and its application to two classical benchmarks for Machine Learning is discussed, by comparing the results with the best ones published in the literature.

[630]
Filippo Neri. First Order Logic Concept Learning by means of a Distributed Genetic Algorithm. PhD thesis, University of Milano, Italy, 1997.

[631]
Filippo Neri. Comparing local search with respect to genetic evolution to detect intrusions in computer networks. In Proceedings of the 2000 Congress on Evolutionary Computation (CEC00) [185], pages 238-243.
The detection of intrusions over computer networks (i.e., network access by non-authorized users) can be cast to the task of detecting anomalous patterns of network traffic. In this case, models of normal traffic have to be determined and compared against the current network traffic. Data mining systems based on Genetic Algorithms can contribute powerful search techniques for the acquisition of patterns of the network traffic from the large amount of data made available by audit tools. We compare models of network traffic acquired by a system based on a distributed genetic algorithm with the ones acquired by a system based on greedy heuristics. Also we provide an empirical proof that representation change of the network data can result in a significant increase in the classification performances of the traffic models. Network data made available from the Information Exploration Shootout project and the 1998 DARPA Intrusion Detection Evaluation have been chosen as experimental testbed.

[632]
Filippo Neri. Relating two cooperative learning strategies to the features of the found concept description. In Lee Spector, Erik D. Goodman, Annie Wu, W.B. Langdon, Hans-Michael Voigt, Mitsuo Gen, Sandip Sen, Marco Dorigo, Shahram Pezeshk, Max H. Garzon, and Edmund Burke, editors, Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2001), page 986, San Francisco, California, USA, 7-11 July 2001. Morgan Kaufmann.

[633]
Filippo Neri. Cooperative concept learning by means of A distributed GA. In W. B. Langdon, E. Cantú-Paz, K. Mathias, R. Roy, D. Davis, R. Poli, K. Balakrishnan, V. Honavar, G. Rudolph, J. Wegener, L. Bull, M. A. Potter, A. C. Schultz, J. F. Miller, E. Burke, and N. Jonoska, editors, GECCO 2002: Proceedings of the Genetic and Evolutionary Computation Conference, page 953. Morgan Kaufmann Publishers, 2002.

[634]
Volker Nissen and Jörg Biethahn. Determining a Good Inventory Policy with a Genetic Algorithm. In Jörg Biethahn and Volker Nissen, editors, Evolutionary Algorithms in Management Applications, pages 240-249. Springer Verlag, 1995.

[635]
M. O. Odetayo and D. R. McGregor. Genetic algorithm for inducing control rules for a dynamic system. In Schaffer [718], pages 177-182. It could be argued this is a GA as opposed to a classifier system approach.
We present a Genetic Algorithm (GA)-based method for automatically inducing control rules for a dynamic physical system. A GA is used to induce control rules for a dynamic physical system: a simulated pole-cart system. The task is to move a wheeled cart, with a rigid pole hinged on top of it, along a bounded straight track without the pole falling beyond a predefined vertical angle and without the cart going off the ends of the track limits. This is achieved by applying a force of fixed magnitude to the left or right of the cart. The dynamics of the system are unknown to the GA. The only information for evaluating performance is a failure signal indicating that the pole-cart system is out of control. This presents a genuinely difficult assignment problem. We compare the performance of the method with other learning algorithm for the same task. We also compare the ability of the algorithms to adapt to changing conditions. We conclude that genetic algorithm is both effective and robust.

[636]
T. O'Hara and L. Bull. Building anticipations in an accuracy-based learning classifier system by use of an artificial neural network. In IEEE Congress on Evolutionary Computation (CEC 2005), pages 2046-2052. IEEE, 2005.

[637]
T. O'Hara and L. Bull. A memetic accuracy-based neural learning classifier system. In Proceedings of the IEEE congress on evolutionary computation (CEC 2005), pages 2040-2045. IEEE, 2005.

[638]
Toby O'Hara and Larry Bull. Backpropagation in accuracy-based neural learning classifier systems. In Tim Kovacs, Xavier LLòra, Keiki Takadama, Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson, editors, Learning Classifier Systems. International Workshops, IWLCS 2003-2005, Revised Selected Papers, volume 4399 of LNCS, pages 25-39. Springer, 2007.

[639]
Jae Chan Oh. Improved Classifier System Using Genetic Algorithms. Master's thesis, Wright State University, (year unknown -- pre-2000).

[640]
Norihiko Ono and Adel T. Rahmani. Self-Organization of Communication in Distributed Learning Classifier Systems. In Albrecht et al. [7], pages 361-367.
In this paper, an application of learning classifier systems is presented. An artificial multi-agent environment has been designed. Mate finding problem, a learning task inspired by nature, is considered which needs cooperation by two distinct agents to achieve the goal. The main feature of our system is existence of two parallel learning subsystems which have to agree on a common communication protocol to succeed in accomplishing the task. Apart from standard learning algorithms, a unification mechanism has been introduced to encourage coordinated behavior among the agents belonging to the same class. Experimental results are presented which demonstrate the effectiveness of this mechanism and the learning capabilities of classifier systems.

[641]
G. Deon Oosthuizen. Machine Learning: A mathematical framework for neural network, symbolic and genetics-based learning. In Schaffer [718], pages 385-390.
The relationship between genetics-based learning, neural network learning and typical AI-type symbolic learning approaches is highlighted by showing how each of the approaches can be mapped onto a common mathematical framework. This involves the mapping of the respective representations onto a structure called a lattice. We describe how the graphical representation of a lattice constructed and explain how it models the learning processes manifested in all three the approaches mentioned.

[642]
F. Oppacher and D. Deugo. The Evolution of Hierarchical Representations. In Proceedings of the 3rd European Conference on Artificial Life. Springer-Verlag, 1995.

[643]
Albert Orriols-Puig and Ester Bernadó-Mansilla. The class imbalance problem in learning classifier systems: a preliminary study. In GECCO Workshops 2005, pages 74-78. ACM Press, 2005.

[644]
Albert Orriols-Puig and Ester Bernadó-Mansilla. The Class Imbalance Problem in UCS Classifier System: Fitness Adaptation. In Proceedings of the 2005 Congress on Evolutionary Computation, volume 1, pages 604-611. IEEE, 2005.

[645]
A. Orriols-Puig and E. Bernadó-Mansilla. Bounding XCS's parameters for unbalanced datasets. In Maarten Keijzer et al., editor, Proceedings of the 2006 Genetic and Evolutionary Computation Conference (GECCO 2006), pages 1561-1568. ACM, 2006.

[646]
Albert Orriols-Puig and Ester Bernadó-Mansilla. Revisiting UCS: Description, Fitness Sharing, and Comparison with XCS. In Jaume Bacardit, Ester Bernadó-Mansilla, Martin Butz, Tim Kovacs, Xavier Llorà, and Keiki Takadama, editors, Learning Classifier Systems. 10th and 11th International Workshops (2006-2007), volume 4998/2008 of Lecture Notes in Computer Science, pages 96-111. Springer, 2008.
This paper provides a deep insight into the learning mechanisms of UCS, a learning classifier system (LCS) derived from XCS that works under a supervised learning scheme. A complete description of the system is given with the aim of being useful as an implementation guide. Besides, we review the fitness computation, based on the individual accuracy of each rule, and introduce a fitness sharing scheme to UCS. We analyze the dynamics of UCS both with fitness sharing and without fitness sharing over five binary-input problems widely used in the LCSs framework. Also XCS is included in the comparison to analyze the differences in behavior between both systems. Results show the benefits of fitness sharing in all the tested problems, specially those with class imbalances. Comparison with XCS highlights the dynamics differences between both systems.

[647]
Albert Orriols-Puig and Ester Bernadó-Mansilla. The Class Imbalance Problem in UCS Classifier System: A Preliminary Study. In Tim Kovacs, Xavier LLòra, Keiki Takadama, Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson, editors, Learning Classifier Systems. International Workshops, IWLCS 2003-2005, Revised Selected Papers, volume 4399 of LNCS, pages 161-180. Springer, 2007.

[648]
A. Orriols-Puig, J. Casillas, and E. Bernadò-Mansilla. Fuzzy-UCS: preliminary results. In H. Lipson, editor, Genetic and Evolutionary Computation Conference, GECCO 2007, Proceedings, pages 2871-2874. ACM, 2007.

[649]
A. Orriols-Puig, D.E. Goldberg, K. Sastry, and E. Bernadó-Mansilla. Modeling XCS in class imbalances: population size and parameter settings. In H. Lipson et al., editor, Genetic and evolutionary computation conference, GECCO 2007, pages 1838-1845. ACM, 2007.

[650]
A. Orriols-Puig, D.E. Goldberg, K. Sastry, and E. Bernadó-Mansilla. Modeling XCS in class imbalances: population size and parameter settings. In H. Lipson, editor, Genetic and Evolutionary Computation Conference, GECCO 2007, Proceedings, pages 1838-1845. ACM, 2007.
This paper analyzes the scalability of the population size required in XCS to maintain nichesthat are infrequently activated.Facetwise models have been developed to predict the effect of the imbalance ratio-ratio betweenthe number of instances of the majority class and the minority class that are sampled to XCS-on population initialization, andon the creation and deletion of classifiers of the minority class. While theoretical models show that, ideally, XCS scales linearly with the imbalance ratio, XCS with standard configuration scales exponentially. The causes that are potentially responsible for this deviation from the ideal scalability are also investigated. Specifically, the inheritance procedure of classifiers' parameters, mutation, and subsumption are analyzed, and improvements in XCS's mechanisms are proposed to effectively and efficiently handle imbalanced problems. Once the recommendations are incorporated to XCS, empirical results show that the population size in XCS indeed scales linearly with the imbalance ratio.

[651]
A. Orriols-Puig, K. Sastry, P.L. Lanzi, D.E. Goldberg, and E. Bernadò-Mansilla. Modeling selection pressure in xcs for proportionate and tournament selection. In H. Lipson, editor, Genetic and Evolutionary Computation Conference, GECCO 2007, Proceedings, page 1846–1853. ACM, 2007.

[652]
Albert Orriols-Puig, Jorge Casillas, and Ester Bernadó-Mansilla. Evolving fuzzy rules with ucs: Preliminary results. In Jaume Bacardit, Ester Bernadó-Mansilla, Martin Butz, Tim Kovacs, Xavier Llorà, and Keiki Takadama, editors, Learning Classifier Systems. 10th and 11th International Workshops (2006-2007), volume 4998/2008 of Lecture Notes in Computer Science, pages 57-76. Springer, 2008.
This paper presents Fuzzy-UCS, a Michigan-style Learning Fuzzy-Classifier System designed for supervised learning tasks. Fuzzy-UCS combines the generalization capabilities of UCS with the good interpretability of fuzzy rules to evolve highly accurate and understandable rule sets. Fuzzy-UCS is tested on a large collection of real-world problems, and compared to UCS and three highly-used machine learning techniques: the decision tree C4.5, the support vector machine SMO, and the fuzzy boosting algorithm Fuzzy LogitBoost. The results show that Fuzzy-UCS is highly competitive with respect to the four learners in terms of performance, and that the fuzzy representation permits a much better understandability of the evolved knowledge. These promising results of the online architecture of Fuzzy-UCS allow for further research and application of the system to new challenging problems.

[653]
Albert Orriols-Puig, Jorge Casillas, and Ester Bernadó-Mansilla. Genetic-based machine learning systems are competitive for pattern recognition. Evolutionary Intelligence, 1(3):209-232, 2008.
During the last decade, research on Genetic-Based Machine Learning has resulted in several proposals of supervised learning methodologies that use evolutionary algorithms to evolve rule-based classification models. Usually, these new GBML approaches are accompanied by little experimentation and there is a lack of comparisons among different proposals. Besides, the competitiveness of GBML systems with respect to non-evolutionary, highly-used machine learning techniques has only been partially studied. This paper reviews the state of the art in GBML, selects some of the best representatives of different families, and compares the accuracy and the interpretability of their models. The paper also analyzes the behavior of the GBML approaches with respect to some of the most influential machine learning techniques that belong to different learning paradigms such as decision trees, support vector machines, instance-based classifiers, and probabilistic classifiers. The experimental observations emphasize the suitability of GBML systems for performing classification tasks. Moreover, the analysis points out the strengths of the different systems, which can be used as recommendation guidelines on which systems should be employed depending on whether the user prefers to maximize the accuracy or the interpretability of the models.

[654]
Albert Orriols-Puig, Kumara Sastry, David E. Goldberg, and Ester Bernadó-Mansilla. Substructural surrogates for learning decomposable classification problems. In Jaume Bacardit, Ester Bernadó-Mansilla, Martin Butz, Tim Kovacs, Xavier Llorà, and Keiki Takadama, editors, Learning Classifier Systems. 10th and 11th International Workshops (2006-2007), volume 4998/2008 of Lecture Notes in Computer Science, pages 235-254. Springer, 2008.
This paper presents a learning methodology based on a substructural classification model to solve decomposable classification problems. The proposed method consists of three important components: (1) a structural model, which represents salient interactions between attributes for a given data, (2) a surrogate model, which provides a functional approximation of the output as a function of attributes, and (3) a classification model, which predicts the class for new inputs. The structural model is used to infer the functional form of the surrogate. Its coefficients are estimated using linear regression methods. The classification model uses a maximally-accurate, least-complex surrogate to predict the output for given inputs. The structural model that yields an optimal classification model is searched using an iterative greedy search heuristic. Results show that the proposed method successfully detects the interacting variables in hierarchical problems, groups them in linkages groups, and builds maximally accurate classification models. The initial results on non-trivial hierarchical test problems indicate that the proposed method holds promise and also shed light on several improvements to enhance the capabilities of the proposed method.

[655]
Ramon Alfonso Palacios-Durazo and Manuel Valenzuela-Rendon. Lessons learned from LCSs: An incremental non-generational coevolutionary algorithm. In Bart Rylander, editor, Genetic and Evolutionary Computation Conference (GECCO) Late Breaking Papers, pages 248-254, 2003.

[656]
Alexandre Parodi and P. Bonelli. The Animat and the Physician. In Meyer and Wilson [613], pages 50-57.

[657]
Alexandre Parodi and Pierre Bonelli. A New Approach to Fuzzy Classifier Systems. In Forrest [335], pages 223-230.
In this article, we present a new approach to fuzzy classifier system learning, and experiment with it on three different functions. This new approach is generic in the sense that it not only allows learning of fuzzy rules, but also of membership functions and output weights. Moreover, our algorithm is simple and yet exhibits good results.

[658]
Mukesh J. Patel and Marco Dorigo. Adaptive Learning of a Robot Arm. In Fogarty [324], pages 180-194.
ALECSYS, an implementation of a learning classifier system (LCS) on a net of transputers was utilised to train a robot arm to solve a light approaching task. This task, as well as more complicated ones, has already been learnt by ALECSYS implemented on AutonoMouse, a small autonomous robot. The main difference between the present and previous applications are, one, the robot arm has asymmetric constraints on its effectors, and two, given its higher number of internal degrees of freedom, and its non anthropomorphic shape, it was not obvious, as it was with the AutonoMouse, where to place the visual sensors and what sort of proprioceptive (the angular position of the arm joints) information to provide to support learning. We report results of a number of exploratory simulations of the robot arm's relative success in learning to perform the light approaching task with a number of combinations of visual and proprioceptive sensors. On the bases of results of such trials it was possible to derive a near optimum combination of sensors which is now being implemented on a real robot arm (an IBM 7547 with a SCARA geometry). Finally, the implications these findings, particularly with reference to LCS based evolutionary approach to learning, are discussed.

[659]
Mukesh J. Patel and U. Schnepf. Concept Formation as Emergent Phenomena. In Francisco J. Varela and P. Bourgine, editors, Proceedings First European Conference on Artificial Life, pages 11-20. MIT Press, 1992.

[660]
Mukesh J. Patel, Marco Colombetti, and Marco Dorigo. Evolutionary Learning for Intelligent Automation: A Case Study. Intelligent Automation and Soft Computing, 1(1):29-42, 1995.

[661]
Ray C. Paton. Designing Adaptable Systems through the Study and Application of Biological Sources. In Vic Rayward-Smith, editor, Applications of Modern Heuristic Methods, pages 39-54. Alfred Waller Ltd, 1995.
The adaptive capabilities of biological systems have provided the inspiration for numerous developments in computer science, this article focuses on evolutionary computation. Following a review the general selectionist agenda in evolutionary computing, a number of new biological sources ideas are discussed. Two experimental designs are discussed: directed adaptation and an enhancement to Genetic Based Learning in a Classifier System called the Life History GA. Some preliminary results using these ideas are then presented. The concluding section reflects on some general issues associated with the design of evolutionary systems.

[662]
Nicolas Pech-Gourg and Jin-Kao Hao. A genetic algorithm for the classification of natural corks. In Lee Spector, Erik D. Goodman, Annie Wu, W.B. Langdon, Hans-Michael Voigt, Mitsuo Gen, Sandip Sen, Marco Dorigo, Shahram Pezeshk, Max H. Garzon, and Edmund Burke, editors, Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2001), pages 1382-1388, San Francisco, California, USA, 7-11 July 2001. Morgan Kaufmann.

[663]
Rolf Pfeifer, Bruce Blumberg, Jean-Arcady Meyer, and Stewart W. Wilson, editors. From Animals to Animats 5. Proceedings of the Fifth International Conference on Simulation of Adaptive Behavior (SAB98). A Bradford Book. MIT Press, 1998.

[664]
Steven E. Phelan. Using Artificial Adaptive Agents to Explore Strategic Landscapes. PhD thesis, School of Business, Faculty of Law and Management, La Trobe University, Australia, 1997.
Recent developments in computer science have made it possible to simulate whole populations of artificial agents within the confines of a single personal computer. These artificial agents can be programmed to act in ways that mimic the behaviour of physical, biological and economic agents. Such developments come at a time when researchers in strategic management have been complaining about a lack of data to test theoretical predictions. There is a perception that the ability to generate novel hypotheses has greatly exceeded the ability to empirically test them. This dissertation investigates whether agent-based simulation can be used to test hypotheses in strategic management. SELESTE is a highly-abstract artificial world that was developed using concepts from Amit and Schoemaker's (1993) integrated theory of strategy. SELESTE provided the environment, or strategic landscape, for our artificial agents to explore. The agents themselves were modelled using an algorithm from artificial intelligence known as a learning classifier system. Artificial agents in SELESTE were shown to behave in similar ways to human agents. Three studies were selected to showcase the range of problems that agent-based simulation can address. The first problem investigated whether differences in the cognitive capacity of firms led to sustainable differences in performance. The importance of differences in cognitive capacity was shown to decline as the absolute level of cognitive capacity increased. The second study investigated the conditions under which imitation proved a superior strategy to innovation. It was revealed that imitation was a powerful strategy under all but the most extreme conditions. The final study revealed that firms that divided their operations into multiple profit-centres or 'patches' performed better than firms organised as a single profit-centre. It was concluded that agent-based simulation represented a useful method for exploring problems in strategic management. Calls were made to establish a research program in agent-based simulation to build on these findings and to refine existing techniques.

[665]
Wolfgang Stolzmann Pier Luca Lanzi and Stewart W. Wilson (guest editors). Journal of evolutionary computing, special issue on learning classifier systems, 11(3), 2003.

[666]
A. G. Pipe and Brian Carse. A Comparison between two Architectures for Searching and Learning in Maze Problems. In Fogarty [324], pages 238-249.
We present two architectures, each designed to search 2-Dimensional mazes in order to locate a ``goal'' position, both of which perform on-line learning as the search proceeds. The first architecture is a form of Adaptive Heuristic Critic which uses a Genetic Algorithm to determine the Action Policy and a Radial Basis Function Neural Network to store the acquired knowledge of the Critic. The second is a stimulus-response Classifier System (CS) which uses a Genetic Algorithm, applied ``Michigan'' style, for rule generation and the ``Bucket Brigade'' algorithm for rule reinforcement. Experiments conducted using agents based upon each architectural model lead us to a comparison of performance, and some observations on the nature and relative levels of abstraction in the acquired knowledge.

[667]
A. G. Pipe and Brian Carse. Autonomous Acquisition of Fuzzy Rules for Mobile Robot Control: First Results from two Evolutionary Computation Approaches. In Whitely et al. [878], pages 849-856.
We describe two architectures that autonomously acquire fuzzy control rules to provide reactive behavioural competencies in a simulated mobile robotics application. One architecture is a `Pittsburgh'-style Fuzzy Classifier System (Pitt1). The other architecture is a `Michigan'-style Fuzzy Classifier System (Mich1). We tested the architectures on the ability to acquire an ``investigative'' obstacle avoidance competency. We found that Mich1 implemented a more local incremental search than the other architecture. In simpler environments Mich1 was typically able to find adequate solutions with significantly fewer fitness evaluations. Since fitness evaluation can be very time consuming in this application, it could be a strong positive factor. However, when the rule set must implement a competency in more complex environments, the situation is somewhat different. The superior ability of Pitt1 to retain a number of schema in the population during the process of optimisation, is then a crucial strength.

[668]
A. G. Pipe and B. Carse. Michigan and pittsburgh fuzzy classifier systems for learning mobile robot control rules: an experimental comparison. In 14th Int. Conf. of Florida Artificial Intelligence Research Society FLAIRS-2001, pages 493-498. AAAI Press, 2001.

[669]
A. G. Pipe and B. Carse. First results from experiments in fuzzy classifier system architectures for mobile robotics. In H.-P. Schwefel J.-J. Merelo Guervós, P. Adamidis, H.-G. Beyer, J.-L. Fernández-Villacañas, editor, Parallel Problem Solving from Nature - PPSN VII, 7th International Conference, Granada, Spain, September 7-11, 2002. Proceedings, number 2439 in Lecture Notes in Computer Science, LNCS, page 578 ff. Springer-Verlag, 2002.

[670]
R. Piroddi and R. Rusconi. A Parallel Classifier System to Solve Learning Problems. Master's thesis, Dipartimento di Elettronica e Informazione, Politecnico di Milano, Milano, Italy., 1992.

[671]
Mitchell A. Potter, Kenneth A. De Jong, and John J. Grefenstette. A Coevolutionary Approach to Learning Sequential Decision Rules. In Eshelman [298], pages 366-372.
We present a coevolutionary approach to learning sequential decision rules which appears to have a number of advantages over non-coevolutionary approaches. The coevolutionary approach encourages the formation of stable niches representing simpler subbehaviors. The evolutionary direction of each subbehavior can be controlled independently, providing an alternative to evolving complex behavior using intermediate training steps. Results are presented showing a significant learning rate speedup over a noncoevolutionary approach in a simulated robot domain. In addition, the results suggest the coevolutionary approach may lead to emergent problem decompositions.

[672]
C. L. Ramsey and John J. Grefenstette. Case-based initialization of genetic algorithms. In Forrest [335], pages 84-91. http://www.ib3.gmu.edu/gref/.
In this paper, we introduce a case-based method of initializing genetic algorithms that are used to guide search in changing environments. This is incorporated in an anytime learning system. Anytime learning is a general approach to continuous learning in a changing environment. The agent's learning module continuously tests new strategies against a simulation model of the task environment, and dynamically updates the knowledge base used by the agent on the basis of the results. The execution module includes a monitor that can dynamically modify the simulation model based on its observations of the external environment; an update to the simulation model causes the learning system to restart learning. Previous work has shown that genetic algorithms provide an appropriate search mechanism for anytime learning. This paper extends the approach by including strategies, which are learned under similar environmental conditions, in the initial population of the genetic algorithm. Experiments show that case-based initialization of the population results in a significantly improved performance.

[673]
C. L. Ramsey and John J. Grefenstette. Case-based anytime learning.. In D. W. Aha, editor, Case-Based Reasoning: Papers from the 1994 Workshop. AAAI Press: Menlo Park, CA, 1994. Also Technical Report WS-94-07 http://www.ib3.gmu.edu/gref/.
We discuss a case-based method of initializing genetic algorithms that are used to guide search in changing environments. This is incorporated in an anytime learning system. Anytime learning is a general approach to continuous learning in a changing environment. A genetic algorithm with a case-based component provides an appropriate search mechanism for anytime learning. When the genetic algorithm is restarted, strategies which were previously learned under similar environmental conditions are included in the initial population of the genetic algorithm. We have evaluated the system by comparing performance with and without the case-based component, and case-based initialization of the population results in a significantly improved performance.

[674]
Gregory J. E. Rawlins, editor. Proceedings of the First Workshop on Foundations of Genetic Algorithms (FOGA91). Morgan Kaufmann: San Mateo, 1991.

[675]
Colin Reveley. A learning classifier system adapted for hold'em poker. Master's thesis, Birkbeck College, University of London, UK, 2002.
This paper stems from the observation that in a game of poker, if our opponents follow a strategy and do not play randomly, there must be some set of rules such that if we follow those rules we will maximise our profits. Perhaps a Genetic Algorithm would be able to discover such a set of rules. This report looks into how feasible such an approach might be, and after some analysis, makes an attempt at an implementation.

[676]
Robert A. Richards and Sheri D. Sheppard. Classifier System Based Structural Component Shape Improvement Utilizing I-DEAS. In Iccon User's Conference Proceeding. Iccon, 1992.

[677]
Robert A. Richards and Sheri D. Sheppard. Learning Classifier Systems in Design Optimization. In Design Theory and Methodology '92. The American Society of Mechanical Engineers, 1992.

[678]
Robert A. Richards and Sheri D. Sheppard. Two-dimensional Component Shape Improvement via Classifier System. In Artificial Intelligence in Design '92. Kluwer Academic Publishers, 1992.

[679]
Robert A. Richards and Sheri D. Sheppard. A Learning Classifier System for Three-dimensional Shape Optimization. In H. M. Voigt, W. Ebeling, I. Rechenberg, and H.-P. Schwefel, editors, Parallel Problem Solving from Nature -- PPSN IV, volume 1141 of LNCS, pages 1032-1042. Springer-Verlag, 1996.
A learning classifier system complex is developed in order to accomplish the broader goal of developing a methodology to perform generalized zeroth-order two-and three-dimensional shape optimization. Specifically, the methodology has the objective of determining the optimal boundary to minimize mass while satisfying constraints on stress and geometry. Even with the enormous advances in shape optimization no method has proven to be satisfactory across the broad spectrum of optimization problems facing the modern engineer. Similarly the available software in the field of learning classifier systems is so embryonic that a new software package has to be developed for this application. The shape optimization via hypothesizing inductive classifier system complex (SPHINcsX) instantiates the methodology in a software package overcoming many of the limitations of today's conventional shape optimization techniques, while advancing the state-of-the-art in classifier system software tools.

[680]
Robert A. Richards and Sheri D. Sheppard. Three-Dimensional Shape Optimization Utilizing a Learning Classifier System. In Koza et al. [527], pages 539-546.

[681]
Robert A. Richards. Zeroth-Order Shape Optimization Utilizing a Learning Classifier System. PhD thesis, Stanford University, 1995. http://www-leland.stanford.edu/~buc/SPHINcsX/book.html.
A methodology to perform generalized zeroth-order two- and three-dimensional shape optimization utilizing a learning classifier system is developed and applied. To this end, the applicability of machine learning to mechanical engineering is investigated. Specifically, the methodology has the objective of determining the optimal boundary to minimize mass while satisfying constraints on stress and geometry. Even with the enormous advances in shape optimization no method has proven to be satisfactory across the broad spectrum of optimization problems facing the modern engineer. The methodology developed in this book is based upon a classifier system (CS) and exploits the CS's adaptability and generality. It thereby overcomes many of the limitations of today's conventional shape optimization techniques. A CS learns rules, postulated as if-then statements, in order to improve its performance in an arbitrary environment, (which for this investigation consists of stress and mass information from components). From this input, and from a population of initially randomly generated rules, the classifier system is expected to learn to make the appropriate component shape modifications to reach a minimum mass design while satisfying all stress constraints. The CS learns by utilizing the design improvement success or failure feedback. Nearly all shape optimization algorithms developed to date depend on sensitivity information in order to function. This research does not present sensitivity information to the classifier system. Thus, the classifier system must not only learn from a clean slate, but confronts the additional challenge of learning without information that most other shape optimization algorithms deem essential. Therefore, the main deliverable is a zeroth-order shape optimization methodology. After a review of mechanical engineering shape optimization methods, an explanatory presentation of CSs and their underlying genetic algorithm (GA) describes how classifier systems learn from feedback and the GA. With this foundation set, the coupling of the shape optimization domain with the classifier system proceeds to form, the Shape oPtimization via Hypothesizing Inductive classifier system compleX (SPHINcsX). The complex learns shape optimization by its application to a suite of sizing optimization problems. The most tangible artifact of this research is the successful development of the zeroth-order shape optimization complex. The complex proved adept at solving both two- and three-dimensional shape optimization problems. The research also provides a demonstrative example of the power and flexibility of machine learning in general and CSs in particular -- how they may be leveraged as tools for mechanical engineering design, and insights into their proper application.

[682]
Robert A. Richards. Classifier System Metrics: Graphical Depictions. In Koza et al. [529], pages 652-657.

[683]
Rick L. Riolo. Bucket Brigade Performance: I. Long Sequences of Classifiers. In Grefenstette [391], pages 184-195.
In Holland-type classifier systems the bucket brigade algorithm allocates strength (``credit'') to classifiers that lead to rewards from environment. This paper presents results that show the bucket brigade algorithm basically works as designed -- strength is passed down sequences of coupled classifiers from those classifiers that receive rewards directly from the environment to those that are stage setters. Results indicate it can take a fairly large number of trials for a classifier system to respond to changes in its environment by reallocating strength down competing sequences of classifiers that implement simple reflex and non-reflex behaviors. However, ``bridging classifiers'' are shown to dramatically decrease the number of times a long sequence must be executed in order reallocate strength to all the classifiers in the sequence. Bridging classifiers also were shown to be one way to avoid problems caused by sharing classifiers across competing sequences.

[684]
Rick L. Riolo. Bucket Brigade Performance: II. Default Hierarchies. In Grefenstette [391], pages 196-201.
Learning systems that operate in environments with huge numbers of states must be able to categorize the states into equivalence classes that can be treated alike. Holland-type classifier systems can learn to categorize states by building default hierarchies of classifiers (rules). However, for default hierarchies to work properly classifiers that implement exception rules must be able to control the system when they are applicable, thus preventing the default rules from making mistakes. This paper presents results that show the standard bucket brigade algorithm does not lead to correct exception rules always winning the competition with the default rules they protect. A simple modification to the bucket brigade algorithm is suggested, and results are presented that show this modification works as desired: default hierarchies can be made to achieve payoff rates as near to optimal as desired.

[685]
Rick L. Riolo. CFS-C: A Package of Domain-Independent Subroutines for Implementing Classifier Systems in Arbitrary User-Defined Environments. Technical report, University of Michigan, 1988.

[686]
Rick L. Riolo. Empirical Studies of Default Hierarchies and Sequences of Rules in Learning Classifier Systems. PhD thesis, University of Michigan, 1988.

[687]
Rick L. Riolo. The Emergence of Coupled Sequences of Classifiers. In Schaffer [718], pages 256-264.
Sequences of coupled classifiers are a basic component of higher level knowledge and control structures in classifier systems that follow the ``Michigan'' approach. This paper explores mechanisms for promoting the emergence of coupled sequences in task domains that provide payoff only after a series of coordinated actions. The results indicate that useful coupled chains do emerge once some minor changes are made to the basic classifier system.

[688]
Rick L. Riolo. The Emergence of Default Hierarchies in Learning Classifier Systems. In Schaffer [718], pages 322-327.
Default hierarchies have been proposed as a way for classifier systems to categorize events efficiently and as accurately as necessary. This paper studies the emergence of default hierarchies for tasks that favor use of default hierarchies over homomorphic models. The results indicate that default hierarchies do rapidly emerge, but that they also tend to be replaced by homomorphic models.

[689]
Rick L. Riolo. Lookahead Planning and Latent Learning in a Classifier System. In Meyer and Wilson [613], pages 316-326.
Classifier systems (CSs) have been used to simulate and describe the behavior of adaptive organisms, animats, and robots. However, classifier system implementations to date have all been reactive systems, which use simple S-R rules and which base their learning algorithms on trial-and-error reinforcement techniques similar to the Hullian Law of Effect. While these systems have exhibited interesting behavior and good adaptive capacity, they cannot do other types of learning which require having explicit internal models of the external world, e.g., using complex plans as humans do, or doing latent learning of the type observed in rats. This paper describes a classifier system that is able to learn and use internal models both to greatly decrease the time to learn general sequential decision tasks and to enable the system to exhibit latent learning.

[690]
Rick L. Riolo. Modelling Simple Human Category Learning with a Classifier System. In Booker and Belew [74], pages 324-333.
A classifier system is used to model human performance on a simple deterministic discrimination task in which subjects must acquire categories based on their experience. The performance of the classifier system is compared to data from experiments with humans and to the performance of an adaptive neural net model described by Gluck and Bower. The classifier system is able to replicate data on human performance, including one observation not replicated by the neural net model. The classifier system also misses one prediction the neural net makes correctly. Three keys to the classifier system's performance are: (1) default hierarchies in which exceptions usually overrule more general rules; (2) a bucket brigade algorithm in which each classifier pays the average bid made by co-winning rules that produce the same message it does (rather than just paying its own bid) and receives an equal share of any payoff; and (3) the use of a bid tax.

[691]
Rick L. Riolo. The discovery and use of forward models for adaptive classifier systems. In Collected Abstracts for the First International Workshop on Learning Classifier System (IWLCS-92) [486]. October 6-8, NASA Johnson Space Center, Houston, Texas.
Paper is an extended abstract

[692]
Joaquin Rivera and Roberto Santana. Improving the Discovery Component of Classifier Systems by the Application of Estimation of Distribution Algorithms. In Proceedings of Student Sessions ACAI'99: Machine Learning and Applications, pages 43-44, Chania, Greece, July 1999.

[693]
A. Robert, F. Chantemargue, and M. Courant. Grounding Agents in EMud Artificial Worlds. In Proceedings of the First International Conference on Virtual Worlds, Paris (France), July 1-3, 1998.
This paper suggests that in the context of autonomous agents and generation of intelligent behavior for such agents, a more important focus should be held on the symbolic context that forms the basis of computer programs. Basically, software agents are symbolic entities living in a symbolic world and this has an effect on how one should think about designing frameworks for their evolution or learning. We will relate the problem of symbol grounding to that of sensory information available to agents. We will then introduce an experimental environment based on virtual worlds called EMuds, where both human and artificial agents can interact. Next, we show how it can be applied in the framework of multi-agent systems to address emergence based problems and report preliminary results. We then conclude with some ongoing and future work.

[694]
Gary Roberts. A Rational Reconstruction of Wilson's Animat and Holland's CS-1. In Schaffer [718], pages 317-321.
In the short history of Genetic Algorithms, there have been a plethora of techniques used. Even within the subfield of classifier systems, many differing implementation exist. It becomes difficult to compare ones results with others, and to determine the cause of actual performance differences. To that end I have attempted a rational reconstruction encompassing two systems described in the literature: Wilson's Animat and the CS-1 system of Holland and Reitman. The results obtained differ, sometimes markedly, from the published versions they attempt to duplicate.

[695]
Gary Roberts. Dynamic Planning for Classifier Systems. In Forrest [335], pages 231-237.
Classifier systems were developed as a scheme for applying genetic algorithms to problems where they would otherwise be difficult to utilize. The bucket brigade algorithm has been used to handle the credit assignment problem associated with classifier systems. This research demonstrates an alternative to bucket brigade based upon dynamic planning and Q-learning. The advantages of the new system, Dyna-Q-CS, include the construction of an explicit world model that speeds learning in the absence of plentiful reward information and is amenable to sophisticated planning techniques. The experimental results show that Dyna-Q-CS learns as fast, or faster than the particular bucket brigade system used (based upon Wilson's Animat), but that the asymptotic performance of Dyna-Q-CS needs improvement. Some form of annealing could solve this asymptotic performance problem.

[696]
George G. Robertson and Rick L. Riolo. A Tale of Two Classifier Systems. Machine Learning, 3:139-159, 1988.
This paper describes two classifier systems that learn. These are rulebased systems that use genetic algorithms, which are based on an analogy with natural selection and genetics, as their principal learning mechanism, and an economic model as their principal mechanism for apportioning credit. CFS-C is a domain independent learning system that has been widely tested on serial computers. *CFS is a parallel implementation of CFS-C that makes full use of the inherent parallelism of classifier systems and genetic algorithms, and that allows the exploration of large-scale tasks that were formerly impractical. As with other approaches to learning, classifier systems in their current form in their current form work well for moderately-sized tasks but break down for larger tasks. In order to shed light on this issue, we present several empirical studies of known issues in classifier systems, including the effects of population size, the actual contribution of genetic algorithms, the use of rule chaining in solving higher-order tasks, and issues of task representation and dynamic population convergence. We conclude with a discussion of some major unresolved issues in learning classifier systems and some possible approaches to making them more effective on complex tasks.

[697]
George G. Robertson. Parallel Implementation of Genetic Algorithms in a Classifier System. In Grefenstette [391], pages 140-147. Also Technical Report TR-159 RL87-5 Thinking Machines Corporation.

[698]
George G. Robertson. Population Size in Classifier Systems. In Proceedings of the Fifth International Conference on Machine Learning, pages 142-152. Morgan Kaufmann, 1988.

[699]
George G. Robertson. Parallel Implementation of Genetic Algorithms in a Classifier System. In Davis [234], pages 129-140.

[700]
J. A. Meyer H. L. Roitblat and S. W. Wilson, editors. From Animals to Animats 2. Proceedings of the Second International Conference on Simulation of Adaptive Behavior (SAB92). A Bradford Book. MIT Press, 1992.

[701]
Wesley Romão, Alex A. Freitas, and Roberto C. S. Pacheco. A genetic algorithm for discovering interesting fuzzy prediction rules: Applications to science and technology data. In W. B. Langdon, E. Cantú-Paz, K. Mathias, R. Roy, D. Davis, R. Poli, K. Balakrishnan, V. Honavar, G. Rudolph, J. Wegener, L. Bull, M. A. Potter, A. C. Schultz, J. F. Miller, E. Burke, and N. Jonoska, editors, GECCO 2002: Proceedings of the Genetic and Evolutionary Computation Conference, pages 1188-1195, New York, 9-13 July 2002. Morgan Kaufmann Publishers.

[702]
Peter Ross, Sonia Schulenburg, Javier Marín-Blázquez, and Emma Hart. Hyper-heuristics: Learning to combine simple heuristics in bin-packing problems. In W. B. Langdon, E. Cantú-Paz, K. Mathias, R. Roy, D. Davis, R. Poli, K. Balakrishnan, V. Honavar, G. Rudolph, J. Wegener, L. Bull, M. A. Potter, A. C. Schultz, J. F. Miller, E. Burke, and N. Jonoska, editors, GECCO 2002: Proceedings of the Genetic and Evolutionary Computation Conference, pages 942-948. Morgan Kaufmann Publishers, 2002.

[703]
S. Ross. Accurate Reaction or Reflective Action? Master's thesis, School of Cognitive and Computing Sciences, University of Sussex, 1994.

[704]
S. E. Rouwhorst and A. P. Engelbrecht. Searching the forest: Using decision trees as building blocks for evolutionary search in classification databases. In Proceedings of the 2000 Congress on Evolutionary Computation (CEC00) [185], pages 633-638.
A new evolutionary search algorithm, called BGP, to be used for classification tasks in data mining, is introduced. It is different from existing evolutionary techniques in that it does not use indirect representations of a solution, such as bit strings or grammars. The algorithm uses decision trees of various sizes as individuals in the populations and operators, e.g. crossover, are performed directly on the trees. when compared to C4.5 and CN2 on a benchmark of problems, BGP shows very good results.

[705]
A. Sanchis, J. M. Molina, P. Isasi, and J. Segovia. Knowledge acquisition including tags in a classifier system. In Angeline et al. [8], pages 137-144.

[706]
Adrian V. Sannier and Erik D. Goodman. Midgard: A Genetic Approach to Adaptive Load Balancing for Distributed Systems. In John E. Laird, editor, Proc. Fifth Int. Conf. Machine Learning (ICML), pages 174-180. Morgan Kaufmann, 1988.

[707]
Manuel Filipe Santos. Learning Classifiers in Distributed Environments. PhD thesis, Departamento de Sistemas de Informação, Universidade do Minho, Portugal, 2000.
Over the last decade Learning Classifier Systems (LCS) have received increasing attention from researchers motivated to develop flexible machine learning devices. However, they were not scalable; i.e., if used to solve real world problems, LCS are outperformed. One has to deal with vast sets of classifiers with large bit length. Learning in distributed environments, on the other hand, constitutes another potential application area for the LCS, if one can dispose of conceptual models for LCS' distribution. One way to by-pass these problems is implementing LCS on parallel hardware, improving the systems performance, and the use of efficient structural organisational models to avoid complexity. Working on this direction an agent-oriented architecture based on the blackboard paradigm, a model that brings parallelism to the LCS functionalities and caters for the distribution of classifiers and associates processes, was object of study. The resulting system is a distributed processing system called DICE (A DIstributed Computational Environment for Genetic-Based Classifier Systems). To achieve these goals, various domains of knowledge were exploited including distributed processing systems, agent-based systems, blackboard-based systems, logic programming and parallel genetic algorithms. The DICE system is the consolidation of such knowledge, a computation medium that improves LCS' performance and allows for the parallelization and distribution of classifiers in a modular and flexible way, augmenting the scope of application of LCS.

[708]
Cédric Sanza, Christophe Destruel, and Yves Duthen. Agents autonomes pour l'interaction adaptative dans les mondes virtuels. In 5ème Journées de l'Association Francaise d'Informatique Graphique. Décembre 1997, Rennes, France, 1997. In French.
Dans le domaine des interfaces pour les dispositifs de la réalité virtuelle, les périphériques de dialogue perturbent les sens et augmentent les difficultés de la communication homme/machine. Afin de minimiser cette perturbation et pour fournir une assistance efficace à chaque utilisateur, nous proposons d'utiliser un système d'apprentissage basé sur les outils de la vie artificielle.

[709]
Cédric Sanza, Christophe Destruel, and Yves Duthen. A learning method for adaptation and evolution in virtual environments. In 3rd International Conference on Computer Graphics and Artificial Intelligence, April 1998, Limoges, France, 1998.
This document has no abstract

[710]
Cédric Sanza, Christophe Destruel, and Yves Duthen. Autonomous actors in an interactive real-time environment. In ICVC'99 International Conference on Visual Computing Feb. 1999, Goa, India, 1999.
This paper presents a learning system based on artificial life that uses short-term memory and knowledge sharing. Inspired from classifier systems, the model allows to generate behaviors for agents integrated in a multi-task environment. The user, immersed in the scene, interacts through his clone with autonomous actors. By his own behavior, he influences the agents' one. An agent perceives the world through sensors and acts through effectors in order to produce rules (called classifiers). Rewards from the environment allow to adjust the strength of every rule that is used in order to define the best behavior. The ``sending message'' protocol has been included to increase the performances of t he system in complex environment. By combining communication and evolution, we then produce a realtime application (a virtual soccer) where the user plays with the other agents. After a short period of adaptation, the simulation gives some positive results: a coherent global behavior is built by the teams.

[711]
Cédric Sanza, Christophe Destruel, and Yves Duthen. Learning in real-time environment based on classifiers system. In 7th International Conference in Central Europe on Computer Graphics, Visualization and Interactive Digital Media'99, Plzen, Czech Republic, 1999.
This paper presents a new architecture of a classifier system for learning in virtual environments. The model will be integrated in our multi-user platform to provide interaction between intelligent agents and user clones. An agent is an autonomous entity equipped with sensors and effectors. Its behavior is guided by rewards coming from the environment that produce rules called classifiers. The knowledge is shared between agents by using the ``sending-message'' protocol to increase the global efficiency of the group. The classifier system is specially adapted to a multi-task environment and incorporates a short-term memory to record the recent events of the simulation. These ideas have been implemented and used to develop a virtual soccer where the user plays with autonomous agents that combine communication and evolution.

[712]
Cédric Sanza, Cyril Panatier, Hervé Luga, and Yves Duthen. Adaptive Behavior for Cooperation: a Virtual Reality Application. In 8th IEEE International Workshop on Robot and Human Interaction September 1999, Pisa, Italy, 1999.
In this paper, we present a behavioral system based on artificial life for animating actors in a virtual reality application. Through a virtual soccer game, we show how a set of autonomous players (called agents) can cooperate and communicate to perform common tasks. The user is immersed in the game. He interacts with the other agents and he is integrated in the cooperation and in the communication systems. Every entity reacts in real-time by using a classifiers system which is composed of a set of binary rules and a reward system. The originality of such method is the ability to build a behavior (by emergence) without initial knowledge. The analysis of the simulation gives interesting results: after convergence, the global behavior of the teams produces coherent movements. Moreover, the introduction of disturbances does not affect the performances of the classifiers system.

[713]
Shaun Saxon and Alwyn Barry. XCS and the Monk's problem. In Wu [923], pages 272-281.
It has been known for some time that Learning Classifier Systems (Holland, 1986) have potential for application as Data Mining tools. Parodi and Bonelli (1990) applied the Boole LCS (Wilson, 1985) to a Lymphography data set and reported 82% classification rates. More recent work, such as GA-Miner (Flockhart,1995) has sought to extend the application of LCS to larger commercial data sets, introducing more complex attribute encoding techniques, static niching, and hybrid genetic operators in order to address the problems presented by large search spaces. Despite these results, the traditional LCS formulation has shown itself to be unreliable in the formation of accurate optimal generalisations, which are vital for the reduction of results to a human readable form. XCS (Wilson, 1995, 1998) has been shown to be capable of generating a complete and optimally accurate mapping of a test environment (Kovacs, 1996) and therefore presents a new opportunity for the application of Learning Classifier Systems to Data Mining. As part of a continuing research effort this paper presents some first results in the application of XCS to a Data Mining task. It demonstrates that XCS is able to produce a classification performance and rule set which exceeds the performance of most current Machine Learning techniques when applied to the Monk's problems (Thrun, 1991).

[714]
Shaun Saxon and Alwyn Barry. XCS and the Monk's Problems. In Lanzi et al. [544], pages 223-242.
It has been known for some time that Learning Classifier Systems (LCS) have potential for application as Data Mining tools. Parodi and Bonelli applied the Boole LCS to the Lymphography data set and reported 82 classification rates. More recent work, such as GA-Miner has sought to extend the application of the GA-based classification system to larger commercial data sets, introducing more complex attribute encoding techniques, static niching, and hybrid genetic operators in order to address the problems presented by large search spaces. Despite these results, the traditional LCS formulation has shown itself to be unreliable in the formation of accurate optimal generalisations, which are vital for the reduction of results to a human readable form. XCS has been shown to be capable of generating a complete and optimally accurate mapping of a test environment and therefore presents a new opportunity for the application of Learning Classifier Systems to the classification task in Data Mining. As part of a continuing research effort this paper presents some first results in the application of XCS to a particular Data Mining task. It demonstrates that XCS is able to produce a classification performance and rule set which exceeds the performance of most current Machine Learning techniques when applied to the Monk's problems.

[715]
Andreas Schachtner. A classifier system with integrated genetic operators. In H.-P. Schwefel and R. Männer, editors, Parallel Problem Solving from Nature, volume 496 of Lecture Notes in Computer Science, pages 331-337, Berlin, 1990. Springer.
This text presents a classifier system (CS), which is able to adapt to an environment by adjusting th activation probabilities of the rules and changing the rules itself. The operators for changing the rules are incorporated into the CS, thus allowing for an adaption of the rates of change on-line during the search process for better rules. An age is attached to the rules. Removal of rules from the rule set is done according to age. Experiments show that this approach to adapting the rule set by means of internal genetic operators (GO) is superior to exogenous genetic operators.

[716]
J. David Schaffer. Some experiments in machine learning using vector evaluated genetic algorithms. PhD thesis, Vanderbilt University, Nashville, 1984.

[717]
J. David Schaffer. Learning Multiclass Pattern Discrimination. In Grefenstette [389], pages 74-79.

[718]
J. David Schaffer, editor. Proceedings of the 3rd International Conference on Genetic Algorithms (ICGA89), George Mason University, June 1989. Morgan Kaufmann.

[719]
Sonia Schulenburg and Peter Ross. An Adaptive Agent Based Economic Model. In Lanzi et al. [544], pages 263-282.
In this paper we describe a simple model of adaptive agents of different types, represented by Learning Classifier Systems (LCS), which make investment decisions about a risk free bond and a risky asset under a well defined stock market environment. Our main aim is to explore the degree of reliability that artificially intelligent agents can have when applied to real life economic problems. We do this by evaluating whether an LCS is able to represent competent traders in a real market scenario in which daily stock prices and dividends are given to the agents exogenously, so permitting us to focus on the dynamics and evolution of the behavior of these evolving traders without having to be concerned about how their actions affect the market. We present results of adaptive and non-adaptive simulations over a period of ten years of real data of a specific stock and show that the artificial agents, by displaying different and rich behaviours evolved throughout the simulations, are able to discover and refine novel and successful sets of market strategies that can outperform baseline strategies such as buy-and-hold or merely keeping money in the bank at a good rate of interest, even though the agents pay commission on every trade.

[720]
Sonia Schulenburg and Peter Ross. Strength and Money: An LCS Approach to Increasing Returns. In Proceedings of the International Workshop on Learning Classifier Systems (IWLCS-2000), in the Joint Workshops of SAB 2000 and PPSN 2000 [487]. Extended abstract.
Paper is an extended abstract

[721]
Sonia Schulenburg and Peter Ross. Strength and money: An LCS approach to increasing returns. In Lanzi et al. [545], pages 114-137.
This paper reports on a number of experiments where three different groups of artificial agents learn, forecast and trade their holdings in a real stock market scenario given exogeneously in the form of easily-obtained stock statistics such as various price moving averages, first difference in prices, volume ratios, etc. These artificial agent-types trade while learning during -- in most cases -- a ten year period. They normally start at the beginning of the year 1990 with a fixed initial wealth to trade over two assets (a bond and a stock) and end in the second half of the year 2000. The adaptive agents are represented as Learning Classifier Systems (LCSs), that is, as sets of bit-encoded rules. Each condition bit expresses the truth or falsehood of a certain real market condition. The actual conditions used differ between agents. The forecasting performance is then compared against the performance of the buy-and-hold strategy, and trend-following strategy and finally against the bank investment over the same period at a fixed compound interest rate. To make the experiments as real as possible, agents pay comissions on every trade. The results so far suggest that this is an excellent approach to make trading decisions in the stock market.

[722]
Sonia Schulenburg and Peter Ross. Explorations in LCS models of stock trading. In Lanzi et al. [546], pages 151-180.

[723]
Alan C. Schultz and John J. Grefenstette. Evolving Robot Behaviors. Poster at the 1994 Artificial Life Conference. (NCARAI Report: AIC-94-017) http://www.ib3.gmu.edu/gref/.
This paper discusses the use of evolutionary computation to evolve behaviors that exhibit emergent intelligent behavior. Genetic algorithms are used to learn navigation and collision avoidance behaviors for robots. The learning is performed under simulation, and the resulting behaviors are then used to control the actual robot. Some of the emergent behavior is described in detail.

[724]
Alan C. Schultz and John J. Grefenstette. Improving Tactical Plans with Genetic Algorithms. In Proceedings of the Second International Conference on Tools for Artificial Intelligence. IEEE, 1990.

[725]
Alan C. Schultz, Connie Logia Ramsey, and John J. Grefenstette. Simulation assisted learning by competition: Effects of noise differences between training model and target environment. In Proceedings of Seventh International Conference on Machine Learning (ICML), pages 211-215. Morgan Kaufmann, 1990.
The problem of learning decision rules for sequential tasks is addressed, focusing on the problem of learning tactical plans from a simple flight simulator where a plane must avoid a missile. The learning method relies on the notion of competition and employs genetic algorithms to search the space of decision policies. Experiments are presented that address issues arising from differences between the simulation model on which learning occurs and the target environment on which the decision rules are ultimately tested. Specifically, either the model or the target environment may contain noise. These experiments examine the effect of learning tactical plans without noise and then testing the plans in a noisy environment, and the effect of learning plans in a noisy simulator and then testing the plans in a noise-free environment. Empirical results show that, while best result are obtained when the training model closely matches the target environment, using a training environment that is more noisy than the target environment is better than using using a training environment that has less noise than the target environment.

[726]
Alan C. Schultz, John Grefenstette, and Kenneth De Jong. Learning to break things: adaptive testing of intelligent controllers. In Thomas Bäck, David B. Fogel, and Zbigniew Michalewicz, editors, Handbook of Evolutionary Computation, pages G3.4:1-G3.4:10. IOP Publishing Ltd and Oxford University Press, 1997.

[727]
Dale Schuurmans and Jonathan Schaeffer. Representational Difficulties with Classifier Systems. In Schaffer [718], pages 328-333. http://www.cs.ualberta.ca/ jonathan/Papers/Papers/classifier.ps.
Classifier systems are currently in vogue as a way of using genetic algorithms to demonstrate machine learning. However, there are a number of difficulties with the formalization that can influence how knowledge is represented and the rate at which the system can learn. Some of the problems are inherent in classifier systems, and one must learn to cope with them, while others are pitfalls waiting to catch the unsuspecting implementor. This paper identifies some of these difficulties, suggesting directions for the further evolution of classifier systems.

[728]
Hans-Paul Schwefel and Reinhard Männer, editors. Parallel Problem Solving from Nature: Proceedings of the First International Workshop. Dortmund, FRG, 1-3 Oct 1990, number 496 in Lecture Notes in Computer Science, Heidelberg, 1990. Springer.

[729]
Tod A. Sedbrook, Haviland Wright, and Richard Wright. Application of a Genetic Classifier for Patient Triage. In Booker and Belew [74], pages 334-338.
This research develops and applies a genetic classifier system (CS) to triage patients presenting with symptoms of upper respiratory infections (URIs). The CS searches among 66 dichotomous patient signs and symptoms to evolve classifiers that best explain care provider triage decisions. The systems search is directed by specifying relative costs of false positives and false negatives. The model achieved a sensitivity and specificity of 100% and 42%, respectively, when applied to a triage case base of URI patients. A split-sample validation of the system shows its accuracy is comparable to that achieved with a triage protocol developed by Infectious Disease Specialists.

[730]
Sandip Sen and Mahendra Sekaran. Multiagent Coordination with Learning Classifier Systems. In Gerhard Weiß and Sandip Sen, editors, Proceedings of the IJCAI Workshop on Adaption and Learning in Multi-Agent Systems, volume 1042 of LNAI, pages 218-233. Springer Verlag, 1996.

[731]
Sandip Sen. Classifier system learning of multiplexer function. Dept. of Electrical Engineering, University of Alabama, Tuscaloosa, Alabama. Class Project, 1988.

[732]
Sandip Sen. Sequential Boolean Function Learning by Classifier System. In Proc. of 1st International Conference on Industrial and Engineering Applications of Artificial Intelligence and Expert Systems, 1988.

[733]
Sandip Sen. Noise Sensitivity in a simple classifier system. In Proc. 5th Conf. on Neural Networks & Parallel Distributed Processing, 1992.

[734]
Sandip Sen. Improving classification accuracy through performance history. In Forrest [335], pages 652-652.

[735]
Sandip Sen. A Tale of two representations. In Proc. 7th International Conference on Industrial and Engineering Applications of Artificial Intelligence and Expert Systems, pages 245-254, 1994.

[736]
Sandip Sen. Modelling human categorization by a simple classifier system. In WSC1: 1st Online Workshop on Soft Computing. Aug 19-30, 1996. http://www.bioele.nuee.nagoya-u.ac.jp/wsc1/papers/p020.html, 1996.
NEWBOOLE is a simple stimulus-response classifier system that has been used successfully in a number of supervised concept learning and classification problems. In this paper we use NEWBOOLE on a relatively simple categorization problem involving medical diagnosis, and compare its performance to that of human subjects on the same problem. The system is provided with the relevant rules and learning involves generating appropriate strengths for the different rules so as to categorize both seen and unseen examples. Results obtained with the simple classifier system exhibit the same trends as demonstrated by the humans. We present an analysis explaining the working of the system on the given task. Other experiments presented in this paper include experiments to replicate human performance on filtering and condensation tasks as described by Kruschke Kruschke92:ALCOVE, and experiments involving learning to attend to relevant dimensions for proper categorization Shepard61:Learning. In the latter task, the system is not provided the relevant rule set, and hence has to discover the appropriate rules as well as learn to assign credit to the useful ones.

[737]
Tiago Sepulveda and Mario Rui Gomes. A Study on the Evolution of Learning Classifier Systems. In Proceedings of the International Workshop on Learning Classifier Systems (IWLCS-2000), in the Joint Workshops of SAB 2000 and PPSN 2000 [487]. Extended abstract.
Paper is an extended abstract

[738]
F. Seredynski and C. Z. Janikow. Learning nash equilibria by coevolving distributed classifier systems. In Angeline et al. [8], pages 1619-1626.

[739]
F. Seredynski, Pawel Cichosz, and G. P. Klebus. Learning classifier systems in multi-agent environments. In Proceedings of the First IEE/IEEE International Conference on Genetic Algorithms in Engineering Systems: Innovations and Applications (GALESIA'95), 1995.
The paper is devoted to the problem of learning decision policies in multi-agent games. This problem is a simple, but appealing computational model of several important real-world problems in such domains as parallel computing, optimization, and control on one hand, and economy, social, and political sciences on the other hand. We describe a general framework for studying games of intelligent agents, extending the basic model of games with limited interactions, and its specific realization based on learning classifier systems. Simulation results are presented that illustrate the convergence properties of the resulting system. Avenues for future work in this area are outlined.

[740]
S. Sette and L. Boullart. An implementation of genetic algorithms for rule based machine learning. Engineering Applications of Artificial Intelligence, 13(4):381-390, 2000.
Genetic algorithms have given rise to two new fields of research where (global) optimisation is of crucial importance: `Genetic Programming' and `Genetic based Machine Learning' (GBML). In this paper the second domain (GBML) will be introduced. An overview of one of the first GBML implementations by Holland, also known as the Learning Classifier Systems (LCS) will be given. After describing and solving a well-known basic (educational) problem a more complex application of GBML is presented. The goal of this application is the automatic development of a rule set for an industrial production process. To this end, the case study on generating a rule set for predicting the spinnability in the fibre-to-yarn production process will be presented. A largely modified LCS, called Fuzzy Efficiency based Classifier System (FECS), originally designed by one of the authors, is used to solve this problem successfully.

[741]
Kamran Shafi and Hussein A. Abbass. An Adaptive Genetic-Based Signature Learning System for Intrusion Detection. Expert Systems With Applications, 37, 2009. In Press.

[742]
Kamran Shafi, Hussein A. Abbass, and Weiping Zhu. The Role of Early Stopping and Population Size in XCS for Intrusion Detection. In Proceedings of the 6th International Conference on Simulated Evolution and Learning, Lecture Notes in Computer Science (LNCS), 4247, pages 50-57. Springer, Heidelberg, 2006.

[743]
Kamran Shafi, Hussein A. Abbass, and Weiping Zhu. Real Time Signature Extraction From a Supervised Classifier System. In Proceedings of the 2007 Congress on Evolutionary Computation, 2007.

[744]
Kamran Shafi, Tim Kovacs, Hussein A. Abbass, and Weiping Zhu. Intrusion Detection with Evolutionary Learning Classifier Systems. Natural Computing, 8(1):3-27, 2009.
Evolutionary Learning Classifier Systems (LCSs) combine reinforcement learning or supervised learning with effective genetics-based search techniques. Together these two mechanisms enable LCSs to evolve solutions to decision problems in the form of easy to interpret rules called classifiers. Although LCSs have shown excellent performance on some data mining tasks, many enhancements are still needed to tackle features like high dimensionality, huge data sizes, non-uniform distribution of classes, etc. Intrusion detection is a real world problem where such challenges exist and to which LCSs have not previously been applied. An intrusion detection problem is characterised by huge network traffic volumes, difficult to realize decision boundaries between attacks and normal activities and highly imbalanced attack class distribution. Moreover, it demands high accuracy, fast processing times and adaptability to a changing environment. We present the results and analysis of two classifier systems (XCS and UCS) on a subset of a publicly available benchmark intrusion detection dataset which features serious class imbalances and two very rare classes. We introduce a better approach for handling the situation when no rules match an input on the test set and recommend this be adopted as a standard part of XCS and UCS. We detect little sign of overfitting in XCS but somewhat more in UCS. However, both systems tend to reach near-best performance in very few passes over the training data. We improve the accuracy of these systems with several modifications and point out aspects that can further enhance their performance. We also compare their performance with other machine learning algorithms and conclude that LCSs are a competitive approach to intrusion detection.

[745]
Jiefu Shi. Genetic Algorithms for Game Playing. In C. Karr and L. M. Freeman, editors, Industrial Applications of Genetic Algorithms, pages 321-338. CRC Press, 1998.
This chapter examines genetic algorithms (GA) and machine learning using the game of tic-tac-toe. After `learning' acceptable strategies for playing the game, the GA-driven computer player is able to play a competent game of tic-tac-toe. Results obtained using a GA are compared to results obtained using alternative AI techniques.

[746]
Sotaro Shimada and Yuichiro Anzai. Component-Based Adaptive Architecture with Classifier Systems. In Pfeifer et al. [663].

[747]
Sotaro Shimada and Yuichiro Anzai. Fast and Robust Convergence of Chained Classifiers by Generating Operons through Niche Formation. In Banzhaf et al. [32], page 810. One page poster paper.
This work has no abstract

[748]
Sotaro Shimada and Yuichiro Anzai. On Niche Formation and Corporation in Classifier System. In Takadama [815].

[749]
Takayuki Shiose and Tetsuo Sawaragi. Extended learning classifier systems by dual referencing mechanism. In Takadama [815].

[750]
Lingyan Shu and Jonathan Schaeffer. VCS: Variable Classifier System. In Schaffer [718], pages 334-339. http://www.cs.ualberta.ca/~jonathan/Papers/Papers/vcs.ps.
Genetic-algorithms-based learning classifier systems suffer from a number of problems that cause system instability, resulting in poor performance. These problems include genetic operation disruptions and difficulties in maintaining good classifiers and classifier structures in the population. A method is proposed in which structural ties are used to achieve coherence, impose cooperation and encourage co-adaptation among classifiers. A hierarchically structured classifier system (HCS) has been implemented to show the effect of this structuring. At the lowest level, classifiers (individuals) are grouped into families. Higher-order structures, such as a community of families, can be introduced if necessary. The experimental results show a significant improvement in system performance and stability. The relationships between the HCS framework and the Michigan and Pittsburgh approaches are discussed.

[751]
Lingyan Shu and Jonathan Schaeffer. Improving the Performance of Genetic Algorithm Learning by Choosing a Good Initial Population. Technical Report TR-90-22, University of Alberta, CS DEPT, Edmonton, Alberta, Canada, 1990.

[752]
Lingyan Shu and Jonathan Schaeffer. HCS: Adding Hierarchies to Classifier Systems. In Booker and Belew [74], pages 339-345.
Genetic-algorithms-based learning classifier systems suffer from a number of problems that cause instability, resulting in poor performance. These problems include genetic operation disruptions and difficulties in maintaining good classifiers and classifier structures in the population. A method is proposed in which structural ties are used to achieve coherence, impose cooperation and encourage co-adaptation among classifiers. A hierarchically structured classifier system (HCS) has been implemented to show the effect of this structuring. At the lowest level, classifiers (individuals) are grouped into families. Higher-order structures, such as communities of families, can be introduced if necessary. The experimental results show a significant improvement in system performance and stability. The relationships between the HCS framework and the Michigan and Pittsburgh approaches are discussed.

[753]
Olivier Sigaud and Pierre Gerard. Being reactive by exchanging roles: an empirical study. In Balancing reactivity and Social Deliberation in Multiagent Systems, volume 2103 of LNAI, pages 150-172. Springer-Verlag, 2001.

[754]
Olivier Sigaud and Pierre Gerard. Using classifier systems as adaptive expert systems for control. In Advances in Classifier Systems, number 1996 in LNAI, pages 138-157. Springer-Verlag, 2001.

[755]
Olivier Sigaud. On the usefulness of a semi-automated Classifier System: the engineering perspective. In Proceedings of the International Workshop on Learning Classifier Systems (IWLCS-2000), in the Joint Workshops of SAB 2000 and PPSN 2000 [487]. Extended abstract.
Paper is an extended abstract

[756]
Michael C. Simon. Using XCS to form hyper-heuristics for the set covering problem. In Alwyn M. Barry, editor, GECCO 2003: Proceedings of the Bird of a Feather Workshops, Genetic and Evolutionary Computation Conference, pages 246-249. AAAI, 2003.

[757]
M.G. Smith and L. Bull. Genetic programming with a genetic algorithm for feature construction and selection. Genetic Programming and Evolvable Machines, 6(3):265-281, 2005.

[758]
Robert E. Smith and H. Brown Cribbs. Is a Learning Classifier System a Type of Neural Network? Evolutionary Computation, 2(1):19-36, 1994.
This paper suggests a simple analogy between learning classifier systems (LCSs) and neural networks (NNs). By clarifying the relationship between LCSs and NNs, the paper indicates how techniques from one can he utilized in the other. The paper points out that the primary distinguishing characteristic of the LCS is its use of a co-adaptive genetic algorithm (GA), where the end product of evolution is a diverse population of individuals that cooperate to perform useful computation. This stands in contrast to typical GA/NN schemes, where a population of networks is employed to evolve a single, optimized network. To fully illustrate the LCS/NN analogy used in this paper, an LCS-like NN is implemented and tested. The test is constructed to run parallel to a similar GA/NN study that did not employ a co-adaptive GA. The test illustrates the LCS/NN analogy and suggests an interesting new method for applying GAs in NNs. Final comments discuss extensions of this work and suggest how LCS and NN studies can further benefit each other.

[759]
Robert E. Smith and Bruce A. Dike. An application of genetic algorithms to air combat maneuvering. In Thomas Bäck, David B. Fogel, and Zbigniew Michalewicz, editors, Handbook of Evolutionary Computation, pages G3.1:1-G3.1:14. IOP Publishing Ltd and Oxford University Press, 1997.

[760]
Robert E. Smith and David E. Goldberg. Reinforcement Learning with Classifier Systems: Adaptive Default Hierarchy Formation. Technical Report 90002, TCGA, University of Alabama, 1990.

[761]
Robert E. Smith and David E. Goldberg. Variable Default Hierarchy Separation in a Classifier System. In Rawlins [674], pages 148-170.
A learning classifier system (LCS) is a machine learning system that incorporates a production-system framework and a genetic algorithm (GA) for rule discovery (Goldberg, 1989; Holland, 1975). A primary feature of LCSs is their potential to exploit overlapping sets of rules called default hierarchies. Default hierarchies increase rule set parsimony, enlarge the solution set, and lend themselves to graceful refinement by the GA (Holland, Holyoak, Nisbett, & Thagard, 1986). Traditionally, auction-based, specificity-biased credit allocation (CA) and conflict resolution (CR) schemes have been used to encourage default hierarchy formation in an LCS. Analyses presented in this paper suggest that these schemes cannot be expected to perform adequately in arbitrary LCS environments. This paper presents an alternate CA/CR that associates two measures with each classifier in place of the single, traditional strength measure. The first measure is a payoff estimate, which is tuned by the linear-update scheme usually used for strength. The second measure is a priority factor that is tuned to control the outcome of a necessity auction. In the necessity auction the winning classifier pays out the payoff estimate of its nearest competitor, rather than a fraction of its own payoff estimate. Results and analyses are presented that show that this CA/CR scheme can induce variable bid separation that responds to the demands of the LCS environment. Additional analyses show that this scheme allows an LCS to adequately exploit a broader class of default hierarchies than traditional schemes. Several avenues are suggested for further study.

[762]
Robert E. Smith and David E. Goldberg. Reinforcement learning with classifier systems: adaptative default hierarchy formation. Applied Artificial Intelligence, 6, 1992.

[763]
S. F. Smith and D. P. Greene. Cooperative Diversity using Coverage as a Constraint. In Collected Abstracts for the First International Workshop on Learning Classifier System (IWLCS-92) [486]. October 6-8, NASA Johnson Space Center, Houston, Texas.
Paper is an extended abstract

[764]
Robert E. Smith and H. B. Cribbs III. Cooperative Versus Competitive System Elements in Coevolutionary Systems. In Maes et al. [595], pages 497-505.

[765]
Robert E. Smith and H. B. Cribbs III. Combined biological paradigms. Robotics and Autonomous Systems, 22(1):65-74, 1997.
This paper introduces an autonomous systems strategy that combines two biological inspirations: neural networks and genetic algorithms (GAs). These ideas have been combined in a variety of ways in other systems, but the scheme presented here has several unique features. The system presented is based on an analogy between learning classifier systems (LCSs) and neural networks first presented by Smith and Cribbs [Evolutionary Computation 2(1) (1994) 19-36]. However, Smith and Cribbs focused on supervised learning. The work presented in this paper transfers these ideas to the realm of autonomous systems by considering reinforcement learning. In the new system, a neural network is used to map environmental states to Q value. The neural network structure is based on an LCS. The GA acts to shape neural connectivity, and the number of hidden layer nodes. The GAs action is similar to its action in the LCS. The suggested system is evaluated in a simulated mobile robot test environment. Experimental results suggest that the system is effective in learning and evolving parsimonious strategy representations for autonomous systems. Future directions for investigation of this system are discussed.

[766]
Robert Elliott Smith and Max Kun Jiang. A learning classifier system with mutual-information-based fitness. In Jaume Bacardit, Ester Bernadó-Mansilla, Martin Butz, Tim Kovacs, Xavier Llorà, and Keiki Takadama, editors, Learning Classifier Systems. 10th and 11th International Workshops (2006-2007), volume 4998/2008 of Lecture Notes in Computer Science, pages 136-153. Springer, 2008.
This paper introduces a new variety of learning classifier system (LCS), called MILCS, which utilizes mutual information as fitness feedback. Unlike most LCSs, MILCS is specifically designed for supervised learning. We present preliminary results, and contrast them to results from XCS. We discuss the explanatory power of the resulting rule sets and introduce a new technique for visualizing explanatory power. Final comments include future directions of this research, including investigations in neural networks and other systems.

[767]
Robert E. Smith and Manuel Valenzuela-Rendón. A Study of Rule Set Development in a Learning Classifier System. In Schaffer [718], pages 340-346.
In this paper a set of simultaneous equations is used to examine the development of rule sets in a classifier system using a genetic algorithm (GA) as its primary discovery device. These equations are developed for a stimulus-response classifier system with a single, two-bit condition and a binary action. Computations are presented that show the importance of niching operators in a classifier system's GA. Further experiments compare and contrast the effects of fitness sharing and mating restriction as niching and speciation operators in the classifier system.

[768]
Robert E. Smith, Stephanie Forrest, and Alan S. Perelson. Population Diversity in an Immune System Model: Implications for Genetic Search. In L. Darrell Whitley, editor, Foundations of Genetic Algorithms 2, pages 153-165. Morgan Kaufmann, 1992.

[769]
Robert E. Smith, Stephanie Forrest, and A. S. Perelson. Searching for diverse, cooperative subpopulations with Genetic Algorithms. Evolutionary Computation, 1(2):127-149, 1993.

[770]
George D. Smith, Nigel C. Steele, and Rudolf F. Albrecht, editors. Artificial Neural Networks and Genetic Algorithms. Springer, 1997.

[771]
Robert E. Smith, B. A. Dike, B. Ravichandran, A. El-Fallah, and R. K. Mehra. The Fighter Aircraft LCS: A Case of Different LCS Goals and Techniques. In Wu [923], pages 282-289.
There are a number of common difficulties and open issues that pertain to the ``traditional'' LCS model. Many of these topics were central at The First International Workshop on Learning Classifier Systems (Houston, Texas, 1992). Since the first workshop, several significant, theoretically-supported advances in LCS practice have addressed these issues. However, a system employed by the authors to acquire novel fighter aircraft maneuvers from combat simulation is more akin to the traditional LCS model than to more recent systems. Given the difficulties often experienced in LCS research on simple problems, one must ask how a relatively primitive LCS has had consistent success in the complex domain of fighter aircraft maneuvering? This paper overviews the troublesome issues discussed at the first workshop, and recent advances. It then presents the fighter aircraft LCS, in greater detail than in previous publications. Positive results from the system are discussed. The paper then focuses on the primary reasons the fighter aircraft LCS has avoided the difficulties of the traditional LCS. The authors believe the system's success has three primary origins: differences in credit assignment, differences in action encoding, and (possibly most importantly) a difference in system goals. In the fighter aircraft system, the goal has been simply the discovery of innovative, novel tactics, rather than online control. The paper concludes by discussing the most salient features of the fighter aircraft learning system, and how those features may be profitably combined with other LCS developments.

[772]
Robert E. Smith, B. A. Dike, R. K. Mehra, B. Ravichandran, and A. El-Fallah. Classifier Systems in Combat: Two-sided Learning of Maneuvers for Advanced Fighter Aircraft. Computer Methods in Applied Mechanics and Engineering, 186(2-4):421-437, 2000.

[773]
Robert E. Smith, B. A. Dike, B. Ravichandran, A. El-Fallah, and R. K. Mehra. The Fighter Aircraft LCS: A Case of Different LCS Goals and Techniques. In Lanzi et al. [544], pages 283-300.
A system employed by the authors to acquire novel fighter aircraft manoeuvres from combat simulation is more akin to the traditional LCS model than to more recent systems. Given the difficulties often experienced in LCS research on simple problems, one must ask how a relatively primitive LCS has had consistent success in the complex domain of fighter aircraft manoeuvering. This paper presents the fighter aircraft LCS, in greater detail than in previous publications. Positive results from the system are discussed. The paper then focuses on the primary reasons the fighter aircraft LCS has avoided the difficulties of the traditional LCS. The authors believe the system's success has three primary origins: differences in credit assignment, differences in action encoding, and (possibly most important) a difference in system goals. In the fighter aircraft system, the goal has been simply the discovery of innovative, novel tactics, rather than online control. The paper concludes by discussing the most salient features of the fighter aircraft system, and how those features may be profitably combined with other LCS developments.

[774]
S. F. Smith. A Learning System Based on Genetic Adaptive Algorithms. PhD thesis, University of Pittsburgh, 1980.

[775]
S. F. Smith. Flexible Learning of Problem Solving Heuristics through Adaptive Search. In Proceedings Eight International Joint Conference on Artificial Intelligence, pages 422-425, 1983.

[776]
S. F. Smith. Adaptive learning systems. In R. Forsyth, editor, Expert Systems: Principles and Case Studies, pages 169-189. Chapman and Hall, 1984.

[777]
Robert E. Smith. Default Hierarchy Formation and Memory Exploitation in Learning Classifier Systems. PhD thesis, University of Alabama, 1991.

[778]
Robert E. Smith. A Report on The First International Workshop on Learning Classifier Systems (IWLCS-92). NASA Johnson Space Center, Houston, Texas, Oct. 6-9. ftp://lumpi.informatik.uni-dortmund.de/pub/LCS/papers/lcs92.ps.gz or from ENCORE, The Electronic Appendix to the Hitch-Hiker's Guide to Evolutionary Computation (ftp://ftp.krl.caltech.edu/pub/EC/Welcome.html) in the section on Classifier Systems, 1992.
This paper has no abstract

[779]
Robert E. Smith. Is a classifier system a type of neural network? In Collected Abstracts for the First International Workshop on Learning Classifier System (IWLCS-92) [486]. October 6-8, NASA Johnson Space Center, Houston, Texas.
Paper is an extended abstract

[780]
Robert E. Smith. Memory exploitation in learning classifier systems. In Collected Abstracts for the First International Workshop on Learning Classifier System (IWLCS-92) [486]. October 6-8, NASA Johnson Space Center, Houston, Texas.
Paper is an extended abstract

[781]
Robert E. Smith. Genetic Learning in Rule-Based and Neural Systems. In Proceedings of the Third International Workshop on Neural Networks and Fuzzy Logic, volume 1, page 183. NASA. Johnson Space Center, January 1993.
The design of neural networks and fuzzy systems can involve complex, nonlinear, and ill-conditioned optimization problems. Often, traditional optimization schemes are inadequate or inapplicable for such tasks. Genetic Algorithms (GA's) are a class of optimization procedures whose mechanics are based on those of natural genetics. Mathematical arguments show how GAs bring substantial computational leverage to search problems, without requiring the mathematical characteristics often necessary for traditional optimization schemes (e.g., modality, continuity, availability of derivative information, etc.). GA's have proven effective in a variety of search tasks that arise in neural networks and fuzzy systems. This presentation begins by introducing the mechanism and theoretical underpinnings of GA's. GA's are then related to a class of rule-based machine learning systems called learning classifier systems (LCS's). An LCS implements a low-level production-system that uses a GA as its primary rule discovery mechanism. This presentation illustrates how, despite its rule-based framework, an LCS can be thought of as a competitive neural network. Neural network simulator code for an LCS is presented. In this context, the GA is doing more than optimizing and objective function. It is searching for an ecology of hidden nodes with limited connectivity. The GA attempts to evolve this ecology such that effective neural network performance results. The GA is particularly well adapted to this task, given its naturally-inspired basis. The LCS/neural network analogy extends itself to other, more traditional neural networks. Conclusions to the presentation discuss the implications of using GA's in ecological search problems that arise in neural and fuzzy systems.

[782]
Robert E. Smith. Memory Exploitation in Learning Classifier Systems. Evolutionary Computation, 2(3):199-220, 1994.
Learning classifier systems (LCSs) offer a unique opportunity to study the adaptive exploitation of memory. Because memory is manipulated in the form of simple internal messages in the LCS, one can easily and carefully examine the development of a system of internal memory symbols. This study examines the LCS applied to a problem whose only performance goal is the effective exploitation of memory. Experimental results show that the genetic algorithm forms a relatively effective set of internal memory symbols, but that this effectiveness is directly limited by the emergence of parasite rules. The results indicate that the emergence of parasites may be an inevitable consequence in a system that must evolve its own set of internal memory symbols. The paper's primary conclusion is that the emergence of parasites is a fundamental obstacle in such problems. To overcome this obstacle, it is suggested that the LCS must form larger, multirule structures. In such structures, parasites can be more accurately evaluated and thus eliminated. This effect is demonstrated through a preliminary evaluation of a classifier corporation scheme. Final comments present future directions for research on memory exploitation in the LCS and similar evolutionary computing systems.

[783]
George D. Smith. Economic Applications of Genetic Algorithms. In Vic Rayward-Smith, editor, Applications of Modern Heuristic Methods, pages 71-90. Alfred Waller Ltd, 1995. Contains 2 pages on LCS.
This work has no abstract

[784]
Robert E. Smith. Derivative Methods: Learning Classifier Systems. In Bäck et al. [22], pages B1.2:6-B1.5:11. http://www.iop.org/Books/Catalogue/.

[785]
Lee Spector, Erik D. Goodman, Annie Wu, W. B. Langdon, Hans-Michael Voigt, Mitsuo Gen, Sandip Sen, Marco Dorigo, Shahram Pezeshk, Max H. Garzon, and Edmund Burke, editors. Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2001). Morgan Kaufmann, 2001.

[786]
Piet Spiessens. PCS: A Classifier System that Builds a Predictive Internal World Model. In PROC of the 9th European Conference on Artificial Intelligence, Stockholm, Sweden, Aug. 6-10, pages 622-627, 1990.

[787]
Bryan G. Spohn and Philip H. Crowley. Complexity of Strategies and the Evolution of Cooperation. In Koza et al. [528], pages 521-528.
We address the complexity of strategies that simulate those used in repeated pairwise social interactions by individuals capable of cooperating (C) or defecting (D). Strategies are composed of some number of interacting rules, each specifying a response (C or D) to the recent history of previous responses by the two individuals. Here we consider the extent of memory and especially the number of rules as components of complexity. Using a classifier system based on small populations playing the Iterated Prisoner's Dilemma game, we show that mutual cooperation (and thus fitness) is maximized for strategies of about 20 rules when the number of rules is fixed; that less memory generally yields more mutual cooperation; and that longer interaction sequences generate more mutual cooperation -- all in accord with previous work. Allowing rule number to evolve with short interaction sequences produces mutual cooperation near or below the low levels associated with random responses. But even with these short sequences, weak selection on rule number can be detected. Selection seems to favor fewer rules (approximately 7) when rule number can evolve than the fixed number of rules that maximzes mutual cooperation. We expect stronger selection on rule number for longer interaction sequences and larger population sizes.

[788]
Wolfgang Stolzmann and Martin Butz. Latent Learning and Action-Planning in Robots with Anticipatory Classifier Systems. In Lanzi et al. [544], pages 301-317.
Two applications of Anticipatory Classifier Systems (ACS) in robotics are discussed. The first one is a simulation of an experiment about latent learning in rats with a mobile robot. It shows than an ACS is able to learn latently, i.e. in the absence of environmental reward and that ACS can do action planning. The second one is about learning of the hand-eye coordination of a robot arm in conjunction with a camera. Goal-directed learning will be introduced. This combination of action planning and latent learning leads to a substantial reduction of the number of trials which are required to learn a complete model of a prototypical environment.

[789]
Wolfgang Stolzmann, Martin Butz, J. Hoffmann, and D. E. Goldberg. First cognitive capabilities in the anticipatory classifier system. In et al. [299], pages 287-296. Also Technical Report 2000008 of the Illinois Genetic Algorithms Laboratory.
This paper adds a new viewpoint to the Anticipatory Classifier System (ACS). It approaches the system from a psychological perspective and thus provides new insights to the current system. By simulating previously published rat experiments, the paper compares the behavior of the ACS with the behavior of the rats. Two further cognitive mechanisms are introduced to the ACS resulting in an animal-like behavior in the presented simulations. Moreover, the paper gives empirical evidence that the evolving generalized, internal environmental model is usable in the ACS for the mental adaptation of actions and thus enables reinforcement learning by mental simulation.

[790]
Wolfgang Stolzmann. Learning Classifier Systems using the Cognitive Mechanism of Anticipatory Behavioral Control, detailed version.. In Proceedings of the First European Workshop on Cognitive Modelling, pages 82-89. Berlin: TU, 1996. http://www.psychologie.uni-wuerzburg.de/stolzmann/.
A classifier system is a machine learning system that learns a collection of rules, called classifiers. Mostly, classifiers can be regarded as simple stimulus-response rules. A first level of learning called credit assignment level, consists of reinforcement learning on these classifiers. A classifier is reinforced in dependence on the result of an interaction between the CS and its environment. A second level that is independent of the first one consists of rule discovery. For that a CS usually uses genetic algorithms that can only use very indirect information about the interaction between the system and the environment in the form of rule strengths. It is often the problem with CSs that hierarchical chunks of classifiers are destroyed when the rule discovery is applied. Therefore in some applications CSs don't use the rule discovery level or don't delete classifiers (e.g. Riolo 1991). This paper gives an introduction to a new kind of CSs that learn with anticipatory behavioral control. These classifier systems are called anticipatory classifier systems (ACSs). Anticipatory behavioral control is a development of reinforcement learning on stimulus-response units and enables us to learn an internal model of an external environment. The main difference between ACSs and other CSs is that in an ACS the rule discovery level is integrated into the credit assignment level. The rule discovery algorithm of an ACS uses immediate environmental information, i.e. it's a kind of intentional rule discovery. This is a particular feature of ACSs. For example, there are no problems with hierarchical chunks of classifiers. After the introduction we prove the performance of ACSs by comparing them with other CSs. A simulation of an experiment about the latent learning of rats is then discussed and it is shown that ACSs solve the locality/globality dilemma for reactive classifier systems.

[791]
Wolfgang Stolzmann. Antizipative Classifier Systeme. PhD thesis, Fachbereich Mathematik/Informatik, University of Osnabrück, 1997.
Abstract: Antizipative Classifier Systems oder kurz ACSs sind Classifier Systems, die mittels antizipativer Verhaltenssteuerung lernen. Classifier Systems wurden 1978 von J. Holland eingeführt und bilden neben künstlichen neuronalen Netzen und Multi-Agenten-Systemen eine wichtige Klasse lernender Systeme in der Künstlichen Intelligenz. Antizipative Verhaltenssteuerung, wie sie 1992 von J. Hoffmann postuliert wurde, ist eine psychologische Lerntheorie, bei der Verhalten eine Grundvoraussetzung für Lernen ist. In der vorliegenden Arbeit ist es gelungen, antizipative Verhaltenssteuerung in Classifier Systems zu integrieren und somit zu einem Lernalgorithmus weiterzuentwickeln. Dabei wurde das Ziel verfolgt, die Ideen der antizipativen Verhaltenssteuerung möglichst unmittelbar im Algorithmus wiederzufinden. Im einzelnen gliedert sich die Arbeit in 5 Kapitel. In Kapitel 1 wird eine kurze Einführung in die Theorie der antizipativen Verhaltenssteuerung gegeben. Kapitel 2 umfaßt eine ausführliche Diskussion verschiedener Varianten von Cassifier Systems. Der Kern der Arbeit besteht aus Kapitel 3. Hier werden auf der Grundlage von Kapitel 1 und Kapitel 2 ACSs formal definiert. Das 4. Kapitel dient der Evaluation von ACSs. Dazu werden zwei Anwendungen antizipativer Classifier Systems diskutiert. Zum einen wird ein Tierexperiment aus der Verhaltensforschung und zum anderen eine Lernaufgabe für einen Roboter simuliert. Im 5. Kapitel werden Grenzen und Erweiterungsmöglichkeiten von ACSs diskutiert.

[792]
Wolfgang Stolzmann. Two Applications of Anticipatory Classifier Systems (ACSs). In Proceedings of the 2nd European Conference on Cognitive Science, pages 68-73. Manchester, U.K., 1997. http://www.psychologie.uni-wuerzburg.de/stolzmann/.
Anticipatory classifier systems (ACSs) are a new kind of classifier systems (CSs) that learn by using the cognitive mechanism of anticipatory behavioral control. At first this paper gives a brief introduction to ACSs. Then two applications of ACSs are discussed. The first one is a simulation of an experiment about latent learning that was done by Seward (1949) and first simulated by Riolo (1991). The second one consists of a simulation of a robot that has to learn its eye-hand coordination, starting without any knowledge, that was described and simulated by Birk (1995) using Drescher's schemata (Drescher, 1991 p.9).

[793]
Wolfgang Stolzmann. Anticipatory classifier systems. In Proceedings of the Third Annual Genetic Programming Conference, pages 658-664. Morgan Kaufmann, 1998. http://www.psychologie.uni-wuerzburg.de/stolzmann/gp-98.ps.gz.
Anticipatory Classifier Systems (ACS) are a new kind of classifier system (CS) that learn by using the cognitive mechanism of anticipatory behavioral control that was introduced in cognitive psychology by Hoffmann (1992). This paper gives at first a brief introduction to Hoffmann's learning mechanism. Then ACS are introduced. To prove the performance of ACS they are compared to Riolo's CFSC2(1991). In addition to a theoretical comparison a simulation of an experiment about latent learning in rats is discussed that was developed by Seward (1949).

[794]
Wolfgang Stolzmann. Untersuchungen zur adäquatheit des postulats einer antizipativen verhaltenssteuerung zur erklärung von verhalten mit ACSs. In W. Krause and U. Kotkamp, editors, Intelligente Informationsverarbeitung, pages 130-138. Deutscher Universitäts Verlag, 1998.

[795]
Wolfgang Stolzmann. Latent Learning in Khepera Robots with Anticipatory Classifier Systems. In Wu [923], pages 290-297.
Seward (1949) developed an experiment about latent learning in rats. During a learning phase rats learn the topology of a T-maze without getting any reward. This experiment is replicated with a Khepera robot that latently learns by using an Anticipatory Classifier System (ACS). The robot and its environment are simulated with the Open Mobile Robots Simulator Webots. Latent learning is defined as learning in the absence of reinforcement. Therefore this experiment cannot be simulated with usual reinforcement learning techniques. A mobile robot can observe its environment only partially, so that the Markov property is not necessarily given. Indeed, the T-maze used here is a Non-Markov environment for a Khepera robot. Non-Markov environments can be learned by adding memory to Learning Classifier Systems (cf. Cliff & Ross 1995, Lanzi 1998). An alternative for ACS is to use classifiers with behavioral sequences. This alternative is discussed. Besides it must be possible to test whether the topology of the T-maze is learned or not. For this purpose the robot is told to reach a certain point in the maze. If the robot needs more than one behavioral act to do this, then it is necessary to have a mechanism that enables the robot to do look ahead planning (cf. Riolo 1991). Such a mechanism is introduced.

[796]
Wolfgang Stolzmann. An Introduction to Anticipatory Classifier Systems. In Lanzi et al. [544], pages 175-194.
Anticipatory Classifier Systems (ACS) are classifier systems that learn by using the cognitive mechanism of anticipatory behavioral control which was introduced in cognitive psychology by Hoffmann. They can learn in deterministic multi-step environments. A stepwise introduction to ACS is given. We start with the basic algorithm and apply it in simple ``woods'' environments. It will be shown that this algorithm can only learn in a special kind of deterministic multi-step environments. Two extensions are discussed. The first one enables an ACS to learn in any deterministic multi-step environment. The second one allows an ACS to deal with a special kind of non-Markov state.

[797]
Chris Stone and Larry Bull. For real! XCS with continuous-valued inputs. Evolutionary Computation, 11(3):298-336, 2003.

[798]
Christopher Stone and Larry Bull. Towards learning classifier systems for continuous-valued online environments. In E. Cantú-Paz, J. A. Foster, K. Deb, D. Davis, R. Roy, U.-M. O'Reilly, H.-G. Beyer, R. Standish, G. Kendall, S. Wilson, M. Harman, J. Wegener, D. Dasgupta, M. A. Potter, A. C. Schultz, K. Dowsland, N. Jonoska, and J. Miller, editors, Genetic and Evolutionary Computation -- GECCO-2003, pages 1924-1925, Berlin, 2003. Springer-Verlag.

[799]
K. Takadama and T. Terano. Good solutions will emerge without a global objective function: Applying organizational-learning oriented classifier system to printed circuit board design. In The IEEE 1997 International Conference On Systems, Man and Cybernetics (SMC'97), pages 3355-3360, 1997.

[800]
Keiki Takadama, S. Nakasuka, and Takao Terano. On the credit assignment algorithm for organizational-learning oriented classifier system. In The 1997 System/information joint Symposium of SICE (The Society of Instrument and Control Engineers), pages 41-46, 1997. In Japanese.

[801]
Keiki Takadama, S. Nakasuka, and Takao Terano. Organizational-learning oriented classifier system. In The 11th Annual Conference of JSAI (Japanese Society for Artificial Intelligence), pages 201-204, 1997. In Japanese.

[802]
Keiki Takadama, S. Nakasuka, and Takao Terano. Organizational-learning oriented classifier system for intelligent multiagent systems. In The 6th Multi Agent and Cooperative Computation (MACC '97) of JSSST (Japan Society for Software Science and Technology), 1997. In Japanese.

[803]
K. Takadama, S. Nakasuka, and T. Terano. Multiagent reinforcement learning with organizational-learning oriented classifier system. In The IEEE 1998 International Conference On Evolutionary Computation (ICEC'98), pages 63-68, 1998.

[804]
Keiki Takadama, S. Nakasuka, and Takao Terano. Analyzing the roles of problem solving and learning in organizational-learning oriented classifier system. In H. Y. Lee and H. Motoda, editors, Lecture Notes in Artificial Intelligence, volume 1531, pages 71-82. Springer-Verlag, 1998.

[805]
Keiki Takadama, Takao Terano, and Katsunori Shimohara. Agent-based model toward organizational computing: From organizational learning to genetics-based machine learning. In The IEEE 1999 International Conference On Systems, Man and Cybernetics (SMC'99), volume 2, pages 604-609, 1999.

[806]
Keiki Takadama, Takao Terano, and Katsunori Shimohara. Can multiagents learn in organization? -- analyzing organizational learning-oriented classifier system. In IJCAI'99 Workshop on Agents Learning about, from and other Agents, 1999.

[807]
Keiki Takadama, Takao Terano, Katsunori Shimohara, H. Hori, and S. Nakasuka. Making Organizational Learning Operational: Implications from Learning Classifier System. Computational and Mathematical Organization Theory (CMOT), 5(3):229-252, 1999.

[808]
Keiki Takadama, Takao Terano, Katsunori Shimohara, H. Hori, and S. Nakasuka. Toward emergent problem solving by distributed classifier systems based on organizational learning. Transactions of SICE (the Society of Instrument and Control Engineers), 35(11):1486-1495, 1999. In Japanese.

[809]
K. Takadama, H. Inoue, and K. Shimohara. How to autonomously decide boundary between self and others? In The Third Asia-Pacific Conference on Simulated Evolution And Learning (SEAL'2000), 2000.

[810]
K. Takadama, T. Terano, and K. Shimohara. Designing multiple agents using learning classifier systems. In The 4th Japan-Australia Joint Workshop on Intelligent and Evolutionary Systems (JA'2000), 2000.

[811]
Keiki Takadama, Shinichi Nakasuka, and Kasunori Shimohara. Designing multiple agents using learning classifier systems - suggestions from three levels analyses. In Takadama [815].

[812]
Keiki Takadama, Takao Terano, and Katsunori Shimohara. Learning Classifier Systems meet Multiagent Environments. In Proceedings of the International Workshop on Learning Classifier Systems (IWLCS-2000), in the Joint Workshops of SAB 2000 and PPSN 2000 [487]. Extended abstract.
Paper is an extended abstract

[813]
K. Takadama, H. Inoue, M. Okada, K. Shimohara, , and O. Katai. Agent architecture based on interactive self-reflection classifier system. International Journal of Artificial Life and Robotics (AROB), 2001.

[814]
K. Takadama, S. Nakasuka, and K. Shimhara. Robustness in Organizational-learning Oriented Classifier System. Journal of Soft Computing, 6(3-4):229-239, 2002.

[815]
Keiki Takadama, editor. Exploring New Potentials in Learning Classifier Systems. A Session of the 4th Japan-Australia Joint Workshop on Intelligent and Evolutionary Systems. Ashikaga Institute of Technology, 2000.

[816]
Keiki Takadama. Organizational-learning oriented classifier system. Technical Report TR-H-290, ATR, 2000. In Japanese.

[817]
K. Tammee, L. Bull, and P. Ouen. Towards clustering with learning classifier systems. In L. Bull, E. Bernadò-Mansilla, and J. Holmes, editors, Learning Classifier Systems in Data Mining, pages 191-204. Springer, 2008.

[818]
Takao Terano and Z. Muro. On-the-fly knowledge base refinement by a classifier system. AI Communications, 4(2), 1994.

[819]
Takao Terano and Keiki Takadama. An organizational learning model of multiagents with a learning classifier system. In The 1997 Fall Conference of JASMIN (Japan Society for Management Information), pages 128-131, 1997. In Japanese.

[820]
K. Tharakannel and D. Goldberg. XCS with average reward criterion in multi-step environment. Technical report, Illinois Genetic Algorithms Laboratory, University of Illinois at Urbana-Champaign, 2002.

[821]
Kurian K. Tharakunnel, Martin V. Butz, and David E. Goldberg. Towards building block propagation in XCS: A negative result and its implications. In E. Cantú-Paz, J. A. Foster, K. Deb, D. Davis, R. Roy, U.-M. O'Reilly, H.-G. Beyer, R. Standish, G. Kendall, S. Wilson, M. Harman, J. Wegener, D. Dasgupta, M. A. Potter, A. C. Schultz, K. Dowsland, N. Jonoska, and J. Miller, editors, Genetic and Evolutionary Computation -- GECCO-2003, volume 2724 of LNCS, pages 1906-1917. Springer-Verlag, 2003.

[822]
P. Thrift. Fuzzy logic synthesis with genetic algorithms. In Lashon B. Booker and Richard K. Belew, editors, Proceedings of 4th international conference on genetic algorithms (ICGA'91), pages 509-513. Morgan Kaufmann, 1991.

[823]
S. Tokinaga and A. B. Whinston. Applying Adaptive Credit Assignment Algorithm for the Learning Classifier System Based upon the Genetic Algorithm. IEICE Transactions on Fundamentals of Electronics Communications and Computer Sciences, VE75A(5):568-577, May 1992.

[824]
Andy Tomlinson and Larry Bull. A Corporate Classifier System. In A. E. Eiben, T. Bäck, M. Shoenauer, and H.-P. Schwefel, editors, Proceedings of the Fifth International Conference on Parallel Problem Solving From Nature -- PPSN V, number 1498 in LNCS, pages 550-559. Springer Verlag, 1998.
Based on the proposals of Wilson and Goldberg we introduce a macro-level evolutionary operator which creates structural links between rules in the ZCS model and thus forms ``corporations'' of rules within the classifier system population. Rule co-dependencies influence both the behaviour of the discovery components of the system and the production system, where a corporation can take control for a number of time-steps. The system is compared to ZCS and also ZCSM in a number of maze environments which include Woods1 and Woods7. The corporate classifier system is shown to be the most suitable design to tackle a range of these types of problems.

[825]
Andy Tomlinson and Larry Bull. A Corporate XCS. In Wu [923], pages 298-305.
Previously we have applied rule linkage to ZCS and shown that the resultant system demonstrates performance improvements over ZCS in a series of sequential tasks, particularly tasks which present ambiguous stimuli to the system. In this paper we show that similar benefits can be gained by applying rule linkage to the more complex XCS. We then show that the benefits of rule-linkage can be increased by further XCS specific modifications to the system's rule-linkage mechanisms.

[826]
Andy Tomlinson and Larry Bull. On Corporate Classifier Systems: Increasing the Benefits of Rule Linkage. In Banzhaf et al. [32], pages 649-656.
Our previous implementation of a Corporate Classifier System (CCS), which introduces rule-linkage to a ZCS-based system, has been shown to demonstrate performance improvements over ZCS in a series of sequential tasks, particularly those which present arbitrary sensory ambiguities to the system. In this paper, the functionality of our CCS is enhanced to provide increased benefits regarding the same class of sequential evaluation tasks.

[827]
Andy Tomlinson and Larry Bull. A zeroth level corporate classifier system. In Wu [923], pages 306-313.
It has long been recognised that increased co-operation amongst the classifiers in a Michigan-style classifier system may resolve some of the established difficulties associated with the design. One approach to this was proposed by Wilson and Goldberg -- the ``corporate'' classifier system. In this paper we implement the ``corporate'' classifier system design, within Wilson's ZCS, in such a way that it complies with their theoretical proposals. In the resultant system , a zeroth-level corporate classifier system, all classifiers initially stand alone but during the course of evolution, a mutation-type operator is used to couple together classifiers by means of structural links. Linked classifiers are considered to represent a corporation, and are treated as a unit by the discovery mechanism of the system. This is achieved by the use of a macro-level evolutionary operator called ``corporate crossover''. In this design the production system remains oblivious to corporations and operates as ZCS. A technique referred to as concept analysis is introduced which is used to clarify the effects of such rule associations, as implemented here, within a Michigan-style classifier system.

[828]
Andy Tomlinson and Larry Bull. A Corporate XCS. In Lanzi et al. [544], pages 194-208.
Previously we have applied rule linkage to ZCS and shown that the resultant system demonstrates performance improvements over ZCS in a series of sequential tasks, particularly tasks which present ambiguous stimuli to the system. In this paper we show that similar benefits can be gained by applying rule linkage to the more complex XCS. We then show that the benefits of rule-linkage can be increased by further XCS specific modifications to the system's rule-linkage mechanisms.

[829]
Andy Tomlinson and Larry Bull. Cxcs: Improvements and corporate generalization. In Lee Spector, Erik D. Goodman, Annie Wu, W.B. Langdon, Hans-Michael Voigt, Mitsuo Gen, Sandip Sen, Marco Dorigo, Shahram Pezeshk, Max H. Garzon, and Edmund Burke, editors, Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2001), pages 966-973, San Francisco, California, USA, 7-11 July 2001. Morgan Kaufmann.

[830]
Andy Tomlinson and Larry Bull. An accuracy-based corporate classifier system. Journal of Soft Computing, 6(3-4):200-215, 2002.

[831]
Andy Tomlinson. Corporate Classifier Systems. PhD thesis, University of the West of England, 1999.

[832]
Dave Toney. Evolutionary Reinforcement Learning of Spoken Dialogue Strategies. PhD thesis, University of Edinburgh, 2007.
From a system developer's perspective, designing a spoken dialogue system can be a time-consuming and difficult process. A developer may spend a lot of time anticipating how a potential user might interact with the system and then deciding on the most appropriate system response. These decisions are encoded in a dialogue strategy, essentially a mapping between anticipated user inputs and appropriate system outputs. To reduce the time and effort associated with developing a dialogue strategy, recent work has concentrated on modelling the development of a dialogue strategy as a sequential decision problem. Using this model, reinforcement learning algorithms have been employed to generate dialogue strategies automatically. These algorithms learn strategies by interacting with simulated users. Some progress has been made with this method but a number of important challenges remain. For instance, relatively little success has been achieved with the large state representations that are typical of real-life systems. Another crucial issue is the time and effort associated with the creation of simulated users. In this thesis, I propose an alternative to existing reinforcement learning methods of dialogue strategy development. More specifically, I explore how XCS, an evolutionary reinforcement learning algorithm, can be used to find dialogue strategies that cover large state spaces. Furthermore, I suggest that hand-coded simulated users are sufficient for the learning of useful dialogue strategies. I argue that the use of evolutionary reinforcement learning and hand-coded simulated users is an effective approach to the rapid development of spoken dialogue strategies. Finally, I substantiate this claim by evaluating a learned strategy with real users. Both the learned strategy and a state-of-the-art hand-coded strategy were integrated into an end-to-end spoken dialogue system. The dialogue system allowed real users to make flight enquiries using a live database for an Edinburgh-based airline. The performance of the learned and hand-coded strategies were compared. The evaluation results show that the learned strategy performs as well as the hand-coded one (81% and 77% task completion respectively) but takes much less time to design (two days instead of two weeks). Moreover, the learned strategy compares favourably with previous user evaluations of learned strategies.

[833]
T.H. Tran, C. Sanza, Y. Duthen, and T.D. Nguyen. XCSF with computed continuous action. In Genetic and evolutionary computation conference (GECCO 2007), pages 1861-1869. ACM, 2007.

[834]
Kwok Ching Tsui and Mark Plumbley. A New Hillclimber for Classifier Systems. In GALESI97, 1997.
Multi-state artificial environments such as mazes represent a class of tasks that can be solved by many different multi-step methods. When different rewards are available in different places of the maze, a problem solver is required to evaluate different positions effectively and remembers the best one. A new hillclimbing strategy for the Michigan style classifier system is suggested which is able to find the shortest path and discarding sub-optimal solutions. Knowledge reuse is also shown to be possible.

[835]
Patrick Tufts. Evolution of a Clustering Scheme for Classifier Systems: Beyond the Bucket Brigade. PhD Thesis proposal. http://www.cs.brandeis.edu/~zippy/papers.htm, 1994.
The Classifier System is a learning mechanism that explores the space of steps leading to a reward. It credits rules leading to a reward through a temporal difference method called a bucket brigade in which each step passes some of its reward to the step that preceded it. The bucket brigade rewards all the steps in a chain, given enough time. However, the classifiers in a population are in competition with each other, so delays in rewarding key steps may result in those steps being crowded out by others. The main hypothesis of this proposal is that the combination of a clustering operator with a classifier system will aid in the formation of long chains. This combination has potential applications for discovering useful variable-length representations for classifier systems, further enhancing their ability to learn, and can be applied to the field of Genetic Programming to discover useful building blocks.

[836]
Patrick Tufts. Dynamic Classifiers: Genetic Programming and Classifier Systems. In E. V. Siegel and J. R. Koza, editors, Working Notes for the AAAI Symposium on Genetic Programming, pages 114-119, MIT, Cambridge, MA, USA, 1995. AAAI.
The Dynamic Classifier System extends the traditional classifier system by replacing its fixed-width ternary representation with Lisp expressions. Genetic programming applied to the classifiers allows the system to discover building blocks in a flexible, fitness directed manner. In this paper, I describe the prior art of problem decomposition using genetic programming and classifier systems. I then show how the proposed system builds on work in these two areas, extending them in a way that provides for flexible representation and fitness directed discovery of useful building blocks.

[837]
Kirk Twardowski. Implementation of a Genetic Algorithm based Associative Classifier System (ACS). In Proceedings International Conference on Tools for Artificial Intelligence, 1990.

[838]
Kirk Twardowski. Credit Assignment for Pole Balancing with Learning Classifier Systems. In Forrest [335], pages 238-245.
Effective reinforcement learning methods are essential for credit assignment in learning classifier systems in order to guide the structural rule-base modifications performed by genetic algorithms. Currently, a number of algorithms have been proposed to fulfill this fundamental need; the most frequently employed being the bucket-brigade algorithm. In this paper, an experimental evaluation of a number of reinforcement learning algorithms for a variety of parameter settings is presented. A simplified learning classifier system is used in order to minimize system complexity in order to isolate the behavior of the reinforcement learning algorithms. The problem domain tackled is that of the control of an unstable dynamic system. It was discovered that exploitation of current information is highly favored over exploration for new information and that a hybrid bucket brigade-backward averaging algorithm produced the fastest convergence to a solution.

[839]
Kirk Twardowski. An Associative Architecture for Genetic Algorithm-Based Machine Learning. Computer, 27(11):27-38, November 1994.
Machine-based learning will eventually be applied to solve real-world problems. Here, an associative architecture teams with hybrid AI algorithms to solve a letter prediction problem with promising results.

[840]
Olgierd Unold and Grzegorz Dabrowski. Use of learning classifier system for inferring natural language grammar. In Tim Kovacs, Xavier LLòra, Keiki Takadama, Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson, editors, Learning Classifier Systems. International Workshops, IWLCS 2003-2005, Revised Selected Papers, volume 4399 of LNCS, pages 17-24. Springer, 2007.

[841]
J. Urzelai, Dario Floreano, Marco Dorigo, and Marco Colombetti. Incremental Robot Shaping. Connection Science, 10(3-4):341-360, 1998.

[842]
J. Urzelai, Dario Floreano, Marco Dorigo, and Marco Colombetti. Incremental Robot Shaping. In Koza et al. [529], pages 832-840.

[843]
Manuel Valenzuela-Rendón and Eduardo Uresti-Charre. A Non-Genetic Algorithm for Multiobjective Optimization. In Bäck [24], pages 658-665.
This paper describes a non-generational genetic algorithm for multiobjective optimisation. The fitness of each individual in the population is calculated incrementally based on the degree in which it is dominated in the Pareto sense, or close to other individuals. The closeness of individuals is measured using a sharing function. The performance of the algorithm presented is compared to previous efforts on three multiobjective problems of growing difficulty. The behavior of each algorithm is analyzed with regard to the visited search space, the quality of the final population attained,and the percentage of non-dominated individuals in the population through time. According to all these performance measures, the algorithm presented clearly outperforms previous efforts based on genetic algorithms.

[844]
Manuel Valenzuela-Rendón. Boolean Analysis of Classifier Sets. In Schaffer [718], pages 351-358.
In this paper a new analysis tool for classifier systems is presented: the Boolean analysis of classifier sets. This tool is applied to determine the minimal classifier sets that perform a general learning task in stimulus-response mode. Modifications to classical Boolean minimization techniques to accommodate for the minimization of default hierarchies are studied. This Boolean analysis is used to determine the relation between the size of a Boolean function and the minimal number of levels required in its minimal default hierarchy. Finally, the concept of parsimony or savings of rules produced by the formation of default hierarchies is discussed.

[845]
Manuel Valenzuela-Rendón. Two analysis tools to describe the operation of classifier systems. PhD thesis, University of Alabama, 1989. Also TCGA technical report 89005.
This document has no abstract.

[846]
Manuel Valenzuela-Rendón. The Fuzzy Classifier System: a Classifier System for Continuously Varying Variables. In Booker and Belew [74], pages 346-353.
This paper presents the fuzzy classifier system which merges the ideas behind classifier systems and fuzzy controllers. The fuzzy classifier system learns by creating fuzzy rules which relate the values of the input variables to internal or output variables. It has credit assignment mechanisms which reassemble those of common classifier systems, but with a fuzzy nature. The fuzzy classifier system employs a genetic algorithm to evolve adequate fuzzy rules. Preliminary results show that the fuzzy classifier system can effectively create fuzzy rules that imitate the behavior of simple static systems.

[847]
Manuel Valenzuela-Rendón. The Fuzzy Classifier System: Motivations and First Results. In Hans-Paul Schwefel and Reinhard Männer, editors, Parallel Problem Solving from Nature (PSSN-1), volume 496 of Lecture Notes in Computer Science, pages 338-342, 1991.

[848]
Manuel Valenzuela-Rendón. Reinforcement learning in the fuzzy classifier system. In Collected Abstracts for the First International Workshop on Learning Classifier System (IWLCS-92) [486]. October 6-8, NASA Johnson Space Center, Houston, Texas.
Paper is an extended abstract

[849]
Manuel Valenzuela-Rendón. Reinforcement learning in the fuzzy classifier system. Expert Systems Applications, 14:237-247, 1998.

[850]
R. Vallim, D. Goldberg, X. Llorà, T. Duque, and A. Carvalho. A new approach for multi-label classification based on default hierarchies and organizational learning. In Proceedings of the Genetic and Evolutionary Computation Conference, Worrkshop Sessions: Learning Classifier Systems, pages 2017-2022, 2003.

[851]
Terry van Belle. A New Approach to Genetic-Based Automatic Feature Discovery. Master's thesis, University of Alberta, 1995. http://www.cs.ualberta.ca/~jonathan/.
Systems which take raw data and categorize them into discrete classes are ubiquitous in computer science, having applications in fields such as vision, expert systems, and game playing. These systems work by extracting features from the data and then combining the values of the features to form a judgement. While much work has been done on ways to automatically combine feature values, the task of automatic discovery of these features is recognized to be much more difficult, and so has become one of the holy grails of machine learning. Classifier systems, an outgrowth of genetic algorithms, seemed a promising approach to automatic feature discovery, but it is difficult to get the full power of the classifier system from existing implementations. This thesis simplifies the classifier system into a variant of the genetic algorithm, called the Population Genetic Algorithm (PGA). PGAs are used to automatically discover features for tic-tac-toe and checkers endgame positions, and these features are automatically combined using Bayesian statistics to classify each position as won, lost, or drawn. The theoretical maximum performance of the PGAs is determined by using an exhaustive enumeration technique to serve as a baseline comparison. The results indicate that while PGAs can be made to perform at near-optimal levels, the optimal solution is insufficient to perfectly classify any of the domains studied.

[852]
Patricia Amancio Vargas, Christiano Lyra Filho, and Fernando J. Von Zuben. On-line approach for loss reduction in electric power distribution networks using learning classifier systems. In Lanzi et al. [546], pages 181-196.

[853]
G. Venturini. SIA: A supervised inductive algorithm with genetic search for learning attributes based concepts. In P.B. Brazdil, editor, ECML-93 - Proc. of the European Conference on Machine Learning, pages 280-296. Springer-Verlag, 1993.

[854]
Gilles Venturini. Apprentissage Adaptatif et Apprentissage Supervisé par Algorithme Génétique. PhD thesis, Université de Paris-Sud., 1994.

[855]
Nickolas Vriend. Self-Organization of Markets: An Example of a Computational Approach. Computational Economics, 8(3):205-231, 1995.
A model of decentralized trade is simulated with firms that produce a given commodity, and consumers who repeatedly wish to purchase one unit of that commodity. Consumers 'shop around', while firms may attract the attention of potential customers by sending information signals and offering good service. The main objective of this paper is to present an example of a computational approach to address the following question: How do self-organized markets emerge in the economy, and what are their characteristics?

[856]
A. Wada, K. Takadama, K. Shimohara, and O. Katai. Learning classifier systems with convergence and generalization. In L. Bull and T. Kovacs, editors, Foundations of learning classifier systems, pages 285-304. Springer, 2005.

[857]
Atsushi Wada, Keiki Takadama, and Katsunori Shimohara. Counter example for q-bucket-brigade under prediction problem. In GECCO Workshops 2005, pages 94-99. ACM Press, 2005.
Aiming to clarify the convergence or divergence conditions for Learning Classifier System (LCS), this paper explores: (1) an extreme condition where the reinforcement process of LCS diverges; and (2) methods to avoid such divergence. Based on our previous work that showed equivalence between LCS's reinforcement process and Reinforcement Learning (RL) with Function approximation (FA) method, we present a counter-example for LCS with Q-bucket-brigade based on the 11-state star problem, a counter-example originally proposed to show the divergence of Q-learning with linear FA. Furthermore, the empirical results applying the counter-example to LCS verified the results predicted from the theory: (1) LCS with Q-bucket-brigade diverged under the prediction problem, where the action selection policy was fixed; and (2) such divergence was avoided by using implicit-bucket-brigade or applying residual gradient algorithm to Q-bucket-brigade.

[858]
Atsushi Wada, Keiki Takadama, and Katsunori Shimohara. Learning classifier system equivalent with reinforcement learning with function approximation. In GECCO Workshops 2005, pages 92-93. ACM Press, 2005.
We present an experimental comparison of the reinforcement process between Learning Classifier System (LCS) and Reinforcement Learning (RL) with function approximation (FA) method, regarding their generalization mechanisms. To validate our previous theoretical analysis that derived equivalence of reinforcement process between LCS and RL, we introduce a simple test environment named Gridworld, which can be applied to both LCS and RL with three different classes of generalization: (1) tabular representation; (2) state aggregation; and (3) linear approximation. From the simulation experiments comparing LCS with its GA-inactivated and corresponding RL method, all the cases regarding the class of generalization showed identical results with the criteria of performance and temporal difference (TD) error, thereby verifying the equivalence predicted from the theory.

[859]
Atsushi Wada, Keiki Takadama, and Katsunori Shimohara. Counter example for q-bucket-brigade under prediction problem. In Tim Kovacs, Xavier LLòra, Keiki Takadama, Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson, editors, Learning Classifier Systems. International Workshops, IWLCS 2003-2005, Revised Selected Papers, volume 4399 of LNCS, pages 128-143. Springer, 2007.

[860]
Atsushi Wada, Keiki Takadama, Katsunori Shimohara, and Osamu Katai. Analyzing parameter sensitivity and classifier representations for real-valued xcs. In Tim Kovacs, Xavier LLòra, Keiki Takadama, Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson, editors, Learning Classifier Systems. International Workshops, IWLCS 2003-2005, Revised Selected Papers, volume 4399 of LNCS, pages 1-16. Springer, 2007.

[861]
David Walter and Chilukuri K. Mohan. ClaDia: A Fuzzy Classifier System for Disease Diagnosis. In Proceedings of the 2000 Congress on Evolutionary Computation (CEC00) [185], pages 1429-1435.
This paper describes ClaDia, a learning classifier system applied to the Wisconsin breast cancer data set, using a fuzzy representation of the rules, a median-based fuzzy combination rule, and separate subpopulations for each class. The system achieves a classification rate of over 90%, for many sets of system parameter values.

[862]
L. A. Wang. Classifier System Learning of the Boolean Multiplexer Function. Master's thesis, Computer Science Department, University of Tennessee, Knoxville, TN, 1990.

[863]
Gerhard Weiss. Action-oriented learning in classifier systems. Technical Report FKI-158-91, Technical Univ. München (TUM), 1991.

[864]
Gerhard Weiss. The Action-Oriented Bucket Brigade. Technical Report FKI-156-91, Technical Univ. München (TUM), 1991.

[865]
Gerhard Weiss. Hierarchical chunking in classifier systems. In Proceedings of the 12th National Conference on Artificial Intelligence, pages 1335-1340. AAAI Press/MIT Press, 1994.

[866]
Gerhard Weiss. Learning by chunking in reactive classifier systems. Technical report, Technical Univ. München (TUM), 1994.

[867]
Gerhard Weiss. The locality/globality dilemma in classifier systems and an approach to its solution. Technical Report FKI-187-94, Technical Univ. München (TUM), 1994.

[868]
Gerhard Weiss. An action-oriented perspective of learning in classifier systems. Journal of Experimental and Theoretical Artificial Intelligence, 8:43-62, 1996.
Classifier systems constitute a general model of low-level rule-based systems that are capable of environmental interaction and learning. A central characteristic and drawback of the traditional approaches to learning in such systems is that they exclusively work on the rule level, without taking into consideration that the individual rules possess a very complex activity behavior. This article investigates an alternative, action-oriented perspective of learning in classifier systems which does not suffer from this drawback. According to this perspective learning is realized on the finer action level instead of the coarser rule level. Comparative theoretical and experimental results are presented that show the advantages of the action-oriented over the traditional perspective.

[869]
Thomas H. Westerdale. The bucket brigade is not genetic. In Grefenstette [389], pages 45-59.

[870]
Thomas H. Westerdale. A Reward Scheme for Production Systems with Overlapping Conflict Sets. IEEE Transactions on Systems, Man and Cybernetics, SMC-16(3):369-383, 1986.

[871]
Thomas H. Westerdale. Altruism in the bucket brigade. In Grefenstette [391], pages 22-26.

[872]
Thomas H. Westerdale. A Defence of the Bucket Brigade. In Schaffer [718], pages 282-290.
The bucket brigade does indeed suffer from detrimental biases. Profit sharing can avoid these only if productions are properly penalized for eligibility, and more importantly, if the span over which profit is shared is long enough. Such a long span introduces much sampling noise, with its often unacceptable danger of premature convergence. The bucket brigade is designed to reduce this noise.

[873]
Thomas H. Westerdale. Quasimorphisms or Queasymorphisms? Modelling Finite Automaton Environments. In Rawlins [674], pages 128-147.
The paper examines models that are homomorphic images of the first component of a particular two component cascade decomposition of the environment. The bucket brigade is used to estimate model state values. The discussion is limited to finite automaton environments whose successive input symbols are selected by the system probabilistically, with independent probabilities, according to a probability distribution over the input symbols.

[874]
Thomas H. Westerdale. Redundant Classifiers and Prokaryote Genomes. In Booker and Belew [74], pages 354-360.
Redundant classifiers waste space. This paper suggests an approach to the problem of classifier redundancy based on the cost of gene replication. The approach stems from the view that reduction and simplification are the essence of the evolutionary creative process, and that the most advanced organisms are Prokaryotes, not Eukaryotes.

[875]
Thomas H. Westerdale. Classifier Systems - No Wonder They Don't Work. In Koza et al. [528], pages 529-537.
Classifier system formalism provides valuable conceptual test beds that help us come to grips with fundamental problems of credit assignment in learning systems. Here we discuss the horizon problem, the sampling noise problem, the problem of removing redundancy to save space, and the freeloader problem, this last being a collection of problems including the problem of ill formed conditions. Classifier systems can help provide answers. But until we have some answers, experimental systems will remain flawed.

[876]
Thomas H. Westerdale. An Approach to Credit Assignment in Classifier Systems. Complexity, 4(2), 1999.

[877]
Thomas H. Westerdale. Wilson's Error Measurement and the Markov Property -- Identifying Detrimental Classifiers. In Wu [923], pages 314-321.
Wilson's error measurement is an important conceptual tool in the study of classifier system reward schemes. It tests classifiers to see where the Markov property detrimentally fails. Incorporating this test directly into a reward scheme involves difficulties. But the use of the error measurement in analysis, and indeed in experimentation, should help advance our understanding of credit assignment issues.

[878]
Darrell Whitely, David Goldberg, Erick Cantú-Paz, Lee Spector, Ian Parmee, and Hans-Georg Beyer, editors. Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2000). Morgan Kaufmann, 2000.

[879]
Jason R. Wilcox. Organizational Learning within a Learning Classifier System. Master's thesis, University of Illinois, 1995. Also Technical Report No. 95003 IlliGAL.
This thesis recasts the debate between Michigan-style and Pitt-style classifier systems to a debate on appropriately sizing organizations within a learning classifier system. Motivated by the economic study of transaction costs, an organizational classifier system (OCS) combining explicit use of multiple reputation values and organization sizing operators better distinguishes parasitic (less than optimal) classifiers than a simple classifier system (SCS). The results show that by building a system that autonomously adjusts the degree of individual to collective behavior, it is possible for it to be both efficient and resilient to problem difficulty.

[880]
Stewart W. Wilson and David E. Goldberg. A Critical Review of Classifier Systems. In Schaffer [718], pages 244-255. http://prediction-dynamics.com/.
The current state of classifier system development is examined with emphasis on challenges and unsolved problems. Suggestions related to the bucket-brigade architecture, the mechanics of bidding and payments, and classifier syntax follow a review of past research.

[881]
Stewart W. Wilson. Aubert processing and intelligent vision. Technical report, Polaroid Corporation, 1981.

[882]
Stewart W. Wilson. On the retino-cortical mapping. International Journal of Man-Machine Studies, 18:361-389, 1983.
Based on Hubel & Wiesel's physiological findings on the projection from retina to cortex, a schematic model of that stage of visual processing is constructed and its properties investigated. The projection or mapping appears to carry out an automatic ``normalization of description'' for the same object independent of retinal image size. This property suggests new concepts regarding (1) contrast sensitivity, (2) the nature and role of indirect vision, (3) the role of eye movements and (4) the recognition of patterns and the analysis of scenes.

[883]
Stewart W. Wilson. Adaptive ``cortical'' pattern recognition. In Grefenstette [389], pages 188-196.
It is shown that a certain model of the primate retino-cortical mapping ``sees'' all centered objects with the same ``object-resolution'', or number of distinct signals, independent of apparent size. In an artificial system, this property would permit recognition of patterns using templates in a cortex-like space. It is suggested that with an adaptive production system such as Holland's classifier system, the recognition process could be made self-organizing.

[884]
Stewart W. Wilson. Knowledge Growth in an Artificial Animal. In Grefenstette [389], pages 16-23. Also appeared in Proceedings of the 4th Yale.
Results are presented of experiments with a simple artificial animal model acting in a simulated environment containing food and other objects. Procedures within the model that lead to improved performance and perceptual generalization are discussed. The model is designed in the light of an explicit definition of intelligence which appears to apply to all animal life. It is suggested that study of artificial animal models of increasing complexity would contribute to understanding of natural and artificial intelligence.

[885]
Stewart W. Wilson. Knowledge Growth in an Artificial Animal. In Proceedings of the 4th Yale Workshop on Applications of Adaptive Systems Theory, pages 98-104, 1985.

[886]
Stewart W. Wilson. Classifier System Learning of a Boolean Function. Technical Report RIS 27r, The Rowland Institute for Science, 1986.
A Simplified classifier system was given the task of learning a relatively difficult boolean function drawn from the machine learning literature. The system solved the problem in times comparable to or less than times required by a network method. Classifiers present in the final populations corresponded closely to terms of an efficient boolean representation of the solution. Achievement of the results depended on a selection regime that emphasized classifiers which were both general and accurate. The theoretically predicted superiority of the crossover genetic operator to the point mutation operator was observed. Most experiments used a fixed crossover rate, but in one series the system itself advantageously controlled the rate based on an environment-independent definition of classifier system entropy.

[887]
Stewart W. Wilson. Knowledge Growth in an Artificial Animal. In K. S. Narenda, editor, Adaptive and learning systems: Theory and applications, pages 255-264. Plenum Press: New York, 1986.

[888]
Stewart W. Wilson. Classifier Systems and the Animat Problem. Machine Learning, 2:199-228, 1987. Also Research Memo RIS-36r, the Rowland Institute for Science, Cambridge, MA, 1986.
This paper characterizes and investigates, from the perspective of machine learning and, particularly, classifier systems, the learning problem faced by animals and autonomous robots (here collectively termed animats). We suggest that, to survive in their environments, animats must in effect learn multiple disjunctive concepts incrementally under payoff (needs-satisfying) feedback. A review of machine learning techniques indicates that most relax at least one of these constraints. In theory, classifier systems satisfy the constraints, but tests have been limited. We show how the standard classifier system model applies to the animat learning problem. Then, in the experimental part of the paper, we specialize the model and test it in a problem environment satisfying the constraints and consisting of a difficult, disjunctive Boolean function drawn from the machine learning literature. Results include: learning the function in significantly fewer trials than a neural-network method; learning under payoff regimes that include both noisy payoff and partial reward for suboptimal performance; demonstration, in a classifier system, of a theoretically predicted property of genetic algorithms: the superiority of crossovers to point mutations; and automatic control of variation (search) rate based on system entropy. We conclude that the results support the classifier system approach to the animat problem, but suggest work aimed at the emergence of behavioral hierarchies of classifiers to offset slower learning rates in larger problems.

[889]
Stewart W. Wilson. Hierarchical Credit Allocation in a Classifier System. In Proceedings Tenth International Joint Conference on AI (IJCAI-87), pages 217-220. Morgan Kaufmann Publishers, 1987. Also Research Memo RIS-37r, the Rowland Institute for Science, Cambridge, MA, 1986.
Learning systems which engage in sequential activity face the problem of properly allocating credit to steps or actions which make possible later steps that result in environmental payoff. In the classifier systems studied by Holland and others, credit is allocated by means of a ``bucket-brigade'' algorithm through which, over time, environmental payoff in effect flows back to classifiers which take early, stage-setting actions. The algorithm has advantages of simplicity and locality, but may not adequately reinforce long action sequences. We suggest an alternative form for the algorithm and the system's operating principles designed to induce behavioral hierarchies in which modularity of the hierarchy would keep all bucket-brigade chains short, thus more reinforceable and more rapidly learned, but overall action sequences could be long.

[890]
Stewart W. Wilson. Quasi-Darwinian Learning in a Classifier System. In Proceedings of the Fourth International Workshop on Machine Learning, pages 59-65. Morgan Kaufmann, 1987.
Classifier systems (Holland, 1986) have a distinct Darwinian flavor, and in this respect contrast sharply with most other learning systems. In this paper we bring out various aspects of the contrast, and provide an example of classifier system learning which illustrates its quasi-Darwinian operation.

[891]
Stewart W. Wilson. The genetic algorithm and biological development. In Grefenstette [391], pages 247-251.

[892]
Stewart W. Wilson. Bid Competition and Specificity Reconsidered. Complex Systems, 2(6):705-723, 1988.
Experiments were conducted with respect to two classifier system mechanisms: the bid competition and the use of classifier specificity in bidding and payments. The experiments employed a simplified classifier system and so may not accurately reflect the behavior of the standard system. Nevertheless, the results indicated that, in general, (1) specificity should not be factored into amounts deducted from a classifier's strength, (2) the bid competition does not improve performance and does not encourage default hierarchies, and (3) default hierarchies will form under a somewhat different algorithm than the standard one.

[893]
Stewart W. Wilson. Hierarchical Credit Assignment in a Classifier System. In M. Elzas, T. Oren, and B. P. Zeigler, editors, Modelling and Simulation Methodology: Knowledge Systems Paradigms. North Holland, 1988.

[894]
Stewart W. Wilson. Hierarchical Credit Allocation in a Classifier System. In Davis [234], pages 104-115.
Learning systems which engage in sequential activity face the problem of properly allocating credit to steps or actions which make possible later steps that result in environmental payoff. In the classifier systems studied by Holland and others, credit is allocated by means of a ``bucket-brigade'' algorithm through which, over time, environmental payoff in effect flows back to classifiers which take early, stage-setting actions. The algorithm has advantages of simplicity and locality, but may not adequately reinforce long action sequences. We suggest an alternative form for the algorithm and the system's operating principles designed to induce behavioral hierarchies in which modularity of the hierarchy would keep all bucket-brigade chains short, thus more reinforceable and more rapidly learned, but overall action sequences could be long.

[895]
Stewart W. Wilson. Hierarchical credit allocation in a classifier system. In M. S. Elzas, T. I. Oren, and B. P. Zeigler, editors, Modelling and simulation methodology, pages 351-357. North-Holland: New York, 1989.

[896]
Stewart W. Wilson. The Genetic Algorithm and Simulated Evolution. In Chris Langton, editor, Artificial Life: Proceedings of an Interdisciplinary Workshop on the Synthesis and Simulation of Living Systems, volume VI of Santa Fe Institute Studies in the Sciences of Complexity. Addison-Wesley: Reading, MA, 1989.
A scheme is described for simulating the evolution of multicellular systems. The scheme is based on a representation for biological development in which the genotypes are sets of production-like growth rules that are executed to produce cell aggregates-the phenotypes. Evolution of populations, through phenotype selection and genotype variation, occurs according to the method of the genetic algorithm. Some examples of the development representation in 1-dimensional creatures are given.

[897]
Stewart W. Wilson. Perceptron redux: Emergence of structure. In Special issue of Physica D (Vol. 42) [292], pages 249-256. Republished in Emergent Computation, S. Forrest (ed.), MIT Press/Bradford Books.
Perceptrons were evolved that computed a rather difficult nonlinear Boolean function. The results with this early and basic form of emergent computation suggested that when genetic search is applied to its structure, a perceptron can learn more complex tasks than is sometimes supposed. The results also suggested, in the light of recent work on classifier systems, that to hasten the emergence of an emergent computation it is desirable to provide evaluative feedback at a level as close as possible to that of the constituent local computations.

[898]
Stewart W. Wilson. The Animat Path to AI. In Meyer and Wilson [613], pages 15-21. http://prediction-dynamics.com/.
A research methodology is proposed for understanding intelligence through simulation of artificial animals (``animats'') in progressively more challenging environments while retaining characteristics of holism, pragmatism, perception, categorization, and adaptation that are often underrepresented in standard AI approaches to intelligence. It is suggested that basic elements of the methodology should include a theory/taxonomy of environments by which they can be ordered in difficulty-one is offered-and a theory of animat efficiency. It is also suggested that the methodology offers a new approach to the problem of perception.

[899]
Stewart W. Wilson. Classifier System mapping of real vectors. In Collected Abstracts for the First International Workshop on Learning Classifier System (IWLCS-92) [486]. October 6-8, NASA Johnson Space Center, Houston, Texas.
Paper is an extended abstract

[900]
Stewart W. Wilson. Toward a GA solution of the discovery problem. In Collected Abstracts for the First International Workshop on Learning Classifier System (IWLCS-92) [486]. October 6-8, NASA Johnson Space Center, Houston, Texas.
Paper is an extended abstract

[901]
Stewart W. Wilson. ZCS: A zeroth level classifier system. Evolutionary Computation, 2(1):1-18, 1994. http://prediction-dynamics.com/.
A basic classifier system, ZCS, is presented that keeps much of Holland's original framework but simplifies it to increase understandability and performance. ZCS's relation to Q-learning is brought out, and their performances compared in environments of two difficulty levels. Extensions to ZCS are proposed for temporary memory, better action selection, more efficient use of the genetic algorithm, and more general classifier representation

[902]
Stewart W. Wilson. Classifier Fitness Based on Accuracy. Evolutionary Computation, 3(2):149-175, 1995. http://prediction-dynamics.com/.
In many classifier systems, the classifier strength parameter serves as a predictor of future payoff and as the classifier's fitness for the genetic algorithm. We investigate a classifier system, XCS, in which each classifier maintains a prediction of expected payoff, but the classifier's fitness is given by a measure of the prediction's accuracy. The system executes the genetic algorithm in niches defined by the match sets, instead of panmictically. These aspects of XCS result in its population tending to form a complete and accurate mapping X x A -> P from inputs and actions to payoff predictions. Further, XCS tends to evolve classifiers that are maximally general, subject to an accuracy criterion. Besides introducing a new direction for classifier system research, these properties of XCS make it suitable for a wide range of reinforcement learning situations where generalization over states is desirable.

[903]
Stewart W. Wilson. Explore/exploit strategies in autonomy. In Maes et al. [595], pages 325-332.
Within a reinforcement learning framework, ten strategies for autonomous control of the explore/exploit decision are reviewed, with observations from initial experiments on four of them. Control based on prediction error or its rate of change appears promising. Connections are made with explore/exploit work by Holland (1975), Thrun (1992), and Schmidhuber (1995).

[904]
Stewart W. Wilson. Generalization in XCS. Unpublished contribution to the ICML '96 Workshop on Evolutionary Computing and Machine Learning. http://prediction-dynamics.com/, 1996.

[905]
Stewart W. Wilson. Generalization in evolutionary learning. Presented at the Fourth European Conference on Artificial Life (ECAL97), Brighton, UK, July 27-31. http://prediction-dynamics.com/, 1997.

[906]
Stewart W. Wilson. Generalization in the XCS classifier system. In Koza et al. [529], pages 665-674. http://prediction-dynamics.com/.
This paper studies two changes to XCS, a classifier system in which fitness is based on prediction accuracy and the genetic algorithm takes place in environmental niches. The changes were aimed at increasing XCS's tendency to evolve accurate, maximally general classifiers and were tested on previously employed ``woods'' and multiplexer tasks. Together the changes bring XCS close to evolving populations whose high-fitness classifiers form a near-minimal, accurate, maximally general cover of the input and action product space. In addition, results on the multiplexer, a difficult categorization task, suggest that XCS's learning complexity is polynomial in the input length and thus may avoid the ``curse of dimensionality'', a notorious barrier to scale-up. A comparison between XCS and genetic programming in solving the 6-multiplexer suggests that XCS's learning rate is about three orders of magnitude faster in terms of the number of input instances processed.

[907]
Stewart W. Wilson. Get real! XCS with continuous-valued inputs. In L. Booker, Stephanie Forrest, M. Mitchell, and Rick L. Riolo, editors, Festschrift in Honor of John H. Holland, pages 111-121. Center for the Study of Complex Systems, 1999. http://prediction-dynamics.com/.
Classifier systems have traditionally taken binary strings as inputs, yet in many real problems such as data inference, the inputs have real components. A modified XCS classifier system is described that learns a non-linear real-vector classification task.

[908]
Stewart W. Wilson. State of XCS classifier system research. In Wu [923], pages 322-334. Also Technical Report 99.1.1, Prediction Dynamics, Concord MA. http://prediction-dynamics.com/.
XCS is a new kind of learning classifier system that differs from the traditional one primarily in its definition of classifier fitness and its relation to contemporary reinforcement learning. Advantages of XCS include improved performance and an ability to form accurate maximal generalizations. This paper reviews recent research on XCS with respect to representation, predictive modelling, internal state, noise, and underlying theory and technique. A notation for environmental regularities is introduced.

[909]
Stewart W. Wilson. Get Real! XCS with Continuous-Valued Inputs. In Lanzi et al. [544], pages 209-219.
Classifier systems have traditionally taken binary strings as inputs, yet in many real problems such as data inference, the inputs have real components. A modified XCS classifier system is described that learns a non-linear real-vector classification task.

[910]
Stewart W. Wilson. Mining Oblique Data with XCS. In Proceedings of the International Workshop on Learning Classifier Systems (IWLCS-2000), in the Joint Workshops of SAB 2000 and PPSN 2000 [487]. Extended abstract.
Paper is an extended abstract

[911]
Stewart W. Wilson. Mining Oblique Data with XCS. Technical Report 2000028, University of Illinois at Urbana-Champaign, 2000.
The classifier system XCS was investigated for data mining applications where the dataset discrimination surface (DS) is generally oblique to the attribute axes. Despite the classifiers' hyper-rectangular predicates, XCS reached 100% performance on synthetic problems with diagonal DS's and, in a train/test experiment, competitive performance on the Wisconsin Breast Cancer dataset. Final classifiers in an extended WBC learning run were interpretable to suggest dependencies on one or a few attributes. For data mining of numeric datasets with partially oblique discrimination surfaces, XCS shows promise from both performance and pattern discovery viewpoints.

[912]
Stewart W. Wilson. State of XCS Classifier System Research. In Lanzi et al. [544], pages 63-82.
XCS is a new kind of learning classifier system that differs from the traditional kind primarily in its definition of classifier fitness and its relation to contemporary reinforcement learning. Advantages of XCS include improved performance and an ability to form accurate maximal generalizations. This paper reviews recent research on XCS with respect to representation, internal state, predictive modelling, noise, and underlying theory and technique. A notation for environmental regularities is introduced.

[913]
S. W. Wilson. Mining oblique data with xcs. In P.L. Lanzi, W. Stolzmann, and S.W. Wilson, editors, Advances in learning classifier systems, third international workshop, IWLCS 2000, volume 1996 of LNCS, pages 158-176. Springer, 2001.

[914]
Stewart W. Wilson. Function approximation with a classifier system. In Lee Spector, Erik D. Goodman, Annie Wu, W.B. Langdon, Hans-Michael Voigt, Mitsuo Gen, Sandip Sen, Marco Dorigo, Shahram Pezeshk, Max H. Garzon, and Edmund Burke, editors, Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2001), pages 974-981, San Francisco, California, USA, 7-11 July 2001. Morgan Kaufmann.

[915]
Stewart W. Wilson. Classifiers that approximate functions. Natural Computing, 1(2-3):211-234, 2002.

[916]
Stewart W. Wilson. Compact rulesets from xcsi. In Lanzi et al. [546], pages 196-208.

[917]
Stewart W. Wilson. Three architectures for continuous action. In Tim Kovacs, Xavier LLòra, Keiki Takadama, Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson, editors, Learning Classifier Systems. International Workshops, IWLCS 2003-2005, Revised Selected Papers, volume 4399 of LNCS, pages 239-257. Springer, 2007.

[918]
Stewart W. Wilson. Classifier conditions using gene expression programming. In Jaume Bacardit, Ester Bernadó-Mansilla, Martin Butz, Tim Kovacs, Xavier Llorà, and Keiki Takadama, editors, Learning Classifier Systems. 10th and 11th International Workshops (2006-2007), volume 4998/2008 of Lecture Notes in Computer Science, pages 206-217. Springer, 2008.
The classifier system XCSF was modified to use gene expression programming for the evolution and functioning of the classifier conditions. The aim was to fit environmental regularities better than is typically possible with conventional rectilinear conditions. An initial experiment approximating a nonlinear oblique environment showed excellent fit to the regularities.

[919]
Mark S. Withall, Chris J. Hinde, and Roger G. Stone. Evolving readable Perl. In W. B. Langdon, E. Cantú-Paz, K. Mathias, R. Roy, D. Davis, R. Poli, K. Balakrishnan, V. Honavar, G. Rudolph, J. Wegener, L. Bull, M. A. Potter, A. C. Schultz, J. F. Miller, E. Burke, and N. Jonoska, editors, GECCO 2002: Proceedings of the Genetic and Evolutionary Computation Conference. Morgan Kaufmann Publishers, 2002.

[920]
M.L. Wong and K.S. Leung. Data mining using grammar based genetic programming and applications. Kluwer, 2000.

[921]
Ian Wright. Reinforcement Learning and Animat Emotions. In Maes et al. [595], pages 272-281.
Emotional states, such as happiness or sadness, pose particular problems for information processing theories of mind. Hedonic components of states, unlike cognitive components, lack representational content. Research within Artificial Life, in particular the investigation of adaptive agent architectures, provides insights into the dynamic relationship between motivation, the ability of control sub-states to gain access to limited processing resources, and prototype emotional states. Holland's learning classifier system provides a concrete example of this relationship, demonstrating simple `emotion-like' states, much as a thermostat demonstrates simple `belief-like' and `desire-like' states. This leads to the conclusion that valency, a particular form of pleasure or displeasure, is a self-monitored process of credit-assignment. The importance of the movement of a domain-independent representation of utility within adaptive architectures is stressed. Existing information processing theories of emotion can be enriched by a `circulation of value' design hypothesis. Implications for the development of emotional animats are considered.

[922]
Ian Wright. Reinforcement learning and animat emotions. Technical Report CSRP-96-4, School of Computer Science. University of Birmingham, 1996. ftp://ftp.cs.bham.ac.uk/pub/tech-reports/1996/CSRP-96-04.ps.gz.
Emotional states, such as happiness or sadness, pose particular problems for information processing theories of mind. Hedonic components of states, unlike cognitive components, lack representational content. Research within Artificial Life, in particular the investigation of adaptive agent architectures, provides insights into the dynamic relationship between motivation, the ability of control sub-states to gain access to limited processing resources, and prototype emotional states. Holland's learning classifier system provides a concrete example of this relationship, demonstrating simple `emotion-like' states, much as a thermostat demonstrates simple `belief-like' and `desire-like' states. This leads to the conclusion that valency, a particular form of pleasure or displeasure, is a self-monitored process of credit-assignment. The importance of the movement of a domain-independent representation of utility within adaptive architectures is stressed. Existing information processing theories of emotion can be enriched by a `circulation of value' design hypothesis. Implications for the development of emotional animats are considered.

[923]
Annie S. Wu, editor. Proceedings of the 1999 Genetic and Evolutionary Computation Conference Workshop Program, 1999.

[924]
David Wyatt and Larry Bull. A memtic learning classifier system for for describing continuous-valued problem spaces. In William E. Hart, N. Krasnogor, and J.E. Smith, editors, Recent Advances in Memetic Algorithms, Studies in Fuzziness and Soft Computing, pages 355-396. Springer, 2004.

[925]
Kumara Sastry Xavier Llorá and David E. Goldberg. Binary rule encoding schemes: A study using the compact classifier system. In Tim Kovacs, Xavier LLòra, Keiki Takadama, Pier Luca Lanzi, Wolfgang Stolzmann, and Stewart W. Wilson, editors, Learning Classifier Systems. International Workshops, IWLCS 2003-2005, Revised Selected Papers, volume 4399 of LNCS, pages 40-58. Springer, 2007.

[926]
Derek F. Yates and Andrew Fairley. An Investigation into Possible Causes of, and Solutions to, Rule Strength Distortion Due to the Bucket Brigade Algorithm. In Forrest [335], pages 246-253.

[927]
Derek F. Yates and Andrew Fairley. Evolutionary Stability in Simple Classifier Systems. In Fogarty [324], pages 28-37.
In this paper, the relatively new branch of mathematics known as Evolutionary Game Theory is proposed as a potentially useful tool when seeking to resolve certain of the more global, unanswered questions related to classifier systems. In particular, it is proved that, under certain mild assumptions, the performance of a classifier system plans will, if the Bucket Brigade Algorithm is adopted, conform to what is referred to as `an evolutionary stable state'. A simple example is also provided to confirm the theoretical findings.

[928]
Takahiro Yoshimi and Toshiharu Taura. Hierarchical Classifier System Based on the Concept of Viewpoint. In Koza et al. [529], pages 675-678.

[929]
Takahiro Yoshimi and Toshiharu Taura. A Computational Model of a Viewpoint-Forming Process in a Hierarchical Classifier System. In Banzhaf et al. [32], pages 758-766.
In an environment where input information to machine learning (ML) systems using production rules has many ``properties'' and the amount is huge enough, the authors aim for an ML systems with effective performance in finding solutions. When finding solutions, all of the properties of the input information are not always required. Therefore, the authors assume that a certain mechanism which can select specific properties to be focused on will contribute to this purpose. For the realization and discussion of this mechanism, the authors have focused on the Classifier System (CS) which has more advantages than other ML systems. From the author's point of view, operation processes in the CS are thought to involve this mechanism. However, the CS also involves such ``duality'' that both the optimization processes of rules for solution finding and the abstraction processes of input information are in a single process, which may lead to problems. In this paper, the authors propose a computational model in which these two processes are explicitly separated. The key concept of the proposed model is the Viewpoint-Forming Process for the purpose of using rules for selecting properties to be focused on. This is separate from the standard rules for finding solutions. A computer system is developed to evaluate the utility of this model. The results acquired by applying the model to an example problem are reported here.

[930]
Z.V. Zatuchna and A.J. Bagnall. Classifier system: Self-adjusting vs. gradual approach. In Proceedings of the 2005 Congress on Evolutionary Computation. IEEE, 2005.

[931]
Z.V. Zatuchna and A.J. Bagnall. Modelling of temperament in an associative reinforcement learning agent. In T. Kovacs and J.A.R. Marshall, editors, AISB'06: Adaptation in Artificial and Biological Systems. AISB, 2006.

[932]
Z.V. Zatuchna and A.J. Bagnall. A reinforcement learning agent with associative perception. In T. Kovacs and J.A.R. Marshall, editors, AISB'06: Adaptation in Artificial and Biological Systems. AISB, 2006.

[933]
Z.V. Zatuchna and A.J. Bagnall. Towards the axioms of consciousness: Modelling the rat in a maze. In T. Kovacs and J.A.R. Marshall, editors, AISB'06: Adaptation in Artificial and Biological Systems. AISB, 2006.

[934]
Z.V. Zatuchna. AgentP model: Learning Classifer System with Associative Perception. In 8th Parallel Problem Solving from Nature International Conference (PPSN VIII), pages 1172-1182, 2004.

[935]
Zhanna V. Zatuchna. AgentP: a learning classifier system with associative perception in maze environments. PhD thesis, University of East Anglia, 2005.
First we introduce new metrics for classifying the complexity of mazes based on agent-independent and agent-dependent characteristics of maze environments. We analyze 50 mazes used in the literature by the metrics and then introduce 351 new maze environments, including 271 aliasing mazes of increased difficulty. The purpose of preparing the extensive set of maze environments is to provide a suitable evaluation environment for alternative learning agent architectures. To fulfil our second goal we analyze the major learning theories, design the psychological model of Associative Perception Learning, integrate it into the Reinforcement Learning framework and define a new Learning Classifier System (LCS), AgentP, that utilizes explicitly imprinted images of the environment states. AgentP is designed specifically to find the shortest route through aliasing mazes with rewards only on transitions to terminal states. Such mazes contain the areas that look alike for a learning agent but may be associated with different optimal actions. The mazes represent a form of Partially Observable Markov Decision Processes (POMDP). Unlike many other LCS, AgentP does not generalize over the states. It learns a one-step transition model of the environment and uses two deterministic heuristics. AgentP has a rule structure similar to Anticipatory Classifier Systems. However, unlike them, AgentP perceives consecutive environmental states not only as a cause-effect time vector, but also as a single perceptive image, which is compared with previously memorized images for differentiation purposes. As a result AgentP is able to recognize aliasing in both the initial and resulting environment states, while ACS is meant to recognize aliasing in the initial state only. Each classifier in AgentP is supplemented with an ID system for a refined differentiation of aliasing squares. AgentP uses a distance-based reinforcement process where the expected difference between two successive learning coefficients remains the same with increased distance to food. It eliminates the disadvantages associated with the behaviour of the Q-learning based reinforcement procedure, commonly employed by LCS, in long-distanced mazes. The distance-based reinforcement procedure introduces certain limitations as AgentP is only able to handle specific kinds of reward function. The environment should be discrete and the agent is not able to operate on multi-motivational tasks. However, the reinforcement procedure in its present form provides simple and reliable test facilities for our main purpose, development of the operators for refined differentiation of aliasing squares, while the limitation can be overcome in future versions of AgentP when it is necessary. While experimenting with two versions of AgentP, we discover the phenomenon of aliasing clones, i.e. aliasing conglomerates of a similar graphic pattern that include more than two aliasing states and are located in different areas of the maze. We investigate the impact which makes the presence of aliasing clones in a maze on the ability of AgentP to solve mazes. We find that AgentP is able to solve optimally extensive mazes with dozens of aliasing squares and numerous aliasing conglomerates, provided they are free from aliasing clones. At the same time, for a maze containing at least three aliasing states that are grouped into an aliasing clone, the risk of an error for both versions of AgentP becomes non zero. We analyze the performance of AgentP in detail and show that it is able to solve optimally the majority of aliasing mazes used in the experiments and may be performing better than other LCS agents. We then discuss the potential of the learning model, possible improvements into the agent's structure and the most promising approaches to future research.

[936]
Zhaohua Zhang, Stan Franklin, and Dipankar Dasgupta. Metacognition in Software Agents Using Classifier Systems. In AAAI-98. Proceedings of the Fifteenth National Conference on Artificial Intelligence, pages 83-88, Madison (WI), 1998. AAAI-Press and MIT Press.

[937]
Hayong Harry Zhou and John J. Grefenstette. Learning by Analogy in Genetic Classifier Systems. In Schaffer [718], pages 291-297.
This paper presents an extension to the classifier system model that provides mechanisms to store solutions to learned tasks in a long term memory, to match the stored units against new problems as they arise, and to use the stored knowledge to learn new tasks in an increasingly efficient manner. The extended model has been implemented in a system called CSM (Classifier System with Memory). Experimental results with CSM demonstrate the benefits of learning by analogy in a robot navigation task domain and show significant improvements compared with the current classifier system model.

[938]
Hayong Harry Zhou. Classifier systems with long term memory. In Grefenstette [389], pages 178-182.

[939]
Hayong Harry Zhou. CSM: A genetic classifier system with memory for learning by analogy. PhD thesis, Department of Computer Science, Vanderbilt University, Nashville, TN, 1987.

[940]
Hayong Harry Zhou. CSM: A Computational Model of Cumulative Learning. Machine Learning, 5(4):383-406, 1990.

[941]
Raed Abu Zitar and Mohammad H. Hassoun. Regulator Control via Genetic Search Assisted Reinforcement. In Forrest [335], pages 254-263.