Task Decomposition through Competition in A

A novel modular connectionist architecture is presented in which the networks composing the architecture compete to learn the training patterns. An outcome of the competition is that di erent networks learn di erent training patterns and, thus, learn to compute di erent functions. The architecture performs task decomposition in the sense that it learns to partition a task into two or more functionally independent tasks and allocates distinct networks to learn each task. In addition, the architecture tends to allocate to each task the network whose topology is most appropriate to that task. The architecture's performance on \what" and \where" vision tasks is presented and compared with the performance of two multi{layer networks. Finally, it is noted that function decomposition is an underconstrained problem and, thus, di erent modular architectures may decompose a function in di erent ways. We argue that a desirable decomposition can be achieved if the architecture is suitably restricted in the types of functions that it can compute. Appropriate restrictions can be found through the application of domain knowledge. A strength of the modular architecture is that its structure is well{suited for incorporating domain knowledge.

[1]  P. Werbos,et al.  Beyond Regression : "New Tools for Prediction and Analysis in the Behavioral Sciences , 1974 .

[2]  Peter E. Hart,et al.  Pattern classification and scene analysis , 1974, A Wiley-Interscience publication.

[3]  James S. Albus,et al.  New Approach to Manipulator Control: The Cerebellar Model Articulation Controller (CMAC)1 , 1975 .

[4]  Geoffrey E. Hinton Shape Representation in Parallel Systems , 1981, IJCAI.

[5]  Geoffrey E. Hinton A Parallel Computation that Assigns Canonical Object-Based Frames of Reference , 1981, IJCAI.

[6]  E. P. Gardner The organization of the cerebral cortex Eds F. O. Schmitt, F. G. Worden, G. Adelman and S. G. Dennis. MIT Press, Cambridge, MA: 1981. 592 pp. $50.00, £31.00 , 1982, Neuroscience.

[7]  J. Fodor Modularity of mind , 1983 .

[8]  Leslie G. Ungerleider,et al.  Object vision and spatial vision: two cortical pathways , 1983, Trends in Neurosciences.

[9]  Dana H. Ballard,et al.  Parameter Nets , 1984, Artif. Intell..

[10]  P. Anandan,et al.  Pattern-recognizing stochastic learning automata , 1985, IEEE Transactions on Systems, Man, and Cybernetics.

[11]  David Zipser,et al.  Feature Discovery by Competive Learning , 1986, Cogn. Sci..

[12]  K. Forster,et al.  The psychological status of overgenerated sentences , 1985, Cognition.

[13]  Geoffrey E. Hinton,et al.  Learning internal representations by error propagation , 1986 .

[14]  James L. McClelland The programmable blackboard model of reading , 1986 .

[15]  Dana H. Ballard,et al.  Cortical connections and parallel processing: Structure and function , 1986, Behavioral and Brain Sciences.

[16]  H. Barlow Why have multiple cortical areas? , 1986, Vision Research.

[17]  Stephen Grossberg,et al.  Competitive Learning: From Interactive Activation to Adaptive Resonance , 1987, Cogn. Sci..

[18]  Lawrence D. Jackel,et al.  Large Automatic Learning, Rule Extraction, and Generalization , 1987, Complex Syst..

[19]  Paul J. Werbos,et al.  Building and Understanding Adaptive Systems: A Statistical/Numerical Approach to Factory Automation and Brain Research , 1987, IEEE Transactions on Systems, Man, and Cybernetics.

[20]  Geoffrey E. Hinton,et al.  Learning sets of filters using back-propagation , 1987 .

[21]  S. Kosslyn Seeing and imagining in the cerebral hemispheres: a computational approach. , 1987, Psychological review.

[22]  Dean A. Pomerleau,et al.  The meta-generalized delta rule : a new algorithm for learning in connectionist networks , 1987 .

[23]  Richard Durbin,et al.  An analogue approach to the travelling salesman problem using an elastic net method , 1987, Nature.

[24]  Dana H. Ballard,et al.  Modular Learning in Neural Networks , 1987, AAAI.

[25]  J. R. Hughes Cerebral lateralization: biological mechanisms, associations and pathology , 1987 .

[26]  Jerome A. Feldman,et al.  Connectionist Models and Their Properties , 1982, Cogn. Sci..

[27]  Michael I. Jordan Supervised learning and systems with excess degrees of freedom , 1988 .

[28]  Yann LeCun,et al.  Generalization and network design strategies , 1989 .

[29]  Kumpati S. Narendra,et al.  Learning automata - an introduction , 1989 .

[30]  Peter Földiák,et al.  Adaptation and decorrelation in the cortex , 1989 .

[31]  George A. Bekey,et al.  Using a context-sensitive learning network for robot arm control , 1989 .

[32]  Alan L. Yuille,et al.  A Winner-Take-All Mechanism Based on Presynaptic Inhibition Feedback , 1989, Neural Computation.

[33]  Geoffrey E. Hinton Connectionist Learning Procedures , 1989, Artif. Intell..

[34]  S. Kosslyn,et al.  Why are What and Where Processed by Separate Cortical Visual Systems? A Computational Investigation , 1989, Journal of Cognitive Neuroscience.

[35]  Alexander H. Waibel,et al.  Modular Construction of Time-Delay Neural Networks for Speech Recognition , 1989, Neural Computation.

[36]  Kurt Hornik,et al.  Multilayer feedforward networks are universal approximators , 1989, Neural Networks.

[37]  Tomaso A. Poggio,et al.  Extensions of a Theory of Networks for Approximation and Learning , 1990, NIPS.

[38]  Michael I. Jordan Attractor dynamics and parallelism in a connectionist sequential machine , 1990 .

[39]  Richard Durbin,et al.  A dimension reduction framework for understanding cortical maps , 1990, Nature.

[40]  G. Reeke The society of mind , 1991 .