Localist Attractor Networks

Attractor networks, which map an input space to a discrete output space, are useful for pattern completioncleaning up noisy or missing input features. However, designing a net to have a given set of attractors is notoriously tricky; training procedures are CPU intensive and often produce spurious attractors and ill-conditioned attractor basins. These difficulties occur because each connection in the network participates in the encoding of multiple attractors. We describe an alternative formulation of attractor networks in which the encoding of knowledge is local, not distributed. Although localist attractor networks have similar dynamics to their distributed counterparts, they are much easier to work with and interpret. We propose a statistical formulation of localist attractor net dynamics, which yields a convergence proof and a mathematical interpretation of model parameters. We present simulation experiments that explore the behavior of localist attractor networks, showing that they yield few spurious attractors, and they readily exhibit two desirable properties of psychological and neurobiological models: priming (faster convergence to an attractor if the attractor has been recently visited) and gang effects (in which the presence of an attractor enhances the attractor basins of neighboring attractors).

[1]  Richard O. Duda,et al.  Pattern classification and scene analysis , 1974, A Wiley-Interscience publication.

[2]  Peter E. Hart,et al.  Pattern classification and scene analysis , 1974, A Wiley-Interscience publication.

[3]  D. Rubin,et al.  Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper , 1977 .

[4]  James L. McClelland,et al.  An interactive activation model of context effects in letter perception: I. An account of basic findings. , 1981 .

[5]  J J Hopfield,et al.  Neural networks and physical systems with emergent collective computational abilities. , 1982, Proceedings of the National Academy of Sciences of the United States of America.

[6]  Steven W. Zucker,et al.  On the Foundations of Relaxation Labeling Processes , 1983, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[7]  C. D. Gelatt,et al.  Optimization by Simulated Annealing , 1983, Science.

[8]  J J Hopfield,et al.  Neurons with graded response have collective computational properties like those of two-state neurons. , 1984, Proceedings of the National Academy of Sciences of the United States of America.

[9]  Geoffrey E. Hinton,et al.  A Learning Algorithm for Boltzmann Machines , 1985, Cogn. Sci..

[10]  James L. McClelland,et al.  Distributed memory and the representation of general and specific information. , 1985, Journal of experimental psychology. General.

[11]  Richard M. Golden,et al.  Probabilistic Characterization of Neural Model Computations , 1987, NIPS.

[12]  Fernando J. Pineda,et al.  Generalization of Back propagation to Recurrent and Higher Order Neural Networks , 1987, NIPS.

[13]  Tomaso Poggio,et al.  Cooperative computation of stereo disparity , 1988 .

[14]  L. B. Almeida A learning rule for asynchronous perceptrons with feedback in a combinatorial environment , 1990 .

[15]  Shlomo Geva,et al.  An Exponential Response Neural Net , 1991, Neural Computation.

[16]  Javier R. Movellan,et al.  Learning Continuous Probability Distributions with Symmetric Diffusion Networks , 1993, Cogn. Sci..

[17]  Eytan Ruppin,et al.  Compensatory Mechanisms in an Attractor Neural Network Model of Schizophrenia , 1995, Neural Computation.

[18]  A. Giovagnoli Connectionist modelling in cognitive neuropsychology: A case study , 1995, The Italian Journal of Neurological Sciences.

[19]  James L. McClelland,et al.  Understanding normal and impaired word reading: computational principles in quasi-regular domains. , 1996, Psychological review.

[20]  Walter J. Freeman,et al.  Reafference and Attractors in the Olfactory System During Odor Recognition , 1996, Int. J. Neural Syst..

[21]  Donald W. Mathis,et al.  Conscious and unconscious perception: A computational theory , 1996 .

[22]  Michael I. Jordan,et al.  Mean Field Theory for Sigmoid Belief NetworksMean Field Theory for Sigmoid Belief , 1996 .

[23]  C. R. Neto,et al.  Multivalley structure of attractor neural networks , 1997 .

[24]  Michael C. Mozer,et al.  A Superadditive-Impairment Theory of Optic Aphasia , 1997, NIPS.

[25]  Michael K. Tanenhaus,et al.  Parsing in a Dynamical System: An Attractor-based Account of the Interaction of Lexical and Structural Constraints in Sentence Processing , 1997 .

[26]  Mark S. Seidenberg,et al.  On the nature and scope of featural representations of word meaning. , 1997, Journal of experimental psychology. General.

[27]  S. Becker,et al.  Long-term semantic priming: a computational account and empirical evidence. , 1997, Journal of experimental psychology. Learning, memory, and cognition.

[28]  B L McNaughton,et al.  Path Integration and Cognitive Mapping in a Continuous Attractor Neural Network Model , 1997, The Journal of Neuroscience.

[29]  D. Amit,et al.  Model of global spontaneous activity and local structured activity during delay periods in the cerebral cortex. , 1997, Cerebral cortex.

[30]  David S. Touretzky,et al.  The Role of the Hippocampus in Solving the Morris Water Maze , 1998, Neural Computation.

[31]  Donald W. Mathis,et al.  A computational theory of consciousness in cognition , 1998 .

[32]  Michael I. Jordan Learning in Graphical Models , 1999, NATO ASI Series.

[33]  Geoffrey E. Hinton,et al.  A View of the Em Algorithm that Justifies Incremental, Sparse, and other Variants , 1998, Learning in Graphical Models.

[34]  Holly P. Branigan,et al.  Proceedings of the 21st Annual Conference of the Cognitive Science Society , 1999 .

[35]  John J. Hopfield,et al.  Neural networks and physical systems with emergent collective computational abilities , 1999 .

[36]  Drew McDermott A Computational Theory of Consciousness , 2001 .

[37]  J. Moake,et al.  This article has been cited by other articles , 2003 .

[38]  David C. Noelle,et al.  Methods for Learning Articulated Attractors over Internal Representations , 1999 .