Resonator Networks, 1: An Efficient Solution for Factoring High-Dimensional, Distributed Representations of Data Structures

The ability to encode and manipulate data structures with distributed neural representations could qualitatively enhance the capabilities of traditional neural networks by supporting rule-based symbolic reasoning, a central property of cognition. Here we show how this may be accomplished within the framework of Vector Symbolic Architectures (VSAs) (Plate, 1991; Gayler, 1998; Kanerva, 1996), whereby data structures are encoded by combining high-dimensional vectors with operations that together form an algebra on the space of distributed representations. In particular, we propose an efficient solution to a hard combinatorial search problem that arises when decoding elements of a VSA data structure: the factorization of products of multiple codevectors. Our proposed algorithm, called a resonator network, is a new type of recurrent neural network that interleaves VSA multiplication operations and pattern completion. We show in two examples—parsing of a tree-like data structure and parsing of a visual scene—how the factorization problem arises and how the resonator network can solve it. More broadly, resonator networks open the possibility of applying VSAs to myriad artificial intelligence problems in real-world domains. The companion article in this issue (Kent, Frady, Sommer, & Olshausen, 2020) presents a rigorous analysis and evaluation of the performance of resonator networks, showing it outperforms alternative approaches.

[1]  Jerome Feldman,et al.  The neural binding problem(s) , 2013, Cognitive Neurodynamics.

[2]  Yann LeCun,et al.  The mnist database of handwritten digits , 2005 .

[3]  Pierre Comon,et al.  An iterative deflation algorithm for exact CP tensor decomposition , 2015, 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[4]  Yvonne Feierabend Foundations Of Language Brain Meaning Grammar Evolution , 2016 .

[5]  Friedrich T. Sommer,et al.  A Theory of Sequence Indexing and Working Memory in Recurrent Neural Networks , 2018, Neural Computation.

[6]  Hans-Peter Kriegel,et al.  A Three-Way Model for Collective Learning on Multi-Relational Data , 2011, ICML.

[7]  George Kachergis,et al.  Toward a scalable holographic word-form representation , 2011, Behavior research methods.

[8]  Geoffrey E. Hinton Mapping Part-Whole Hierarchies into Connectionist Networks , 1990, Artif. Intell..

[9]  Tony Plate,et al.  Holographic Reduced Representations: Convolution Algebra for Compositional Distributed Representations , 1991, IJCAI.

[10]  Bruno A. Olshausen,et al.  Resonator Networks, 2: Factorization Performance and Capacity Compared to Optimization-Based Methods , 2020, Neural Computation.

[11]  Ross W. Gayler Vector Symbolic Architectures answer Jackendoff's challenges for cognitive neuroscience , 2004, ArXiv.

[12]  Bruno A. Olshausen,et al.  Learning Intermediate-Level Representations of Form and Motion from Natural Movies , 2012, Neural Computation.

[13]  G. Kane Parallel Distributed Processing: Explorations in the Microstructure of Cognition, vol 1: Foundations, vol 2: Psychological and Biological Models , 1994 .

[14]  Geoffrey E. Hinton Tensor Product Variable Binding and the Representation of Symbolic Structures in Connectionist Systems , 1991 .

[15]  Jan M. Rabaey,et al.  Classification and Recall With Binary Hyperdimensional Computing: Tradeoffs in Choice of Density and Mapping Characteristics , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[16]  M. Ledoux The concentration of measure phenomenon , 2001 .

[17]  J. Wolfe,et al.  The Psychophysical Evidence for a Binding Problem in Human Vision , 1999, Neuron.

[18]  Tony A. Plate,et al.  Holographic reduced representations , 1995, IEEE Trans. Neural Networks.

[19]  Dmitri A. Rachkovskij,et al.  Binding and Normalization of Binary Sparse Distributed Representations by Context-Dependent Thinning , 2001, Neural Computation.

[20]  D. McDermott LANGUAGE OF THOUGHT , 2012 .

[21]  Pentti Kanerva,et al.  Fully Distributed Representation , 1997 .

[22]  J. Fodor,et al.  Connectionism and cognitive architecture: A critical analysis , 1988, Cognition.

[23]  Edward H. Adelson,et al.  The perception of shading and reflectance , 1996 .

[24]  J J Hopfield,et al.  Neural networks and physical systems with emergent collective computational abilities. , 1982, Proceedings of the National Academy of Sciences of the United States of America.

[25]  Aditya Joshi,et al.  Language Geometry Using Random Indexing , 2016, QI.

[26]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[27]  Jussi H. Poikonen,et al.  High-dimensional computing with sparse vectors , 2015, 2015 IEEE Biomedical Circuits and Systems Conference (BioCAS).

[28]  Friedrich T. Sommer,et al.  Robust computation with rhythmic spike patterns , 2019, Proceedings of the National Academy of Sciences.

[29]  Simon D. Levy,et al.  A DISTRIBUTED BASIS FOR ANALOGICAL MAPPING , 2009 .

[30]  G. A. Miller THE PSYCHOLOGICAL REVIEW THE MAGICAL NUMBER SEVEN, PLUS OR MINUS TWO: SOME LIMITS ON OUR CAPACITY FOR PROCESSING INFORMATION 1 , 1956 .

[31]  Jitendra Malik,et al.  Shape, Illumination, and Reflectance from Shading , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[32]  Z. Harris,et al.  Foundations of language , 1941 .

[33]  Ross W. Gayler,et al.  Multiplicative Binding, Representation Operators & Analogy , 1998 .

[34]  Danqi Chen,et al.  Reasoning With Neural Tensor Networks for Knowledge Base Completion , 2013, NIPS.

[35]  Okko Johannes Räsänen,et al.  Generating Hyperdimensional Distributed Representations from Continuous-Valued Multivariate Sensory Input , 2015, CogSci.

[36]  Tony A. Plate,et al.  Holographic Reduced Representation: Distributed Representation for Cognitive Structures , 2003 .

[37]  Friedrich T. Sommer,et al.  Variable Binding for Sparse Distributed Representations: Theory and Applications , 2020, IEEE Transactions on Neural Networks and Learning Systems.

[38]  Christoph von der Malsburg,et al.  The What and Why of Binding The Modeler’s Perspective , 1999, Neuron.

[39]  H. Barrow,et al.  RECOVERING INTRINSIC SCENE CHARACTERISTICS FROM IMAGES , 1978 .

[40]  John E. Hummel,et al.  Distributed representations of structure: A theory of analogical access and mapping. , 1997 .

[41]  Austin Roorda,et al.  High-acuity vision from retinal image motion , 2020, Journal of vision.

[42]  Pentti Kanerva,et al.  Large Patterns Make Great Symbols: An Example of Learning from Example , 1998, Hybrid Neural Systems.

[43]  Pentti Kanerva,et al.  Hyperdimensional Computing: An Introduction to Computing in Distributed Representation with High-Dimensional Random Vectors , 2009, Cognitive Computation.

[44]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[45]  Pentti Kanerva,et al.  Binary Spatter-Coding of Ordered K-Tuples , 1996, ICANN.

[46]  Bartlett W. Mel,et al.  Sigma-Pi Learning: On Radial Basis Functions and Cortical Associative Learning , 1989, NIPS.

[47]  Joshua B. Tenenbaum,et al.  Building machines that learn and think like people , 2016, Behavioral and Brain Sciences.

[48]  Geoffrey E. Hinton,et al.  Learning to Represent Spatial Transformations with Factored Higher-Order Boltzmann Machines , 2010, Neural Computation.

[49]  Ross W. Gayler Multiplicative Binding, Representation Operators & Analogy (Workshop Poster) , 1998 .

[50]  A. Treisman,et al.  A feature-integration theory of attention , 1980, Cognitive Psychology.

[51]  Tony A. Plate,et al.  Analogy retrieval and processing with distributed vector representations , 2000, Expert Syst. J. Knowl. Eng..

[52]  T A Plate,et al.  Randomly connected sigma–pi neurons can form associator networks , 2000, Network.