Parsing Embedded Clauses with Distributed Neural Networks

A distributed neural network model called SPEC for processing sentences with recursive relative clauses is described. The model is based on separating the tasks of segmenting the input word sequence into clauses, forming the case-role representations, and keeping track of the recursive embeddings into different modules. The system needs to be trained only with the basic sentence constructs, and it generalizes not only to new instances of familiar relative clause structures, but to novel structures as well. SPEC exhibits plausible memory degradation as the depth of the center embeddings increases, its memory is primed by earlier constituents, and its performance is aided by semantic constraints between the constituents. The ability to process structure is largely due to a central executive network that monitors and controls the execution of the entire system. This way, in contrast to earlier subsymbolic systems, parsing is modeled as a controlled high-level process rather than one based on automatic reflex responses.

[1]  Risto Miikkulainen,et al.  Subsymbolic natural language processing - an integrated model of scripts, lexicon, and memory , 1993, Neural network modeling and connectionism.

[2]  J. Elman Distributed Representations, Simple Recurrent Networks, And Grammatical Structure , 1991 .

[3]  George Berg,et al.  A Connectionist Parser with Recursive Sentence Structure and Lexical Disambiguation , 1992, AAAI.

[4]  Risto Miikkulainen,et al.  Natural Language Processingwith Modular Neural Networks and Distributed Lexicon , 1991 .

[5]  George A. Miller,et al.  Free Recall of Self-Embedded English Sentences , 1964, Inf. Control..

[6]  Risto Miikkulainen,et al.  Natural Language Processing With Modular PDP Networks and Distributed Lexicon , 1991, Cogn. Sci..

[7]  Donald J. Foss,et al.  Some effects of memory limitation upon sentence comprehension and recall , 1970 .

[8]  A. Caramazza,et al.  Dissociation of algorithmic and heuristic processes in language comprehension: Evidence from aphasia , 1976, Brain and Language.

[9]  Michael I. Jordan,et al.  Task Decomposition Through Competition in a Modular Connectionist Architecture: The What and Where Vision Tasks , 1990, Cogn. Sci..

[10]  Ajay N. Jain Parsing Complex Sentences with Structured Connectionist Networks , 1991, Neural Computation.

[11]  David S. Touretzky Connectionism and Compositional Semantics , 1989 .

[12]  Risto Miikkulainen,et al.  A Pdp Architecture for Processing Sentences With Relative Clauses , 1990, COLING.

[13]  Andreas Stolcke Learning Feature-based Semantics with Simple Recurrent Networks , 1990 .

[14]  M. Huang A Developmental Study of Children's Comprehension of Embedded Sentences with and without Semantic Constraints , 1983 .

[15]  Mark C. Detweiler,et al.  A Connectionist/Control Architecture for Working Memory , 1988 .

[16]  Mark F. St. John,et al.  The Story Gestalt: A Model of Knowledge-Intensive Processes in Text Comprehension , 1992, Cogn. Sci..

[17]  James L. McClelland,et al.  Learning and Applying Contextual Constraints in Sentence Comprehension , 1990, Artif. Intell..

[18]  Jeffrey L. Elman,et al.  Finding Structure in Time , 1990, Cogn. Sci..

[19]  Jordan B. Pollack,et al.  Recursive Distributed Representations , 1990, Artif. Intell..

[20]  Walter S. Stolz,et al.  A study of the ability to decode grammatically novel sentences , 1967 .

[21]  James L. McClelland,et al.  Mechanisms of Sentence Processing: Assigning Roles to Constituents of Sentences , 1986 .

[22]  Geoffrey E. Hinton,et al.  Learning internal representations by error propagation , 1986 .