Language Induction by Phase Transition in Dynamical Recognizers

A higher order recurrent neural network architecture learns to recognize and generate languages after being "trained" on categorized exemplars. Studying these networks from the perspective of dynamical systems yields two interesting discoveries: First, a longitudinal examination of the learning process illustrates a new form of mechanical inference: Induction by phase transition. A small weight adjustment causes a "bifurcation" in the limit behavior of the network. This phase transition corresponds to the onset of the network's capacity for generalizing to arbitrary-length strings. Second, a study of the automata resulting from the acquisition of previously published languages indicates that while the architecture is NOT guaranteed to find a minimal finite automata consistent with the given exemplars, which is an NP-Hard problem, the architecture does appear capable of generating nonregular languages by exploiting fractal and chaotic dynamics. I end the paper with a hypothesis relating linguistic generative capacity to the behavioral regimes of non-linear dynamical systems.

[1]  DANA ANGLUIN,et al.  On the Complexity of Minimum Inference of Regular Sets , 1978, Inf. Control..

[2]  Stephen Wolfram,et al.  Universality and complexity in cellular automata , 1983 .

[3]  P. Grassberger,et al.  Measuring the Strangeness of Strange Attractors , 1983 .

[4]  M. Barnsley,et al.  Solution of an inverse problem for fractals and other sets. , 1986, Proceedings of the National Academy of Sciences of the United States of America.

[5]  Paul Smolensky,et al.  Information processing in dynamical systems: foundations of harmony theory , 1986 .

[6]  Tad Hogg,et al.  Phase Transitions in Artificial Intelligence Systems , 1987, Artif. Intell..

[7]  W. Freeman,et al.  How brains make chaos in order to make sense of the world , 1987, Behavioral and Brain Sciences.

[8]  Meir,et al.  Chaotic behavior of a layered neural network. , 1988, Physical review. A, General physics.

[9]  James L. McClelland,et al.  Learning Subsequential Structure in Simple Recurrent Networks , 1988, NIPS.

[10]  Jordan B. Pollack,et al.  Implications of Recursive Distributed Representations , 1988, NIPS.

[11]  C. Lee Giles,et al.  Higher Order Recurrent Networks and Grammatical Inference , 1989, NIPS.

[12]  W. Pitts,et al.  A Logical Calculus of the Ideas Immanent in Nervous Activity (1943) , 2021, Ideas That Created the Future.

[13]  Jeffrey L. Elman,et al.  Finding Structure in Time , 1990, Cogn. Sci..

[14]  Jordan B. Pollack,et al.  Recursive Distributed Representations , 1990, Artif. Intell..

[15]  W. H. Zurek Complexity, Entropy and the Physics of Information , 1990 .

[16]  James P. Crutchfield,et al.  Computation at the Onset of Chaos , 1991 .