Stabilizing patterns in time: Neural network approach

Recurrent and feedback networks are capable of holding dynamic memories. Nonetheless, training a network for that task is challenging. In order to do so, one should face non-linear propagation of errors in the system. Small deviations from the desired dynamics due to error or inherent noise might have a dramatic effect in the future. A method to cope with these difficulties is thus needed. In this work we focus on recurrent networks with linear activation functions and binary output unit. We characterize its ability to reproduce a temporal sequence of actions over its output unit. We suggest casting the temporal learning problem to a perceptron problem. In the discrete case a finite margin appears, providing the network, to some extent, robustness to noise, for which it performs perfectly (i.e. producing a desired sequence for an arbitrary number of cycles flawlessly). In the continuous case the margin approaches zero when the output unit changes its state, hence the network is only able to reproduce the sequence with slight jitters. Numerical simulation suggest that in the discrete time case, the longest sequence that can be learned scales, at best, as square root of the network size. A dramatic effect occurs when learning several short sequences in parallel, that is, their total length substantially exceeds the length of the longest single sequence the network can learn. This model easily generalizes to an arbitrary number of output units, which boost its performance. This effect is demonstrated by considering two practical examples for sequence learning. This work suggests a way to overcome stability problems for training recurrent networks and further quantifies the performance of a network under the specific learning scheme.

[1]  A. A. Mullin,et al.  Principles of neurodynamics , 1962 .

[2]  Paul C. Bressloff,et al.  Perceptron-like learning in time-summating neural networks , 1992 .

[3]  F ROSENBLATT,et al.  The perceptron: a probabilistic model for information storage and organization in the brain. , 1958, Psychological review.

[4]  H Sompolinsky,et al.  Associative neural network model for the generation of temporal patterns. Theory and application to central pattern generators. , 1988, Biophysical journal.

[5]  Michael I. Jordan Serial Order: A Parallel Distributed Processing Approach , 1997 .

[6]  Harald Haas,et al.  Harnessing Nonlinearity: Predicting Chaotic Systems and Saving Energy in Wireless Communication , 2004, Science.

[7]  L. D. Harmon Neuromimes: Action of a Reciprocally Inhibitory Pair , 1964, Science.

[8]  Daniel L Adams,et al.  The cortical column: a structure without a function , 2005, Philosophical Transactions of the Royal Society B: Biological Sciences.

[9]  L. F. Abbott,et al.  Generating Coherent Patterns of Activity from Chaotic Neural Networks , 2009, Neuron.

[10]  Paul C. Bressloff,et al.  Temporal sequence storage capacity of time-summating neural networks , 1992 .

[11]  Haim Sompolinsky,et al.  Short-term memory in orthogonal neural networks. , 2004, Physical review letters.

[12]  Anders Krogh,et al.  Introduction to the theory of neural computation , 1994, The advanced book program.

[13]  Peter Ford Dominey Complex sensory-motor sequence learning based on recurrent state representation and reinforcement learning , 1995, Biological Cybernetics.

[14]  Jeffrey L. Elman,et al.  Finding Structure in Time , 1990, Cogn. Sci..

[15]  Nicolas Brunel,et al.  Storage of Correlated Patterns in Standard and Bistable Purkinje Cell Models , 2012, PLoS Comput. Biol..

[16]  William H. Press,et al.  Numerical recipes in C , 2002 .

[17]  Haim Sompolinsky,et al.  Associative network models for central pattern generators , 1989 .

[18]  Eduardo D. Sontag,et al.  Computational Aspects of Feedback in Neural Circuits , 2006, PLoS Comput. Biol..

[19]  G. Székely,et al.  Simulation of rhythmic nervous activities , 1968, Kybernetik.

[20]  M. Nissen,et al.  Attentional requirements of learning: Evidence from performance measures , 1987, Cognitive Psychology.

[21]  Henry Markram,et al.  Real-Time Computing Without Stable States: A New Framework for Neural Computation Based on Perturbations , 2002, Neural Computation.

[22]  Corinna Cortes,et al.  Support-Vector Networks , 1995, Machine Learning.

[23]  D. O. Hebb,et al.  The organization of behavior , 1988 .

[24]  Benjamin A. Clegg,et al.  Short article: Implicit motor sequence learning is not represented purely in response locations , 2009, Quarterly journal of experimental psychology.

[25]  Donald M. Wilson,et al.  Models for the generation of the motor output pattern in flying locusts , 1968 .

[26]  Y. Adini,et al.  The time course and characteristics of procedural learning in schizophrenia patients and healthy individuals , 2015, Front. Hum. Neurosci..

[27]  Jochen J. Steil,et al.  Analyzing the weight dynamics of recurrent learning algorithms , 2005, Neurocomputing.