Spatial coherence as an internal teacher for a neural network

Supervised learning procedures for neural networks have recently met with considerable success in learning di cult mappings. So far, however, they have been limited by their poor scaling behaviour, particularly for networks with many hidden layers. A promising alternative is to develop unsupervised learning algorithms by de ning objective functions that characterize the quality of an internal representation without requiring knowledge of the desired outputs of the system. Our major goal is to build self-organizing network modules which capture important regularities in the environment in a simple form. A layered hierarchy of such modules should be able to learn in a time roughly linear in the number of layers. We propose that a good objective for perceptual learning is to extract higher-order features that exhibit simple coherence across time or space. This can be done by transforming the input representation into an underlying representation in which the mutual information between adjacent patches of the input can be expressed in a simple way. We have applied this basic idea to develop several interesting learning algorithms for discovering spatially coherent features in images. Our simulations show that a network can discover depth of surfaces when trained on binary random dot stereograms with discrete global shifts, as well as on real-valued stereograms of surfaces with continuously varying disparities. Once a module of depth-tuned units has developed, we show that units in a higher layer can discover a simple form of surface interpolation of curved surfaces, by learning to predict the depth of one image region based on depth measurements in surrounding regions.

[1]  D. Hubel,et al.  RECEPTIVE FIELDS OF CELLS IN STRIATE CORTEX OF VERY YOUNG, VISUALLY INEXPERIENCED KITTENS. , 1963, Journal of neurophysiology.

[2]  D. Hubel,et al.  Comparison of the effects of unilateral and bilateral eye closure on cortical unit responses in kittens. , 1965, Journal of neurophysiology.

[3]  G. F. Cooper,et al.  Development of the Brain depends on the Visual Environment , 1970, Nature.

[4]  Donald B. Rubin,et al.  Max-imum Likelihood from Incomplete Data , 1972 .

[5]  C. Blakemore,et al.  Innate and environmental factors in the development of the kitten's visual cortex. , 1975, The Journal of physiology.

[6]  D. Rubin,et al.  Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper , 1977 .

[7]  E. Oja Simplified neuron model as a principal component analyzer , 1982, Journal of mathematical biology.

[8]  L. S. Meharg,et al.  Binocular neurons and binocular function in monkeys and children. , 1983, Investigative ophthalmology & visual science.

[9]  E. Smith,et al.  Stereoblind monkeys have few binocular neurons. , 1984, Investigative ophthalmology & visual science.

[10]  David Zipser,et al.  Feature Discovery by Competive Learning , 1986, Cogn. Sci..

[11]  R Linsker,et al.  From basic network principles to neural architecture: emergence of orientation columns. , 1986, Proceedings of the National Academy of Sciences of the United States of America.

[12]  Geoffrey E. Hinton,et al.  Experiments on Learning by Back Propagation. , 1986 .

[13]  M. Stryker,et al.  Binocular impulse blockade prevents the formation of ocular dominance columns in cat visual cortex , 1986, The Journal of neuroscience : the official journal of the Society for Neuroscience.

[14]  Eric Saund Abstraction and Representation of Continuous Variables in Connectionist Networks , 1986, AAAI.

[15]  R Linsker,et al.  From basic network principles to neural architecture: emergence of orientation-selective cells. , 1986, Proceedings of the National Academy of Sciences of the United States of America.

[16]  R Linsker,et al.  From basic network principles to neural architecture: emergence of spatial-opponent cells. , 1986, Proceedings of the National Academy of Sciences of the United States of America.

[17]  Barak A. Pearlmutter,et al.  G-maximization: An unsupervised learning procedure for discovering regularities , 1987 .

[18]  Robin Sibson,et al.  What is projection pursuit , 1987 .

[19]  Stephen Grossberg,et al.  Competitive Learning: From Interactive Activation to Adaptive Resonance , 1987, Cogn. Sci..

[20]  S. Grossberg,et al.  ART 2: self-organization of stable category recognition codes for analog input patterns. , 1987, Applied optics.

[21]  Lokendra Shastri,et al.  Learning Phonetic Features Using Connectionist Networks , 1987, IJCAI.

[22]  Ralph Linsker,et al.  Self-organization in a perceptual network , 1988, Computer.

[23]  M. Sur,et al.  Experimentally induced visual projections into auditory thalamus and cortex. , 1988, Science.

[24]  Terence D. Sanger,et al.  An Optimality Principle for Unsupervised Learning , 1988, NIPS.

[25]  Robert A. Jacobs,et al.  Increased rates of convergence through learning rate adaptation , 1987, Neural Networks.

[26]  Erkki Oja,et al.  Neural Networks, Principal Components, and Subspaces , 1989, Int. J. Neural Syst..

[27]  John Moody,et al.  Fast Learning in Networks of Locally-Tuned Processing Units , 1989, Neural Computation.

[28]  Terence D. Sanger Optimal Unsupervised Learning in Feedforward Neural Networks , 1989 .

[29]  Steven J. Nowlan,et al.  Maximum Likelihood Competitive Learning , 1989, NIPS.

[30]  Terence D. Sanger,et al.  Optimal unsupervised learning in a single-layer linear feedforward neural network , 1989, Neural Networks.

[31]  K. Miller,et al.  Ocular dominance column development: analysis and simulation. , 1989, Science.

[32]  Kurt Hornik,et al.  Neural networks and principal component analysis: Learning from examples without local minima , 1989, Neural Networks.

[33]  Kevin J. Lang A time delay neural network architecture for speech recognition , 1989 .

[34]  S. Lehky,et al.  Neural model of stereoacuity and depth interpolation based on a distributed representation of stereo disparity [published erratum appears in J Neurosci 1991 Mar;11(3):following Table of Contents] , 1990, The Journal of neuroscience : the official journal of the Society for Neuroscience.

[35]  Geoffrey E. Hinton,et al.  Discovering Viewpoint-Invariant Relationships That Characterize Objects , 1990, NIPS.

[36]  Tomaso A. Poggio,et al.  Extensions of a Theory of Networks for Approximation and Learning , 1990, NIPS.

[37]  Geoffrey E. Hinton,et al.  Adaptive Mixtures of Local Experts , 1991, Neural Computation.