Human-level concept learning through probabilistic program induction

Handwritten characters drawn by a model Not only do children learn effortlessly, they do so quickly and with a remarkable ability to use what they have learned as the raw material for creating new stuff. Lake et al. describe a computational model that learns in a similar fashion and does so better than current deep learning algorithms. The model classifies, parses, and recreates handwritten characters, and can generate new letters of the alphabet that look “right” as judged by Turing-like tests of the model's output in comparison to what real humans produce. Science, this issue p. 1332 Combining the capacity to handle noise with probabilistic learning yields humanlike performance in a computational model. People learning new concepts can often generalize successfully from just a single example, yet machine learning algorithms typically require tens or hundreds of examples to perform with similar accuracy. People can also use learned concepts in richer ways than conventional algorithms—for action, imagination, and explanation. We present a computational model that captures these human learning abilities for a large class of simple visual concepts: handwritten characters from the world’s alphabets. The model represents concepts as simple programs that best explain observed examples under a Bayesian criterion. On a challenging one-shot classification task, the model achieves human-level performance while outperforming recent deep learning approaches. We also present several “visual Turing tests” probing the model’s creative generalization abilities, which in many cases are indistinguishable from human behavior.

[1]  H. Harlow,et al.  The formation of learning sets. , 1949, Psychological review.

[2]  J. Goodnow,et al.  “The grammar of action”: Sequence and syntax in children's copying☆ , 1973 .

[3]  T. Ferguson A Bayesian Analysis of Some Nonparametric Problems , 1973 .

[4]  G. Miller,et al.  Language and Perception , 1976 .

[5]  Patrick Henry Winston,et al.  The psychology of computer vision , 1976, Pattern Recognit..

[6]  Edward E. Smith,et al.  On the adequacy of prototype theory as a theory of concepts , 1981, Cognition.

[7]  L. Barsalou,et al.  Ad hoc categories , 1983, Memory & cognition.

[8]  Leslie G. Valiant,et al.  A theory of the learnable , 1984, STOC '84.

[9]  Donald D. Hoffman,et al.  Parts of recognition , 1984, Cognition.

[10]  I. Biederman Recognition-by-components: a theory of human image understanding. , 1987, Psychological review.

[11]  Linda B. Smith,et al.  The importance of shape in early lexical learning , 1988 .

[12]  M. Babcock,et al.  Perception of dynamic information in static handwritten forms. , 1988, The American journal of psychology.

[13]  E. Markman Categorization and naming in children , 1989 .

[14]  Lawrence D. Jackel,et al.  Backpropagation Applied to Handwritten Zip Code Recognition , 1989, Neural Computation.

[15]  Jun S. Huang,et al.  Stroke segmentation by bernstein-bezier curve fitting , 1990, Pattern Recognit..

[16]  J. Sethuraman A CONSTRUCTIVE DEFINITION OF DIRICHLET PRIORS , 1991 .

[17]  Ching Y. Suen,et al.  Thinning Methodologies - A Comprehensive Survey , 1992, IEEE Trans. Pattern Anal. Mach. Intell..

[18]  Elie Bienenstock,et al.  Neural Networks and the Bias/Variance Dilemma , 1992, Neural Computation.

[19]  T. B. Ward Structured Imagination: the Role of Category Structure in Exemplar Generation , 1994, Cognitive Psychology.

[20]  Geoffrey E. Hinton,et al.  Using Generative Models for Handwritten Digit Recognition , 1996, IEEE Trans. Pattern Anal. Mach. Intell..

[21]  Feldman,et al.  The Structure of Perceptual Categories , 1997, Journal of mathematical psychology.

[22]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[23]  A. Markman,et al.  Referential communication and category acquisition. , 1998, Journal of experimental psychology. General.

[24]  Vladimir Vapnik,et al.  An overview of statistical learning theory , 1999, IEEE Trans. Neural Networks.

[25]  Ching Y. Suen,et al.  Identification of Fork Points on the Skeletons of Handwritten Chinese Characters , 1999, IEEE Trans. Pattern Anal. Mach. Intell..

[26]  Jonathan Baxter,et al.  A Model of Inductive Bias Learning , 2000, J. Artif. Intell. Res..

[27]  Linda B. Smith,et al.  Object name Learning Provides On-the-Job Training for Attention , 2002, Psychological science.

[28]  Jean-Luc Velay,et al.  Visual presentation of single letters activates a premotor area involved in writing , 2003, NeuroImage.

[29]  R. Kimchi,et al.  Perceptual organization in vision : behavioral and neural perspectives , 2003 .

[30]  Yann LeCun,et al.  Learning a similarity metric discriminatively, with application to face verification , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[31]  K. James,et al.  Letter processing automatically recruits a sensory–motor brain network , 2006, Neuropsychologia.

[32]  J. Tenenbaum,et al.  Word learning as Bayesian inference. , 2007, Psychological review.

[33]  J. Tenenbaum,et al.  Bayesian Special Section Learning Overhypotheses with Hierarchical Bayesian Models , 2022 .

[34]  Zoubin Ghahramani,et al.  Proceedings of the 24th international conference on Machine learning , 2007, ICML 2007.

[35]  Thomas L. Griffiths,et al.  A Rational Analysis of Rule-Based Concept Learning , 2008, Cogn. Sci..

[36]  Joshua B. Tenenbaum,et al.  Church: a language for generative models , 2008, UAI.

[37]  R. Rosenfeld Nature , 2009, Otolaryngology--head and neck surgery : official journal of American Academy of Otolaryngology-Head and Neck Surgery.

[38]  K. James,et al.  When writing impairs reading: letter perception's susceptibility to motor interference. , 2009, Journal of experimental psychology. General.

[39]  Dima Damen,et al.  Computer Vision and Pattern Recognition (CVPR) , 2009 .

[40]  David A. McAllester,et al.  Object Detection with Discriminatively Trained Part Based Models , 2010, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[41]  A. Wenger,et al.  Reading in the brain , 2010 .

[42]  Guido Sanguinetti,et al.  Advances in Neural Information Processing Systems 24 , 2011 .

[43]  Joseph Jay Williams,et al.  The role of explanation in discovery and generalization: evidence from category learning , 2010, ICLS.

[44]  Susan J. Hespos,et al.  Categorization in 3- and 4-month-old infants: an advantage of words over tones. , 2010, Child development.

[45]  David Poeppel,et al.  Analysis by Synthesis: A (Re-)Emerging Program of Research for Language and Vision , 2010, Biolinguistics.

[46]  Daniel A. Braun,et al.  Structure learning in action , 2010, Behavioural Brain Research.

[47]  Noah D. Goodman,et al.  Learning a theory of causality. , 2011, Psychological review.

[48]  Noah D. Goodman,et al.  Bootstrapping in a language of thought: A formal model of numerical concept learning , 2012, Cognition.

[49]  Masae Sato,et al.  Reading in the brain , 2012 .

[50]  Tara N. Sainath,et al.  FUNDAMENTAL TECHNOLOGIES IN MODERN SPEECH RECOGNITION Digital Object Identifier 10.1109/MSP.2012.2205597 , 2012 .

[51]  Trevor Bekolay,et al.  Supplementary Materials for A Large-Scale Model of the Functioning Brain , 2012 .

[52]  Tara N. Sainath,et al.  The shared views of four research groups ) , 2012 .

[53]  Todd M. Gureckis,et al.  Evaluating Amazon's Mechanical Turk as a Tool for Experimental Behavioral Research , 2013, PloS one.

[54]  Charles Kemp,et al.  A probabilistic account of exemplar and category generation , 2013, Cognitive Psychology.

[55]  Joshua B. Tenenbaum,et al.  Learning with Hierarchical-Deep Models , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[56]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[57]  Rebecca Treiman,et al.  How Children Learn to Write Words , 2014 .

[58]  Trevor Darrell,et al.  Caffe: Convolutional Architecture for Fast Feature Embedding , 2014, ACM Multimedia.

[59]  B. Scassellati,et al.  Proceedings of the 36th Annual Conference of the Cognitive Science Society , 2014 .

[60]  Zoubin Ghahramani,et al.  Probabilistic machine learning and artificial intelligence , 2015, Nature.

[61]  Erica Cosentino,et al.  Proceedings of the 37th Annual Conference of the Cognitive Science Society. , 2015 .

[62]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[63]  Michael S. Bernstein,et al.  ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.

[64]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[65]  Philipp Koehn,et al.  Cognitive Psychology , 1992, Ageing and Society.

[66]  C. Belcher How Children Learn to Write Words , 2017 .