Exact Bayesian bin classification: a fast alternative to Bayesian classification and its application to neural response analysis

We investigate the general problem of signal classification and, in particular, that of assigning stimulus labels to neural spike trains recorded from single cortical neurons. Finding efficient ways of classifying neural responses is especially important in experiments involving rapid presentation of stimuli. We introduce a fast, exact alternative to Bayesian classification. Instead of estimating the class-conditional densities p(x|y) (where x is a scalar function of the feature[s], y the class label) and converting them to P(y|x) via Bayes’ theorem, this probability is evaluated directly and without the need for approximations. This is achieved by integrating over all possible binnings of x with an upper limit on the number of bins. Computational time is quadratic in both the number of observed data points and the number of bins. The algorithm also allows for the computation of feedback signals, which can be used as input to subsequent stages of inference, e.g. neural network training. Responses of single neurons from high-level visual cortex (area STSa) to rapid sequences of complex visual stimuli are analysed. Information latency and response duration increase nonlinearly with presentation duration, suggesting that neural processing speeds adapt to presentation speeds.

[1]  Michael I. Jordan,et al.  An Introduction to Variational Methods for Graphical Models , 1999, Machine Learning.

[2]  Corinna Cortes,et al.  Support-Vector Networks , 1995, Machine Learning.

[3]  Dimitri P. Bertsekas,et al.  Dynamic Programming and Optimal Control, Two Volume Set , 1995 .

[4]  Chih-Jen Lin,et al.  A Practical Guide to Support Vector Classication , 2008 .

[5]  C. E. SHANNON,et al.  A mathematical theory of communication , 1948, MOCO.

[6]  Thomas G. Dietterich What is machine learning? , 2020, Archives of Disease in Childhood.

[7]  E. Rolls,et al.  Representation of olfactory information in the primate orbitofrontal cortex. , 1996, Journal of neurophysiology.

[8]  P. Fldik,et al.  The Speed of Sight , 2001, Journal of Cognitive Neuroscience.

[9]  Marcus Hutter,et al.  Distribution of Mutual Information , 2001, NIPS.

[10]  D. Perrett,et al.  Rapid serial visual presentation for the determination of neural selectivity in area STSa. , 2004, Progress in brain research.

[11]  P. Földiák,et al.  The ‘Ideal Homunculus’: Statistical Inference from Neural Population Responses , 1993 .

[12]  James L. McClelland,et al.  Parallel distributed processing: explorations in the microstructure of cognition, vol. 1: foundations , 1986 .

[13]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[14]  Vladimir Cherkassky,et al.  The Nature Of Statistical Learning Theory , 1997, IEEE Trans. Neural Networks.

[15]  Dominik Endres,et al.  Quadratic programming for learning sparse codes , 1999 .

[16]  Sang Joon Kim,et al.  A Mathematical Theory of Communication , 2006 .

[17]  David Barber,et al.  Bayesian Classification With Gaussian Processes , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[18]  J. Miller,et al.  The Institution of Electrical Engineers , 2006, Nature.

[19]  Radford M. Neal Monte Carlo Implementation of Gaussian Process Models for Bayesian Regression and Classification , 1997, physics/9701026.

[20]  George Cybenko,et al.  Approximation by superpositions of a sigmoidal function , 1992, Math. Control. Signals Syst..

[21]  Bernhard Schölkopf,et al.  Comparison of View-Based Object Recognition Algorithms Using Realistic 3D Models , 1996, ICANN.

[22]  Stefano Panzeri,et al.  Analytical estimates of limited sampling biases in different information measures. , 1996, Network.

[23]  Dominik Endres,et al.  Bayesian and information-theoretic tools for neuroscience , 2006 .

[24]  Geoffrey E. Hinton,et al.  Learning internal representations by error propagation , 1986 .

[25]  George Cybenko,et al.  Approximation by superpositions of a sigmoidal function , 1989, Math. Control. Signals Syst..

[26]  Brendan J. Frey,et al.  Factor graphs and the sum-product algorithm , 2001, IEEE Trans. Inf. Theory.

[27]  David J. C. MacKay,et al.  Information Theory, Inference, and Learning Algorithms , 2004, IEEE Transactions on Information Theory.

[28]  P. Földiák,et al.  Forming sparse representations by local anti-Hebbian learning , 1990, Biological Cybernetics.

[29]  Richard W. Prager,et al.  Experiments with Low-Entropy Neural Networks , 2000, Information Theory and the Brain.

[30]  James M. Bower,et al.  Computation and Neural Systems , 2014, Springer US.

[31]  Vladimir N. Vapnik,et al.  The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.

[32]  Vladimir Vapnik,et al.  Statistical learning theory , 1998 .

[33]  D. Perrett,et al.  The `Ideal Homunculus': decoding neural population signals , 1998, Trends in Neurosciences.

[34]  D. Perrett,et al.  Time course of neural responses discriminating different views of the face and head. , 1992, Journal of neurophysiology.