Monte Carlo Implementation of Gaussian Process Models for Bayesian Regression and Classification

Gaussian processes are a natural way of defining prior distributions over functions of one or more input variables. In a simple nonparametric regression problem, where such a function gives the mean of a Gaussian distribution for an observed response, a Gaussian process model can easily be implemented using matrix computations that are feasible for datasets of up to about a thousand cases. Hyperparameters that define the covariance function of the Gaussian process can be sampled using Markov chain methods. Regression models where the noise has a t distribution and logistic or probit models for classification applications can be implemented by sampling as well for latent values underlying the observations. Software is now available that implements these methods using covariance functions with hierarchical parameterizations. Models defined in this way can discover high-level properties of the data, such as which inputs are relevant to predicting the response.

[1]  Richard Von Mises,et al.  Mathematical Theory of Probability and Statistics , 1966 .

[2]  Harald Bergstriim Mathematical Theory of Probability and Statistics , 1966 .

[3]  G. Wahba Improper Priors, Spline Smoothing and the Problem of Guarding Against Model Errors in Regression , 1978 .

[4]  A. O'Hagan,et al.  Curve Fitting and Optimal Design for Prediction , 1978 .

[5]  K. Mardia,et al.  Maximum likelihood estimation of models for residual covariance in spatial regression , 1984 .

[6]  Charles R. Johnson,et al.  Matrix analysis , 1985, Statistical Inference for Engineers and Data Scientists.

[7]  Ronald A. Thisted,et al.  Elements of statistical computing , 1986 .

[8]  S. Duane,et al.  Hybrid Monte Carlo , 1987 .

[9]  A. Yaglom Correlation Theory of Stationary and Related Random Functions I: Basic Results , 1987 .

[10]  Bryan F. J. Manly,et al.  Maximum likelihood estimation of models , 1990 .

[11]  A. Horowitz A generalized guided Monte Carlo algorithm , 1991 .

[12]  Radford M. Neal An improved acceptance procedure for the hybrid Monte Carlo algorithm , 1992, hep-lat/9208011.

[13]  A. W. Kemp,et al.  Kendall's Advanced Theory of Statistics. , 1994 .

[14]  Carl E. Rasmussen,et al.  In Advances in Neural Information Processing Systems , 2011 .

[15]  Geoffrey E. Hinton,et al.  Bayesian Learning for Neural Networks , 1995 .

[16]  David Barber,et al.  Gaussian Processes for Bayesian Classification via Hybrid Monte Carlo , 1996, NIPS.

[17]  Geoffrey E. Hinton,et al.  Evaluation of Gaussian processes and other methods for non-linear regression , 1997 .

[18]  M. Gibbs,et al.  Efficient implementation of gaussian processes , 1997 .