GAUSSIAN PROCESSES FOR MACHINE LEARNING
Top Cited Papers
- 1 April 2004
- journal article
- research article
- Published by World Scientific Pub Co Pte Ltd in International Journal of Neural Systems
- Vol. 14 (02) , 69-106
- https://doi.org/10.1142/s0129065704001899
Abstract
Gaussian processes (GPs) are natural generalisations of multivariate Gaussian random variables to infinite (countably or continuous) index sets. GPs have been applied in a large number of fields to a diverse range of ends, and very many deep theoretical analyses of various properties are available. This paper gives an introduction to Gaussian processes on a fairly elementary level with special emphasis on characteristics relevant in machine learning. It draws explicit connections to branches such as spline smoothing models and support vector machines in which similar ideas have been investigated.Gaussian process models are routinely used to solve hard machine learning problems. They are attractive because of their flexible non-parametric nature and computational simplicity. Treated within a Bayesian framework, very powerful statistical methods can be implemented which offer valid estimates of uncertainties in our predictions and generic model selection procedures cast as nonlinear optimization problems. Their main drawback of heavy computational scaling has recently been alleviated by the introduction of generic sparse approximations.13,78,31The mathematical literature on GPs is large and often uses deep concepts which are not required to fully understand most machine learning applications. In this tutorial paper, we aim to present characteristics of GPs relevant to machine learning and to show up precise connections to other "kernel machines" popular in the community. Our focus is on a simple presentation, but references to more detailed sources are provided.Keywords
This publication has 30 references indexed in Scilit:
- Sparse On-Line Gaussian ProcessesNeural Computation, 2002
- A Global Geometric Framework for Nonlinear Dimensionality ReductionScience, 2000
- Computation with Infinite Neural NetworksNeural Computation, 1998
- Bayesian classification with Gaussian processesPublished by Institute of Electrical and Electronics Engineers (IEEE) ,1998
- A fast method for calculating the perceptron with maximal stabilityJournal de Physique I, 1993
- Networks for approximation and learningProceedings of the IEEE, 1990
- Generalized Linear ModelsJournal of the Royal Statistical Society. Series A (General), 1972
- A Correspondence Between Bayesian Estimation on Stochastic Processes and Smoothing by SplinesThe Annals of Mathematical Statistics, 1970
- The perceptron: A probabilistic model for information storage and organization in the brain.Psychological Review, 1958
- Theory of reproducing kernelsTransactions of the American Mathematical Society, 1950