Jump to content

Gaussian process approximations

From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by Marcin.jurek (talk | contribs) at 17:59, 19 June 2020. The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

In statistics and machine learning Gaussian process approximation is a computational method that accelerates inference tasks in the context of a Gaussian process model, most commonly likelihood evaluation and prediction. Like approximations of other models, they can often be expressed as additional assumptions imposed on the model, which do not correspond to any actual feature, but which retain its key properties while simplifying calculations. Many of these approximation methods can be expressed in purely linear algebraic or functional analytic terms as matrix or function approximations. Others are purely algorithmic and cannot easily be rephrased as a modification of a statistical model.

Basic ideas

In statistical modeling, it is often convenient to assume that , the phenomenon under investigation is a Gaussian process indexed by which has mean function and covariance function . One can also assume that data are values of a particular realization of this process for indices .

Consequently, the joint distribution of the data can be expressed as

,

where and , i.e. respectively a vector a matrix with the covariance function values and a vector with the mean function values at corresponding (pairs of) indices. The negative log-likelihood of the data then takes the form

Similarly, the best predictor of , the values of for indices , given data has the form

In the context of Gaussian models, especially in geostatistics, prediction using the best predictor, i.e. mean conditional on the data, is also known as kriging.

The most computationally expensive component of the best predictor formula is inverting the covariance matrix , which has cubic complexity . Similarly, evaluating likelihood involves both calculating and the determinant which has the same cubic complexity.

Gaussian process approximations can often be expressed in terms of assumptions on under which and can be calculated with much lower complexity. Since these assumptions are generally not believed to reflect reality, the likelihood and the best predictor obtained in this way are not exact, but they are meant to be close to their original values.

Model-based methods

This class of approximations is expressed through a set of assumptions which are imposed on the original process and which, typically, imply some special structure of the covariance matrix. Although most of these methods were developed independently, most of them can be expressed as special cases of the sparse general Vecchia approximation.

Low-rank methods

While this approach encompasses many methods, the common assumption underlying them all is the assumption, that , the Gaussian process of interest, is effectively low-rank. More precisely, it is assumed, that there exists a set of indices such that every other set of indices

where is an matrix, and and is a diagonal matrix. Depending on the method and application various ways of selecting have been proposed.

Typically, is selected to be much smaller than which means that the computational cost of inverting is manageable ( instead of ).

Other sparse methods

Hierarchical methods

The general principle of hierarchical approximations consists of a repeated application of some other method, such that each consecutive application refines the quality of the approximation. Even though they can be expressed as a set of statistical assumptions, they are often described in terms of a hierarchical matrix approximation (HODLR) or basis function expansion (LatticeKrig, MRA, wavelets). The hierarchical matrix approach can often be represented as a repeated application of a low-rank approximation to successively smaller subsets of the index set . Basis function expansion relies on using functions with compact support. These features can then be exploited by an algorithm who steps through consecutive layers of the approximation. In the most favourable settings some of these methods can achieve quasi-linear () complexity.

Unified framework

Probabilistic graphical models provide a convenient framework for comparing model-based approximations. In this context, value of the process at index can then be represented by a vertex in a directed graph and edges correspond to the terms in the factorization of the joint density of . In general, when no independece relations are assumed, the joint probability distribution can be represented by an arbitrary ordering of vertices and edges that make up a spanning tree. Using a particular approximation can then be expressed as a certain way of ordering the vertices and adding or removing specific edges.

Vecchia approximation

An early attempt at using this framework for approximating Gaussian processes is called Vecchia approximation. For a given ordering of vertices, it is assumed that each vertex is independent of other vertices given preceding vertices. For example, for this implies, that the join density can be written as

while the graphical representation will look as follows.

Graphical representation of the Vecchia approximation. It shows that in the factorization of the joint density the orange vertex '"`UNIQ--postMath-0000002D-QINU`"' will be dependent on light orange vertices '"`UNIQ--postMath-0000002E-QINU`"' and '"`UNIQ--postMath-0000002F-QINU`"'
Graphical representation of the Vecchia approximation. It shows that in the factorization of the joint density the orange vertex will be dependent on light orange vertices and

Methods without a statistical model

References

  • . arXiv:1807.01065 [stat.ML]. {{cite arXiv}}: Missing or empty |title= (help) A bot will complete this citation soon. Click here to jump the queue
  • Heaton, Matthew J.; Datta, Abhirup; Finley, Andrew O.; Furrer, Reinhard; Guinness, Joseph; Guhaniyogi, Rajarshi; Gerber, Florian; Gramacy, Robert B.; Hammerling, Dorit; Katzfuss, Matthias; Lindgren, Finn; Nychka, Douglas W.; Sun, Furong; Zammit-Mangion, Andrew (2018). "A Case Study Competition Among Methods for Analyzing Large Spatial Data". Journal of Agricultural, Biological and Environmental Statistics. 24 (3): 398–425. doi:10.1007/s13253-018-00348-w. ISSN 1085-7117.