Skip to main content

Gaussian processes (GPs) are popular, flexible, and interpretable probabilistic models for functions in geospatial analysis, computer-model emulation, and machine learning. However, direct application of GPs involves dense covariance matrices and is computationally infeasible for large datasets. We consider a framework for fast GP inference based on the so-called Vecchia approximation, which implies a sparse Cholesky factor of the inverse covariance matrix. The approximation can be written in closed form and computed in parallel, and it includes many popular existing approximations as special cases. We discuss various applications and extensions of the framework, including high-dimensional inference and variable selection, variational approximations for latent GPs, and nonparametric learning of non-Gaussian distributions.

Bio: I am a Professor in the Department of Statistics at Texas A&M University. I’m the recipient of an NSF Career Award, a Fulbright Scholarship, and an Early Investigator Award by the ASA Section on Statistics and the Environment.

arrow-left-smallarrow-right-large-greyarrow-right-large-yellowarrow-right-largearrow-right-long-yellowarrow-right-smallfacet-arrow-down-whitefacet-arrow-downCheckedCheckedlink-outmag-glass