Close Menu

Stochastic Linear Algebra for Scalable Gaussian Processes


Jan 28, 2019 - 1:50pm to 2:55pm


Rettaliata Engineering Center, Room 104


David Bindel
Department of Computer Science, Cornell University


Gaussian processes (GPs) define a distribution over functions that generalizes the multivariate normal distribution over vector spaces. Long used as a tool for spatio-temporal statistical modeling, GPs are also a key part of the modern arsenal in machine learning. Unfortunately, Gaussian process regression and kernel hyper-parameter estimation with \(N\) training examples involve manipulating a dense \(N\times N\)kernel matrix, and standard factorization-based approaches to the underlying linear algebra problems have \(O(N^{3})\) scaling. For regression with a fixed covariance kernel, more scalable iterative methods based on fast matrix-vector multiplication with the kernel matrices are available. However, maximum likelihood estimation of kernel hyper-parameters and computation of conditional variances involve operations such as computing log derivatives and their derivatives or extracting the diagonal part of a Schur complement.  New tools are needed to address these problems in a scalable manner. In this talk, we discuss our recent work on one such set of tools, based on a combination of Krylov subspace methods for matrix solves. 
Joint work with Kun Dong, David Eriksson, and Andrew Wilson.

Event Type: 

Department of Applied Mathematics - Colloquia