Timezone: »

Gaussian Process Prior Variational Autoencoders
Francesco Paolo Casale · Adrian Dalca · Luca Saglietti · Jennifer Listgarten · Nicolo Fusi

Thu Dec 07:45 AM -- 09:45 AM PST @ Room 210 #63

Variational autoencoders (VAE) are a powerful and widely-used class of models to learn complex data distributions in an unsupervised fashion. One important limitation of VAEs is the prior assumption that latent sample representations are independent and identically distributed. However, for many important datasets, such as time-series of images, this assumption is too strong: accounting for covariances between samples, such as those in time, can yield to a more appropriate model specification and improve performance in downstream tasks. In this work, we introduce a new model, the Gaussian Process (GP) Prior Variational Autoencoder (GPPVAE), to specifically address this issue. The GPPVAE aims to combine the power of VAEs with the ability to model correlations afforded by GP priors. To achieve efficient inference in this new class of models, we leverage structure in the covariance matrix, and introduce a new stochastic backpropagation strategy that allows for computing stochastic gradients in a distributed and low-memory fashion. We show that our method outperforms conditional VAEs (CVAEs) and an adaptation of standard VAEs in two image data applications.

Author Information

Francesco Paolo Casale (Microsoft Research)
Adrian Dalca (MIT)
Luca Saglietti (Microsoft Research New England (visitor) Italian Institute for Genomic Medicine, Torino, Italy)
Jennifer Listgarten (UC Berkeley)
Nicolo Fusi (Microsoft Research)

More from the Same Authors