Poster
Memoized Online Variational Inference for Dirichlet Process Mixture Models
Michael Hughes · Erik Sudderth

Thu Dec 5th 07:00 -- 11:59 PM @ Harrah's Special Events Center, 2nd Floor #None

Variational inference algorithms provide the most effective framework for large-scale training of Bayesian nonparametric models. Stochastic online approaches are promising, but are sensitive to the chosen learning rate and often converge to poor local optima. We present a new algorithm, memoized online variational inference, which scales to very large (yet finite) datasets while avoiding the complexities of stochastic gradient. Our algorithm maintains finite-dimensional sufficient statistics from batches of the full dataset, requiring some additional memory but still scaling to millions of examples. Exploiting nested families of variational bounds for infinite nonparametric models, we develop principled birth and merge moves allowing non-local optimization. Births adaptively add components to the model to escape local optima, while merges remove redundancy and improve speed. Using Dirichlet process mixture models for image clustering and denoising, we demonstrate major improvements in robustness and accuracy.

Author Information

Mike Hughes (Tufts University)
Erik Sudderth (Brown University)

More from the Same Authors