Research‎ > ‎

Memoized Online Variational Inference for Dirichlet Process Mixture Models

AUTHORS
Michael Hughes
Erik Sudderth

FULL TEXT: PDF

ABSTRACT
Variational inference algorithms provide the most effective framework for large-scale training  of Bayesian nonparametric models. Stochastic online approaches are promising, but are sensitive to the chosen learning rate and often converge to poor local optima.  We present a new algorithm, memoized online variational inference, which scales to very large (yet finite) datasets while avoiding the complexities of stochastic gradient.  Our algorithm maintains finite-dimensional sufficient statistics from batches of the full dataset, requiring some additional memory but still scaling to millions of examples.  Exploiting nested families of variational bounds for infinite nonparametric models, we develop principled birth and merge moves allowing non-local optimization.  Births adaptively add components to the model to escape local optima, while merges remove redundancy and improve speed.  Using Dirichlet process mixture models for image clustering and denoising, we demonstrate major improvements in robustness and accuracy.

CODE
Available online: bitbucket.org/michaelchughes/bnpy/

Ċ
Mike Hughes,
Dec 5, 2013, 2:37 PM
Comments