Skip to yearly menu bar Skip to main content


Poster

Deep Homogeneous Mixture Models: Representation, Separation, and Approximation

Priyank Jaini · Pascal Poupart · Yaoliang Yu

Room 210 #85

Keywords: [ Latent Variable Models ] [ Hierarchical Models ] [ Generative Models ] [ Density Estimation ] [ Graphical Models ]


Abstract: At their core, many unsupervised learning models provide a compact representation of homogeneous density mixtures, but their similarities and differences are not always clearly understood. In this work, we formally establish the relationships among latent tree graphical models (including special cases such as hidden Markov models and tensorial mixture models), hierarchical tensor formats and sum-product networks. Based on this connection, we then give a unified treatment of exponential separation in \emph{exact} representation size between deep mixture architectures and shallow ones. In contrast, for \emph{approximate} representation, we show that the conditional gradient algorithm can approximate any homogeneous mixture within $\epsilon$ accuracy by combining $O(1/\epsilon^2)$ ``shallow'' architectures, where the hidden constant may decrease (exponentially) with respect to the depth. Our experiments on both synthetic and real datasets confirm the benefits of depth in density estimation.

Live content is unavailable. Log in and register to view live content