Poster
Learning to Pass Expectation Propagation Messages
Nicolas Heess · Daniel Tarlow · John Winn

Fri Dec 6th 07:00 -- 11:59 PM @ Harrah's Special Events Center, 2nd Floor #None

Expectation Propagation (EP) is a popular approximate posterior inference algorithm that often provides a fast and accurate alternative to sampling-based methods. However, while the EP framework in theory allows for complex non-Gaussian factors, there is still a significant practical barrier to using them within EP, because doing so requires the implementation of message update operators, which can be difficult and require hand-crafted approximations. In this work, we study the question of whether it is possible to automatically derive fast and accurate EP updates by learning a discriminative model e.g., a neural network or random forest) to map EP message inputs to EP message outputs. We address the practical concerns that arise in the process, and we provide empirical analysis on several challenging and diverse factors, indicating that there is a space of factors where this approach appears promising.

Author Information

Nicolas Heess (Gatsby Unit)
Daniel Tarlow (Google Brain)
John Winn (Microsoft Research)

More from the Same Authors