Poster

Deep Combinatorial Aggregation

Yuesong Shen · Daniel Cremers

Keywords: [ deep combinatorial aggregation ] [ consistency enforcing loss ] [ deep combinatorial weight averaging ] [ Uncertainty Estimation ]

[ Abstract ]
[ Poster [ OpenReview
 
Spotlight presentation: Lightning Talks 3B-1
Wed 7 Dec 9 a.m. PST — 9:15 a.m. PST

Abstract:

Neural networks are known to produce poor uncertainty estimations, and a variety of approaches have been proposed to remedy this issue. This includes deep ensemble, a simple and effective method that achieves state-of-the-art results for uncertainty-aware learning tasks. In this work, we explore a combinatorial generalization of deep ensemble called deep combinatorial aggregation (DCA). DCA creates multiple instances of network components and aggregates their combinations to produce diversified model proposals and predictions. DCA components can be defined at different levels of granularity. And we discovered that coarse-grain DCAs can outperform deep ensemble for uncertainty-aware learning both in terms of predictive performance and uncertainty estimation. For fine-grain DCAs, we discover that an average parameterization approach named deep combinatorial weight averaging (DCWA) can improve the baseline training. It is on par with stochastic weight averaging (SWA) but does not require any custom training schedule or adaptation of BatchNorm layers. Furthermore, we propose a consistency enforcing loss that helps the training of DCWA and modelwise DCA. We experiment on in-domain, distributional shift, and out-of-distribution image classification tasks, and empirically confirm the effectiveness of DCWA and DCA approaches.

Chat is not available.