Timezone: »

Deep Class-Conditional Gaussians for Continual Learning
Thomas Lee · Amos Storkey
Event URL: https://openreview.net/forum?id=MOGpMNc_NdP »

The current state of the art for continual learning with frozen, pre-trained embedding networks are simple probabilistic models defined over the embedding space, for example class conditional Gaussians. As yet, in the task-incremental online setting, it has been an open question how to extend these methods to when the embedding function has to be learned from scratch. In this paper, we propose DeepCCG, an empirical Bayesian method which learns online both a class conditional Gaussian model and an embedding function. The learning process can be interpreted as using a variant of experience replay, known to be effective in continual learning. As part of our framework, we decide which examples to store by selecting the subset that minimises the KL divergence between the true posterior and the posterior induced by the subset. We demonstrate performance task-incremental online settings, including those with overlapping tasks. Our method outperforms all other methods, including several other replay-based methods.

Author Information

Thomas Lee (University of Edinbrugh)
Amos Storkey (University of Edinburgh)

More from the Same Authors