Skip to yearly menu bar Skip to main content


Poster
in
Workshop: Mathematics of Modern Machine Learning (M3L)

First-order ANIL provably learns representations despite overparametrisation

Oguz Kaan Yuksel · Etienne Boursier · Nicolas Flammarion


Abstract:

Meta-learning methods leverage data from previous tasks to learn a new task in a sample-efficient manner. In particular, model-agnostic methods look for initialisation points from which gradient descent quickly adapts to any new task.Although it has been empirically suggested that such methods learns shared representations during pretraining, there is limited theoretical evidence of such behavior. In this direction, this work shows, in the limit of infinite tasks, first-order ANIL with a linear two-layer network successfully learns linear shared representations. This result even holds under overparametrisation; having a width larger than the dimension of the shared representations results in an asymptotically low-rank solution.

Chat is not available.