Timezone: »
Convolutional neural networks have proven very successful for a wide range of modelling tasks. Convolutional layers embed equivariance to discrete translations into the architectural structure neural networks. Recent extensions generalize this notion to continuous Lie groups beyond translation, such as rotation, scale or more complex symmetries. Another recent generalization of the convolution has allowed for relaxed equivariance constraints, which can be to model data that does not fully respect symmetries while still leveraging on useful inductive biases that equivariances provide. Unlike simple grids for regular convolution over the translational group, sampling convolutional filters on Lie groups requires filters that are continuously parameterised. To parameterise sufficiently flexible continuous filters, small MLP hypernetworks are often used in practice. Although this works, it introduces many additional model parameters. To be more parameter-efficient, we propose an alternative approach defining continuous filters on Lie groups with a small finite set of basis functions through pseudo-points. Regular convolutional layers appear as a special case, allowing for practical conversion between regular filters and our basis function filter formulation, at equal memory complexity. We demonstrate that basis function filters can be used to create efficient equivariant and relaxed-equivariant versions of commonly used neural network architectures, outperforming baselines on CIFAR-10 and CIFAR-100 vision classification tasks.
Author Information
Tycho van der Ouderaa (Imperial College London)
The main topic of my PhD is learning structure and inductive biases in neural networks. The focus has been on learning symmetry from data, such as equivariance and invariance. The aim of the research is to make learning inductive biases and structure in machine learning models as easy as learning the weights.
Mark van der Wilk (Imperial College London)
More from the Same Authors
-
2022 : Actually Sparse Variational Gaussian Processes »
Jake Cunningham · So Takao · Mark van der Wilk · Marc Deisenroth -
2022 : Recommendations for Baselines and Benchmarking Approximate Gaussian Processes »
Sebastian Ober · David Burt · Artem Artemev · Mark van der Wilk -
2022 : Causal Discovery using Marginal Likelihood »
Anish Dhir · Mark van der Wilk -
2022 Poster: Invariance Learning in Deep Neural Networks with Differentiable Laplace Approximations »
Alexander Immer · Tycho van der Ouderaa · Gunnar Rätsch · Vincent Fortuin · Mark van der Wilk -
2022 Poster: SnAKe: Bayesian Optimization with Pathwise Exploration »
Jose Pablo Folch · Shiqiang Zhang · Robert Lee · Behrang Shafei · David Walz · Calvin Tsay · Mark van der Wilk · Ruth Misener -
2022 Poster: Memory safe computations with XLA compiler »
Artem Artemev · Yuze An · Tilman Roeder · Mark van der Wilk -
2022 Poster: Relaxing Equivariance Constraints with Non-stationary Continuous Filters »
Tycho van der Ouderaa · David W. Romero · Mark van der Wilk -
2020 Poster: A Bayesian Perspective on Training Speed and Model Selection »
Clare Lyle · Lisa Schut · Robin Ru · Yarin Gal · Mark van der Wilk -
2020 Poster: Stochastic Segmentation Networks: Modelling Spatially Correlated Aleatoric Uncertainty »
Miguel Monteiro · Loic Le Folgoc · Daniel Coelho de Castro · Nick Pawlowski · Bernardo Marques · Konstantinos Kamnitsas · Mark van der Wilk · Ben Glocker