Timezone: »

Riemannian batch normalization for SPD neural networks
Daniel Brooks · Olivier Schwander · Frederic Barbaresco · Jean-Yves Schneider · Matthieu Cord

Wed Dec 11 10:45 AM -- 12:45 PM (PST) @ East Exhibition Hall B + C #131

Covariance matrices have attracted attention for machine learning applications due to their capacity to capture interesting structure in the data. The main challenge is that one needs to take into account the particular geometry of the Riemannian manifold of symmetric positive definite (SPD) matrices they belong to. In the con- text of deep networks, several architectures for these matrices have recently been proposed. In our article, we introduce a Riemannian batch normalization (batch- norm) algorithm, which generalizes the one used in Euclidean nets. This novel layer makes use of geometric operations on the manifold, notably the Riemannian barycenter, parallel transport and non-linear structured matrix transformations. We derive a new manifold-constrained gradient descent algorithm working in the space of SPD matrices, allowing to learn the batchnorm layer. We validate our proposed approach with experiments in three different contexts on diverse data types: a drone recognition dataset from radar observations, and on emotion and action recognition datasets from video and motion capture data. Experiments show that the Riemannian batchnorm systematically gives better classification performance compared with leading methods and a remarkable robustness to lack of data.

Author Information

Daniel Brooks (Thales - LIP6)
Olivier Schwander (Sorbonne Université)
Frederic Barbaresco (THALES LAND & AIR SYSTEMS)
Jean-Yves Schneider (THALES LAND & AIR SYSTEMS)
Matthieu Cord (Sorbonne University)

More from the Same Authors