Timezone: »
Designing learning systems which are invariant to certain data transformations is critical in machine learning. Practitioners can typically enforce a desired invariance on the trained model through the choice of a network architecture, e.g. using convolutions for translations, or using data augmentation. Yet, enforcing true invariance in the network can be difficult, and data invariances are not always known a piori. State-of-the-art methods for learning data augmentation policies require held-out data and are based on bilevel optimization problems, which are complex to solve and often computationally demanding. In this work we investigate new ways of learning invariances only from the training data. Using learnable augmentation layers built directly in the network, we demonstrate that our method is very versatile. It can incorporate any type of differentiable augmentation and be applied to a broad class of learning problems beyond computer vision. We provide empirical evidence showing that our approach is easier and faster to train than modern automatic data augmentation techniques based on bilevel optimization, while achieving comparable results. Experiments show that while the invariances transferred to a model through automatic data augmentation are limited by the model expressivity, the invariance yielded by our approach is insensitive to it by design.
Author Information
Cédric ROMMEL (INRIA - MIND team)
I am currently a postdoctoral researcher in the [Parietal team](https://team.inria.fr/parietal/) (INRIA), working in deep learning and neuroscience under the supervision of [Thomas Moreau](https://tommoral.github.io/about.html) and [Alexandre Gramfort](https://alexandre.gramfort.net/). My research currently revolves around the idea of learning and exploiting data invariances to make deep neural networks more data efficient and robust to domain changes. This includes for example learning optimal data augmentations directly from datasets for which those are not intuitive, such as brain electrical signals. I am also interested in using learned invariances as a tool to better understand how some types of information (e.g. sleep stages) are encoded within neurological signals. Previously, I was the scientific and engineering leader of the AI team at [Ava](https://www.ava.me/), working in speaker recognition technology for deaf and hard-of-hearing accessibility. Me and my team were then mainly focused on real-time speaker recognition and diarization with multiple microphones, which involved works in zero-shot learning, metric learning and neural architectures for speech processing. I obtained my PhD in applied mathematics at [Ecole Polytechnique](https://www.polytechnique.edu/) ([CMAP](https://portail.polytechnique.edu/cmap/fr)) and [INRIA](https://team.inria.fr/commands/), under the supervision of [Frédéric Bonnans](http://www.cmap.polytechnique.fr/~bonnans/) and [Pierre Martinon](http://www.cmapx.polytechnique.fr/~martinon/). My work lied in the intersection between optimal control, machine learning and optimization. My main interest was to learn interpretable and physically plausible models of dynamical systems, and how to optimally control them taking model uncertainty into account. My thesis was funded by the aviation start up [Safety Line](https://www.safety-line.fr/) and the main application of my work was the optimization of real aircraft trajectories for fuel consumption reduction. My algorithms integrated the product [OptiClimb](https://www.sita.aero/solutions/sita-for-aircraft/digital-day-of-operations/opticlimb/), which is currently used daily to compute fuel efficient flights all over the globe by companies such as AirFrance. Before my PhD thesis I studied at [MINES ParisTech](https://www.mines-paristech.fr/) where I obtained a MSc. in engineering and applied mathematics.
Thomas Moreau (Inria)
Alexandre Gramfort (Meta)
More from the Same Authors
-
2021 : Electromagnetic neural source imaging under sparsity constraints with SURE-based hyperparameter tuning »
Pierre-Antoine Bannier · Quentin Bertrand · Joseph Salmon · Alexandre Gramfort -
2022 : Validation Diagnostics for SBI algorithms based on Normalizing Flows »
Julia Linhart · Alexandre Gramfort · Pedro Rodrigues -
2022 Poster: Benchopt: Reproducible, efficient and collaborative optimization benchmarks »
Thomas Moreau · Mathurin Massias · Alexandre Gramfort · Pierre Ablin · Pierre-Antoine Bannier · Benjamin Charlier · Mathieu Dagréou · Tom Dupre la Tour · Ghislain DURIF · Cassio F. Dantas · Quentin Klopfenstein · Johan Larsson · En Lai · Tanguy Lefort · Benoît Malézieux · Badr MOUFAD · Binh T. Nguyen · Alain Rakotomamonjy · Zaccharie Ramzi · Joseph Salmon · Samuel Vaiter -
2022 Poster: A framework for bilevel optimization that enables stochastic and global variance reduction algorithms »
Mathieu Dagréou · Pierre Ablin · Samuel Vaiter · Thomas Moreau -
2022 Poster: Toward a realistic model of speech processing in the brain with self-supervised learning »
Juliette MILLET · Charlotte Caucheteux · pierre orhan · Yves Boubenec · Alexandre Gramfort · Ewan Dunbar · Christophe Pallier · Jean-Remi King -
2021 Poster: HNPE: Leveraging Global Parameters for Neural Posterior Estimation »
Pedro Rodrigues · Thomas Moreau · Gilles Louppe · Alexandre Gramfort -
2021 : The NeurIPS 2021 BEETL Competition: Benchmarks for EEG Transfer Learning + Q&A »
Xiaoxi Wei · Vinay Jayaram · Sylvain Chevallier · Giulia Luise · Camille Jeunet · Moritz Grosse-Wentrup · Alexandre Gramfort · Aldo A Faisal -
2021 Poster: Shared Independent Component Analysis for Multi-Subject Neuroimaging »
Hugo Richard · Pierre Ablin · Bertrand Thirion · Alexandre Gramfort · Aapo Hyvarinen -
2020 Poster: Learning to solve TV regularised problems with unrolled algorithms »
Hamza Cherkaoui · Jeremias Sulam · Thomas Moreau -
2020 Poster: Modeling Shared responses in Neuroimaging Studies through MultiView ICA »
Hugo Richard · Luigi Gresele · Aapo Hyvarinen · Bertrand Thirion · Alexandre Gramfort · Pierre Ablin -
2020 Spotlight: Modeling Shared responses in Neuroimaging Studies through MultiView ICA »
Hugo Richard · Luigi Gresele · Aapo Hyvarinen · Bertrand Thirion · Alexandre Gramfort · Pierre Ablin -
2020 Poster: NeuMiss networks: differentiable programming for supervised learning with missing values. »
Marine Le Morvan · Julie Josse · Thomas Moreau · Erwan Scornet · Gael Varoquaux -
2020 Oral: NeuMiss networks: differentiable programming for supervised learning with missing values. »
Marine Le Morvan · Julie Josse · Thomas Moreau · Erwan Scornet · Gael Varoquaux -
2020 Poster: Statistical control for spatio-temporal MEG/EEG source imaging with desparsified mutli-task Lasso »
Jerome-Alexis Chevalier · Joseph Salmon · Alexandre Gramfort · Bertrand Thirion -
2019 Poster: Handling correlated and repeated measurements with the smoothed multivariate square-root Lasso »
Quentin Bertrand · Mathurin Massias · Alexandre Gramfort · Joseph Salmon -
2019 Poster: Learning step sizes for unfolded sparse coding »
Pierre Ablin · Thomas Moreau · Mathurin Massias · Alexandre Gramfort -
2019 Poster: Manifold-regression to predict from MEG/EEG brain signals without source modeling »
David Sabbagh · Pierre Ablin · Gael Varoquaux · Alexandre Gramfort · Denis A. Engemann -
2018 Poster: Multivariate Convolutional Sparse Coding for Electromagnetic Brain Signals »
Tom Dupré la Tour · Thomas Moreau · Mainak Jas · Alexandre Gramfort -
2017 Poster: Learning the Morphology of Brain Signals Using Alpha-Stable Convolutional Sparse Coding »
Mainak Jas · Tom Dupré la Tour · Umut Simsekli · Alexandre Gramfort -
2016 Poster: GAP Safe Screening Rules for Sparse-Group Lasso »
Eugene Ndiaye · Olivier Fercoq · Alexandre Gramfort · Joseph Salmon -
2015 Poster: GAP Safe screening rules for sparse multi-task and multi-class models »
Eugene Ndiaye · Olivier Fercoq · Alexandre Gramfort · Joseph Salmon -
2010 Poster: Brain covariance selection: better individual functional connectivity models using population prior »
Gaël Varoquaux · Alexandre Gramfort · Jean-Baptiste Poline · Bertrand Thirion