`

Timezone: »

 
Optimal Representations for Covariate Shifts
Yann Dubois · Yangjun Ruan · Chris Maddison
Event URL: https://openreview.net/forum?id=de1kSNxv5BQ »

Machine learning often experiences distribution shifts between training and testing. We introduce a simple objective whose optima are \textit{exactly all} representations on which risk minimizers are guaranteed to be robust to Bayes preserving shifts, e.g., covariate shifts. Our objective has two components. First, a representation must remain discriminative, i.e., some predictor must be able to minimize the source and target risk. Second, the representation's support should be invariant across source and target. We make this practical by designing self-supervised methods that only use unlabelled data and augmentations. Our objectives achieve SOTA on DomainBed, and give insights into the robustness of recent methods, e.g., CLIP.

Author Information

Yann Dubois (Stanford University)
Yangjun Ruan (University of Toronto)
Chris Maddison (University of Toronto)

More from the Same Authors