Timezone: »
Transfer learning aims to leverage models pre-trained on source data to efficiently adapt to target setting, where only limited data are available for model fine-tuning. Recent works empirically demonstrate that adversarial training in the source data can improve the ability of models to transfer to new domains. However, why this happens is not known. In this paper, we provide a theoretical model to rigorously analyze how adversarial training helps transfer learning. We show that adversarial training in the source data generates provably better representations, so fine-tuning on top of this representation leads to a more accurate predictor of the target data. We further demonstrate both theoretically and empirically that semi-supervised learning in the source data can also improve transfer learning by similarly improving the representation. Moreover, performing adversarial training on top of semi-supervised learning can further improve transferability, suggesting that the two approaches have complementary benefits on representations. We support our theories with experiments on popular data sets and deep learning architectures.
Author Information
Zhun Deng (Harvard University)
Linjun Zhang (Rutgers University)
Kailas Vodrahalli (Stanford University)
Kenji Kawaguchi (MIT)
James Zou (Stanford)
More from the Same Authors
-
2021 Spotlight: A Central Limit Theorem for Differentially Private Query Answering »
Jinshuo Dong · Weijie Su · Linjun Zhang -
2021 : Catastrophic Failures of Neural Active Learning on Heteroskedastic Distributions »
Savya Khosla · Alex Lamb · Jordan Ash · Cyril Zhang · Kenji Kawaguchi -
2021 : Noether Networks: Meta-Learning Useful Conserved Quantities »
Ferran Alet · Dylan Doblar · Allan Zhou · Josh Tenenbaum · Kenji Kawaguchi · Chelsea Finn -
2022 : Predicting Immune Escape with Pretrained Protein Language Model Embeddings »
Kyle Swanson · Howard Chang · James Zou -
2022 : Data-driven subgroup identification for linear regression »
Zachary Izzo · Ruishan Liu · James Zou -
2022 : Is Unsupervised Performance Estimation Impossible When Both Covariates and Labels shift? »
Lingjiao Chen · Matei Zaharia · James Zou -
2022 : DrML: Diagnosing and Rectifying Vision Models using Language »
Yuhui Zhang · Jeff Z. HaoChen · Shih-Cheng Huang · Kuan-Chieh Wang · James Zou · Serena Yeung -
2022 : Provable Re-Identification Privacy »
Zachary Izzo · Jinsung Yoon · Sercan Arik · James Zou -
2022 : Recommendation for New Drugs with Limited Prescription Data »
Zhenbang Wu · Huaxiu Yao · Zhe Su · David Liebovitz · Lucas Glass · James Zou · Chelsea Finn · Jimeng Sun -
2022 Panel: Panel 1C-7: Beyond Adult and… & Uncalibrated Models Can… »
Kailas Vodrahalli · Flavio Calmon -
2022 : An Electrocardiogram-Based Risk Score for Cardiovascular Mortality »
John Hughes · David Ouyang · Pierre Elias · James Zou · Euan Ashley · Marco Perez -
2022 : An Electrocardiogram-Based Risk Score for Cardiovascular Mortality »
John Hughes · David Ouyang · Pierre Elias · James Zou · Euan Ashley · Marco Perez -
2022 Poster: Estimating and Explaining Model Performance When Both Covariates and Labels Shift »
Lingjiao Chen · Matei Zaharia · James Zou -
2022 Poster: SkinCon: A skin disease dataset densely annotated by domain experts for fine-grained debugging and analysis »
Roxana Daneshjou · Mert Yuksekgonul · Zhuo Ran Cai · Roberto Novoa · James Zou -
2022 Poster: HAPI: A Large-scale Longitudinal Dataset of Commercial ML API Predictions »
Lingjiao Chen · Zhihua Jin · Evan Sabri Eyuboglu · Christopher Ré · Matei Zaharia · James Zou -
2022 Poster: Uncalibrated Models Can Improve Human-AI Collaboration »
Kailas Vodrahalli · Tobias Gerstenberg · James Zou -
2022 Poster: C-Mixup: Improving Generalization in Regression »
Huaxiu Yao · Yiping Wang · Linjun Zhang · James Zou · Chelsea Finn -
2022 Poster: Mind the Gap: Understanding the Modality Gap in Multi-modal Contrastive Representation Learning »
Victor Weixin Liang · Yuhui Zhang · Yongchan Kwon · Serena Yeung · James Zou -
2022 Poster: WeightedSHAP: analyzing and improving Shapley based feature attributions »
Yongchan Kwon · James Zou -
2021 Poster: Understanding End-to-End Model-Based Reinforcement Learning Methods as Implicit Parameterization »
Clement Gehring · Kenji Kawaguchi · Jiaoyang Huang · Leslie Kaelbling -
2021 Poster: A Central Limit Theorem for Differentially Private Query Answering »
Jinshuo Dong · Weijie Su · Linjun Zhang -
2021 Poster: EIGNN: Efficient Infinite-Depth Graph Neural Networks »
Juncheng Liu · Kenji Kawaguchi · Bryan Hooi · Yiwei Wang · Xiaokui Xiao -
2021 Poster: Noether Networks: meta-learning useful conserved quantities »
Ferran Alet · Dylan Doblar · Allan Zhou · Josh Tenenbaum · Kenji Kawaguchi · Chelsea Finn -
2021 Poster: Tailoring: encoding inductive biases by optimizing unsupervised objectives at prediction time »
Ferran Alet · Maria Bauza · Kenji Kawaguchi · Nurullah Giray Kuru · Tomás Lozano-Pérez · Leslie Kaelbling -
2021 Poster: Discrete-Valued Neural Communication »
Dianbo Liu · Alex Lamb · Kenji Kawaguchi · Anirudh Goyal · Chen Sun · Michael Mozer · Yoshua Bengio -
2020 Session: Orals & Spotlights Track 02: COVID/Health/Bio Applications »
Tristan Naumann · James Zou -
2019 Poster: Making AI Forget You: Data Deletion in Machine Learning »
Antonio Ginart · Melody Guan · Gregory Valiant · James Zou -
2019 Spotlight: Making AI Forget You: Data Deletion in Machine Learning »
Antonio Ginart · Melody Guan · Gregory Valiant · James Zou -
2017 Workshop: Machine Learning in Computational Biology »
James Zou · Anshul Kundaje · Gerald Quon · Nicolo Fusi · Sara Mostafavi -
2017 Poster: NeuralFDR: Learning Discovery Thresholds from Hypothesis Features »
Fei Xia · Martin J Zhang · James Zou · David Tse -
2016 Poster: Deep Learning without Poor Local Minima »
Kenji Kawaguchi -
2016 Oral: Deep Learning without Poor Local Minima »
Kenji Kawaguchi -
2015 Poster: Bayesian Optimization with Exponential Convergence »
Kenji Kawaguchi · Leslie Kaelbling · Tomás Lozano-Pérez