Timezone: »
An oft-cited challenge of federated learning is the presence of heterogeneity. \emph{Data heterogeneity} refers to the fact that data from different clients may follow very different distributions. \emph{System heterogeneity} refers to the fact that client devices have different system capabilities. A considerable number of federated optimization methods address this challenge. In the literature, empirical evaluations usually start federated training from random initialization. However, in many practical applications of federated learning, the server has access to proxy data for the training task that can be used to pre-train a model before starting federated training. We empirically study the impact of starting from a pre-trained model in federated learning using four standard federated learning benchmark datasets. Unsurprisingly, starting from a pre-trained model reduces the training time required to reach a target error rate and enables the training of more accurate models (up to 40\%) than is possible when starting from random initialization. Surprisingly, we also find that starting federated learning from a pre-trained initialization reduces the effect of both data and system heterogeneity. We recommend that future work proposing and evaluating federated optimization methods evaluate the performance when starting from random and pre-trained initializations. We also believe this study raises several questions for further work on understanding the role of heterogeneity in federated optimization.
Author Information
John Nguyen (Facebook)
Jianyu Wang (Meta)
Kshitiz Malik (University of Illinois, Urbana-Champaign)
Maziar Sanjabi (Facebook AI)
Mike Rabbat (Facebook FAIR)
More from the Same Authors
-
2021 : Opacus: User-Friendly Differential Privacy Library in PyTorch »
Ashkan Yousefpour · Igor Shilov · Alexandre Sablayrolles · Karthik Prasad · Mani Malek Esmaeili · John Nguyen · Sayan Ghosh · Akash Bharadwaj · Jessica Zhao · Graham Cormode · Ilya Mironov -
2021 : Stochastic Polyak Stepsize with a Moving Target »
Robert Gower · Aaron Defazio · Mike Rabbat -
2022 : The Interpolated MVU Mechanism For Communication-efficient Private Federated Learning »
Chuan Guo · Kamalika Chaudhuri · Pierre STOCK · Mike Rabbat -
2022 : FedSynth: Gradient Compression via Synthetic Data in Federated Learning »
Shengyuan Hu · Jack Goetz · Kshitiz Malik · Hongyuan Zhan · Zhe Liu · Yue Liu -
2022 : Federated Learning under Distributed Concept Drift »
Ellango Jothimurugesan · Kevin Hsieh · Jianyu Wang · Gauri Joshi · Phillip Gibbons -
2022 : On the Unreasonable Effectiveness of Federated Averaging with Heterogenous Data »
Jianyu Wang -
2019 Poster: Gossip-based Actor-Learner Architectures for Deep Reinforcement Learning »
Mahmoud Assran · Joshua Romoff · Nicolas Ballas · Joelle Pineau · Mike Rabbat -
2018 Poster: On the Convergence and Robustness of Training GANs with Regularized Optimal Transport »
Maziar Sanjabi · Jimmy Ba · Meisam Razaviyayn · Jason Lee -
2017 Poster: Federated Multi-Task Learning »
Virginia Smith · Chao-Kai Chiang · Maziar Sanjabi · Ameet S Talwalkar