Timezone: »
We study the problem of learning a target task when data samples from several auxiliary source tasks are available. Examples of this problem appear in multitask learning, where several tasks are combined jointly, and weak supervision, where multiple programmatic labels are generated for each sample. Because of task data's heterogeneity, negative interference is a critical challenge for solving this problem. Previous works have measured first-order task affinity as an effective metric, yet it becomes less accurate for approximating higher-order transfers. We propose a procedure called task modeling to model first- and higher-order transfers. This procedure samples subsets of source tasks and estimates surrogate functions to approximate multitask predictions. We show theoretical and empirical results that task models can be estimated in nearly-linear time in the number of tasks and accurately approximate multitask predictions. Thus, the target task's performance can be optimized using task models to select source tasks. We validate this approach on various datasets and performance metrics. Our method increases accuracy up to 3.6% over existing methods on five text classification tasks with noisy supervision sources. Additionally, task modeling can be applied to group robustness and fairness metrics. Ablation studies show that task models can accurately predict whether or not a set of up to four source tasks transfer positively to the target task.
Author Information
Dongyue Li (Northeastern University)
Huy Nguyen (Northeastern University)
Hongyang Zhang (Northeastern University)
More from the Same Authors
-
2021 : Correct-N-Contrast: A Contrastive Approach for Improving Robustness to Spurious Correlations »
Michael Zhang · Nimit Sohoni · Hongyang Zhang · Chelsea Finn · Christopher RĂ© -
2023 Poster: Improved Convergence in High Probability of Clipped Gradient Methods with Heavy Tailed Noise »
Ta Duy Nguyen · Thien H Nguyen · Alina Ene · Huy Nguyen -
2023 Poster: On the Generalization Error of Stochastic Mirror Descent for Quadratically-Bounded Losses: an Improved Analysis »
Ta Duy Nguyen · Alina Ene · Huy Nguyen -
2023 Poster: Fast Optimal Locally Private Mean Estimation via Random Projections »
Hilal Asi · Vitaly Feldman · Jelani Nelson · Huy Nguyen · Kunal Talwar -
2023 Poster: Improved Frequency Estimation Algorithms with and without Predictions »
Anders Aamand · Justin Chen · Huy Nguyen · Sandeep Silwal · Ali Vakilian -
2023 Poster: Graph Neural Networks for Road Safety Modeling: Datasets and Evaluations for Accident Analysis »
Abhinav Nippani · Dongyue Li · Haotian Ju · Haris Koutsopoulos · Hongyang Zhang -
2021 Poster: Improved Regularization and Robustness for Fine-tuning in Neural Networks »
Dongyue Li · Hongyang Zhang