Timezone: »
Learning from label proportions (LLP) is a weakly supervised setting for classification in which unlabeled training instances are grouped into bags, and each bag is annotated with the proportion of each class occurring in that bag. Prior work on LLP has yet to establish a consistent learning procedure, nor does there exist a theoretically justified, general purpose training criterion. In this work we address these two issues by posing LLP in terms of mutual contamination models (MCMs), which have recently been applied successfully to study various other weak supervision settings. In the process, we establish several novel technical results for MCMs, including unbiased losses and generalization error bounds under non-iid sampling plans. We also point out the limitations of a common experimental setting for LLP, and propose a new one based on our MCM framework.
Author Information
Clayton Scott (University of Michigan)
Jianxin Zhang (University of Michigan)
More from the Same Authors
-
2020 Poster: Consistent Estimation of Identifiable Nonparametric Mixture Models from Grouped Observations »
Alexander Ritchie · Robert Vandermeulen · Clayton Scott -
2020 Poster: Weston-Watkins Hinge Loss and Ordered Partitions »
Yutong Wang · Clayton Scott -
2017 Poster: Multi-Task Learning for Contextual Bandits »
Aniket Anand Deshmukh · Urun Dogan · Clay Scott -
2014 Poster: Robust Kernel Density Estimation by Scaling and Projection in Hilbert Space »
Robert A Vandermeulen · Clayton Scott -
2011 Poster: Generalizing from Several Related Classification Tasks to a New Unlabeled Sample »
Gilles Blanchard · Gyemin Lee · Clayton Scott -
2010 Poster: Extensions of Generalized Binary Search to Group Identification and Exponential Costs »
Gowtham Bellala · Suresh Bhavnani · Clayton Scott -
2008 Poster: Performance analysis for L_2 kernel classification »
JooSeuk Kim · Clayton Scott -
2008 Spotlight: Performance analysis for L_2 kernel classification »
JooSeuk Kim · Clayton Scott