Timezone: »
Datasets can be biased due to societal inequities, human biases, under-representation of minorities, etc. Our goal is to certify that models produced by a learning algorithm are pointwise-robust to dataset biases. This is a challenging problem: it entails learning models for a large, or even infinite, number of datasets, ensuring that they all produce the same prediction. We focus on decision-tree learning due to the interpretable nature of the models. Our approach allows programmatically specifying \emph{bias models} across a variety of dimensions (e.g., label-flipping or missing data), composing types of bias, and targeting bias towards a specific group. To certify robustness, we use a novel symbolic technique to evaluate a decision-tree learner on a large, or infinite, number of datasets, certifying that each and every dataset produces the same prediction for a specific test point. We evaluate our approach on datasets that are commonly used in the fairness literature, and demonstrate our approach's viability on a range of bias models.
Author Information
Anna Meyer (University of Wisconsin - Madison)
Aws Albarghouthi (University of Wisconsin, Madison)
Loris D'Antoni (University of Wisconsin, Madison)
More from the Same Authors
-
2022 Poster: BagFlip: A Certified Defense Against Data Poisoning »
Yuhao Zhang · Aws Albarghouthi · Loris D'Antoni -
2022 Poster: AutoWS-Bench-101: Benchmarking Automated Weak Supervision with 100 Labels »
Nicholas Roberts · Xintong Li · Tzu-Heng Huang · Dyah Adila · Spencer Schoenberg · Cheng-Yu Liu · Lauren Pick · Haotian Ma · Aws Albarghouthi · Frederic Sala