Timezone: »
The increasing size and complexity of modern ML systems has improved their predictive capabilities but made their behavior harder to explain. Many techniques for model explanation have been developed in response, but we lack clear criteria for assessing these techniques. In this paper, we cast model explanation as the causal inference problem of estimating causal effects of real-world concepts on the output behavior of ML models given actual input data. We introduce CEBaB, a new benchmark dataset for assessing concept-based explanation methods in Natural Language Processing (NLP). CEBaB consists of short restaurant reviews with human-generated counterfactual reviews in which an aspect (food, noise, ambiance, service) of the dining experience was modified. Original and counterfactual reviews are annotated with multiply-validated sentiment ratings at the aspect-level and review-level. The rich structure of CEBaB allows us to go beyond input features to study the effects of abstract, real-world concepts on model behavior. We use CEBaB to compare the quality of a range of concept-based explanation methods covering different assumptions and conceptions of the problem, and we seek to establish natural metrics for comparative assessments of these methods.
Author Information
Eldar D Abraham (Technion - Israel Institute of Technology)
I’m an M.Sc student at the Technion - Israel Institute of Technology, working on Natural Language Processing and Causal Inference. Sometimes I also play with general deep learning and even optimization algorithms. My advisor is Prof. Roi Reichart.
Karel D'Oosterlinck (Ghent University / Stanford University)

PhD Student at Ghent University. Visiting Student Researcher at Stanford University. Computer Science, Explainable AI.
Amir Feder (Columbia University)
Amir Feder is a Postdoctoral Research Scientist in the Data Science Institute, working with Professor David Blei on causal inference and natural language processing. His research seeks to develop methods that integrate causality into natural language processing, and use them to build linguistically-informed algorithms for predicting and understanding human behavior. Through the paradigm of causal machine learning, Amir aims to build bridges between machine learning and the social sciences. Before joining Columbia, Amir received his PhD from the Technion, where he was advised by Roi Reichart and worked closely with Uri Shalit. In a previous (academic) life, Amir was an economics, statistics and history student at Tel Aviv University, the Hebrew University of Jerusalem and Northwestern University. Amir was the organizer of the First Workshop on Causal Inference and NLP (CI+NLP) at EMNLP 2021.
Yair Gat (Technion)
Atticus Geiger (Stanford University)
Christopher Potts (Stanford University)
Roi Reichart (Technion, Israel Institute of Technology)
Zhengxuan Wu (Stanford University)
More from the Same Authors
-
2021 : ReaSCAN: Compositional Reasoning in Language Grounding »
Zhengxuan Wu · Elisa Kreiss · Desmond Ong · Christopher Potts -
2021 Spotlight: Baleen: Robust Multi-Hop Reasoning at Scale via Condensed Retrieval »
Omar Khattab · Christopher Potts · Matei Zaharia -
2022 : An Invariant Learning Characterization of Controlled Text Generation »
Claudia Shi · Carolina Zheng · Keyon Vafa · Amir Feder · David Blei -
2022 : Useful Confidence Measures: Beyond the Max Score »
Gal Yona · Amir Feder · Itay Laish -
2022 : An Invariant Learning Characterization of Controlled Text Generation »
Claudia Shi · Carolina Zheng · Keyon Vafa · Amir Feder · David Blei -
2023 Poster: Interpretability at Scale: Identifying Causal Mechanisms in Alpaca »
Zhengxuan Wu · Atticus Geiger · Christopher Potts · Noah Goodman -
2023 Poster: Causal-structure Driven Augmentations for Text OOD Generalization »
Amir Feder · Yoav Wald · Claudia Shi · Suchi Saria · David Blei -
2023 Poster: Evaluating the Moral Beliefs Encoded in LLMs »
Nino Scherrer · Claudia Shi · Amir Feder · David Blei -
2022 : An Invariant Learning Characterization of Controlled Text Generation »
Claudia Shi · Carolina Zheng · Keyon Vafa · Amir Feder · David Blei -
2022 Poster: ZeroC: A Neuro-Symbolic Model for Zero-shot Concept Recognition and Acquisition at Inference Time »
Tailin Wu · Megan Tjandrasuwita · Zhengxuan Wu · Xuelin Yang · Kevin Liu · Rok Sosic · Jure Leskovec -
2022 Poster: In the Eye of the Beholder: Robust Prediction with Causal User Modeling »
Amir Feder · Guy Horowitz · Yoav Wald · Roi Reichart · Nir Rosenfeld -
2021 : Intuitive Image Descriptions are Context-Sensitive »
Shayan Hooshmand · Elisa Kreiss · Christopher Potts -
2021 Poster: Causal Abstractions of Neural Networks »
Atticus Geiger · Hanson Lu · Thomas Icard · Christopher Potts -
2021 Poster: Decrypting Cryptic Crosswords: Semantically Complex Wordplay Puzzles as a Target for NLP »
Josh Rozner · Christopher Potts · Kyle Mahowald -
2021 Poster: Dynaboard: An Evaluation-As-A-Service Platform for Holistic Next-Generation Benchmarking »
Zhiyi Ma · Kawin Ethayarajh · Tristan Thrush · Somya Jain · Ledell Wu · Robin Jia · Christopher Potts · Adina Williams · Douwe Kiela -
2021 Poster: Baleen: Robust Multi-Hop Reasoning at Scale via Condensed Retrieval »
Omar Khattab · Christopher Potts · Matei Zaharia -
2021 Poster: On Calibration and Out-of-Domain Generalization »
Yoav Wald · Amir Feder · Daniel Greenfeld · Uri Shalit