Timezone: »
We study the problem of identifying the best action in a sequential decision-making setting when the reward distributions of the arms exhibit a non-trivial dependence structure, which is governed by the underlying causal model of the domain where the agent is deployed. In this setting, playing an arm corresponds to intervening on a set of variables and setting them to specific values. In this paper, we show that whenever the underlying causal model is not taken into account during the decision-making process, the standard strategies of simultaneously intervening on all variables or on all the subsets of the variables may, in general, lead to suboptimal policies, regardless of the number of interventions performed by the agent in the environment. We formally acknowledge this phenomenon and investigate structural properties implied by the underlying causal model, which lead to a complete characterization of the relationships between the arms' distributions. We leverage this characterization to build a new algorithm that takes as input a causal structure and finds a minimal, sound, and complete set of qualified arms that an agent should play to maximize its expected reward. We empirically demonstrate that the new strategy learns an optimal policy and leads to orders of magnitude faster convergence rates when compared with its causal-insensitive counterparts.
Author Information
Sanghack Lee (Purdue University)
Elias Bareinboim (Purdue)
More from the Same Authors
-
2021 : Partition-based Local Independence Discovery »
Inwoo Hwang · Byoung-Tak Zhang · Sanghack Lee -
2021 Poster: Causal Identification with Matrix Equations »
Sanghack Lee · Elias Bareinboim -
2021 Poster: Nested Counterfactual Identification from Arbitrary Surrogate Experiments »
Juan Correa · Sanghack Lee · Elias Bareinboim -
2021 Oral: Causal Identification with Matrix Equations »
Sanghack Lee · Elias Bareinboim -
2020 Poster: Characterizing Optimal Mixed Policies: Where to Intervene and What to Observe »
Sanghack Lee · Elias Bareinboim -
2019 Poster: Near-Optimal Reinforcement Learning in Dynamic Treatment Regimes »
Junzhe Zhang · Elias Bareinboim -
2019 Poster: Efficient Identification in Linear Structural Causal Models with Instrumental Cutsets »
Daniel Kumor · Bryant Chen · Elias Bareinboim -
2019 Poster: Characterization and Learning of Causal Graphs with Latent Variables from Soft Interventions »
Murat Kocaoglu · Amin Jaber · Karthikeyan Shanmugam · Elias Bareinboim -
2019 Poster: Identification of Conditional Causal Effects under Markov Equivalence »
Amin Jaber · Jiji Zhang · Elias Bareinboim -
2019 Spotlight: Identification of Conditional Causal Effects under Markov Equivalence »
Amin Jaber · Jiji Zhang · Elias Bareinboim -
2018 : Datasets and Benchmarks for Causal Learning »
Csaba Szepesvari · Isabelle Guyon · Nicolai Meinshausen · David Blei · Elias Bareinboim · Bernhard Schölkopf · Pietro Perona -
2018 : Causality and Transfer Learning »
Elias Bareinboim -
2018 Poster: Equality of Opportunity in Classification: A Causal Approach »
Junzhe Zhang · Elias Bareinboim -
2017 Poster: Experimental Design for Learning Causal Graphs with Latent Variables »
Murat Kocaoglu · Karthikeyan Shanmugam · Elias Bareinboim -
2016 : The Data-Fusion Problem: Causal Inference and Reinforcement Learning »
Elias Bareinboim -
2015 Poster: Bandits with Unobserved Confounders: A Causal Approach »
Elias Bareinboim · Andrew Forney · Judea Pearl -
2014 Poster: Transportability from Multiple Environments with Limited Experiments: Completeness Results »
Elias Bareinboim · Judea Pearl -
2014 Spotlight: Transportability from Multiple Environments with Limited Experiments: Completeness Results »
Elias Bareinboim · Judea Pearl -
2013 Poster: Transportability from Multiple Environments with Limited Experiments »
Elias Bareinboim · Sanghack Lee · Vasant Honavar · Judea Pearl -
2013 Tutorial: Causes and Counterfactuals: Concepts, Principles and Tools. »
Judea Pearl · Elias Bareinboim