Skip to yearly menu bar Skip to main content


Oral
in
Workshop: eXplainable AI approaches for debugging and diagnosis

[O3] Reinforcement Explanation Learning

Siddhant Agarwal · OWAIS IQBAL · Sree Aditya Buridi · Madda Manjusha · Abir Das


Abstract:

Deep Learning has become overly complicated and has enjoyed stellar success in solving several classical problems like image classification, object detection, etc. Several methods for explaining these decisions have been proposed. Black box methods to generate saliency maps are particularly interesting due to the fact that they do not utilize the internals of the model to explain the decision. Major black box methods perturb the input and observe the changes in the output. We formulate saliency map generation as a sequential search problem and leverage upon Reinforcement Learning (RL) to accumulate evidence from input images that most strongly support decisions made by a classifier. Such a strategy encourages to search \emph{intelligently} for the perturbations that will lead to high-quality explanations. While successful black box explanation approaches need to rely on heavy computations and suffer from small sample approximation, the deterministic policy learned by our method makes it a lot more efficient during the inference. Experiments on three benchmark datasets demonstrate the superiority of the proposed approach in inference time over state-of-the-arts without hurting the performance. The anonymized code can be found at https://anonymous.4open.science/r/RExL-88F8