Contrastive Explanation for Machine Learning

Research output: Other contributionAcademic

Abstract

Introduction. Recent advances in Interpretable Machine Learning (iML) and Explainable Artificial Intelligence (XAI) have shown promising approaches that are able to provide human-understandable explanations. However, these approaches have also been criticized for disregarding human behavior in explanation. When humans ask for an explanation they generally contrast the given output against an output of interest. We propose to use this human tendency to ask questions like `Why this output (the fact) instead of the other (the foil)?' as a natural way of limiting an explanation to its key causes.

Method. In this study we present an end-to-end approach for extracting contrastive explanations for machine learning (ML). First, we define how to apply contrastive explanation to ML. Next, we extensively study 84 iML methods in a systematic literature review to overview approaches for enhancing interpretability in machine learning, and finding method parts most suitable for contrastive explanations. We develop Foil Trees: a model-agnostic approach to extracting explanations for finding the set of rules that causes the explanation to be predicted the actual outcome (fact) instead of the other (foil).

Results. Quantitative validation showed that Foil Trees are able to accurately mimic the decision boundaries of the model it aims to explain (94% fidelity), generalizes well on unseen data (88% accuracy), provides 78% shorter explanations than their non-contrastive counterparts (mean length of 1.19 over 5.37) and does this all in real-time (60ms on average per explanation). Moreover, we conducted a user experiment on 121 participants to establish how contrastive and non-contrastive explanations are perceived in terms of general preference, transparency and trust. We found that contrastive explanations are preferred over non-contrastive explanations in terms of understandability, informativeness of contents and alignment with their own decision-making. This preference lead to an increased general preference and willingness to act upon the decision.

Discussion. These results suggest that it is feasible to extract generalizable, objectively transparent contrastive explanations for ML, and that contrastive explanations provide an intuitive means to create informative minimal-length human-understandable explanations that are preferable and more persuasive.
Original languageEnglish
TypeMaster thesis
PublisherUtrecht University
Number of pages176
Publication statusPublished - Jul 2018

Fingerprint

Dive into the research topics of 'Contrastive Explanation for Machine Learning'. Together they form a unique fingerprint.

Cite this