Abstract
We present GLOBAL CAUSAL ANALYSIS (GCA) for text classification. GCA is a technique for global model-agnostic explainability drawing from well-established observational causal structure learning algorithms. GCA generates an explanatory graph from high-level human-interpretable features, revealing how these features affect each other and the black-box output. We show how these high-level features do not always have to be human-annotated, but can also be computationally inferred. Moreover, we discuss how the explanatory graph can be used for global model analysis in natural language processing (NLP): the graph shows the effect of different types of features on model behavior, whether these effects are causal effects or mere (spurious) correlations, and if and how different features interact. We then propose a three-step method for (semi-)automatically evaluating the quality, fidelity and stability of the GCA explanatory graph without requiring a ground truth. Finally, we provide a detailed GCA of a state-of-the-art NLP model, showing how setting a global one-versus-rest contrast can improve explanatory relevance, and demonstrating the utility of our three-step evaluation method.
| Original language | English |
|---|---|
| Title of host publication | Explainable Artificial Intelligence |
| Subtitle of host publication | First World Conference, xAI 2023, Lisbon, Portugal, July 26–28, 2023, Proceedings, Part I |
| Editors | Luca Longo |
| Place of Publication | Cham |
| Publisher | Springer |
| Pages | 299–323 |
| Number of pages | 25 |
| Edition | 1 |
| ISBN (Electronic) | 978-3-031-44064-9 |
| ISBN (Print) | 978-3-031-44063-2 |
| DOIs | |
| Publication status | Published - 30 Oct 2023 |
Publication series
| Name | Communications in Computer and Information Science |
|---|---|
| Publisher | Springer |
| Volume | 1901 |
| ISSN (Print) | 1865-0929 |
| ISSN (Electronic) | 1865-0937 |
Bibliographical note
Publisher Copyright:© 2023, The Author(s), under exclusive license to Springer Nature Switzerland AG.
Funding
This study has been partially supported by the Dutch National Police. The authors would like to thank Elize Herrewijnen and Gizem Sogancioglu for their valuable feedback on earlier versions of this work.
| Funders |
|---|
| Dutch National Police |
Keywords
- Causal explanation
- Explainable Machine Learning (XML)
- Model-agnostic explanation
- Natural Language Processing (NLP)