SIDEs: Separating Idealization from Deceptive 'Explanations' in xAI

  • Emily Sullivan*
  • *Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

Abstract

Explainable AI (xAI) methods are important for establishing trust in using black-box models. However, recent criticism has mounted against current xAI methods that they disagree, are necessarily false, and can be manipulated, which has started to undermine the deployment of black-box models. Rudin (2019) goes so far as to say that we should stop using black-box models altogether in high-stakes cases because xAI explanations 'must be wrong'. However, strict fidelity to the truth is historically not a desideratum in science. Idealizations-the intentional distortions introduced to scientific theories and models-are commonplace in the natural sciences and are seen as a successful scientific tool. Thus, it is not falsehood qua falsehood that is the issue. In this paper, I outline the need for xAI research to engage in idealization evaluation. Drawing on the use of idealizations in the natural sciences and philosophy of science, I introduce a novel framework for evaluating whether xAI methods engage in successful idealizations or deceptive explanations (SIDEs). SIDEs evaluates whether the limitations of xAI methods, and the distortions that they introduce, can be part of a successful idealization or are indeed deceptive distortions as critics suggest. I discuss the role that existing research can play in idealization evaluation and where innovation is necessary. Through a qualitative analysis we find that leading feature importance methods and counterfactual explanations are subject to idealization failure and suggest remedies for ameliorating idealization failure.

Original languageEnglish
Title of host publication2024 ACM Conference on Fairness, Accountability, and Transparency, FAccT 2024
PublisherAssociation for Computing Machinery
Pages1714-1724
Number of pages11
ISBN (Electronic)9798400704505
DOIs
Publication statusPublished - 3 Jun 2024
Event2024 ACM Conference on Fairness, Accountability, and Transparency, FAccT 2024 - Rio de Janeiro, Brazil
Duration: 3 Jun 20246 Jun 2024

Publication series

Name2024 ACM Conference on Fairness, Accountability, and Transparency, FAccT 2024

Conference

Conference2024 ACM Conference on Fairness, Accountability, and Transparency, FAccT 2024
Country/TerritoryBrazil
CityRio de Janeiro
Period3/06/246/06/24

Bibliographical note

Publisher Copyright:
© 2024 Owner/Author.

Keywords

  • explainable AI
  • idealization
  • philosophy of science
  • qualitative evaluation

Fingerprint

Dive into the research topics of 'SIDEs: Separating Idealization from Deceptive 'Explanations' in xAI'. Together they form a unique fingerprint.

Cite this