SIDEs: Separating Idealization from Deceptive ‘Explanations’ in xAI

Proceedings of the 2024 Acm Conference on Fairness, Accountability, and Transparency (forthcoming)
  Copy   BIBTEX

Abstract

Explainable AI (xAI) methods are important for establishing trust in using black-box models. However, recent criticism has mounted against current xAI methods that they disagree, are necessarily false, and can be manipulated, which has started to undermine the deployment of black-box models. Rudin (2019) goes so far as to say that we should stop using black-box models altogether in high-stakes cases because xAI explanations ‘must be wrong’. However, strict fidelity to the truth is historically not a desideratum in science. Idealizations – the intentional distortions introduced to scientific theories and models – are commonplace in the natural sciences and are seen as a successful scientific tool. Thus, it is not falsehood qua falsehood that is the issue. In this paper, I outline the need for xAI research to engage in idealization evaluation. Drawing on the use of idealizations in the natural sciences and philosophy of science, I introduce a novel framework for evaluating whether xAI methods engage in successful idealizations or deceptive explanations (SIDEs). SIDEs evaluates whether the limitations of xAI methods, and the distortions that they introduce, can be part of a successful idealization or are indeed deceptive distortions as critics suggest. I discuss the role that existing research can play in idealization evaluation and where innovation is necessary. Through a qualitative analysis we find that leading feature importance methods and counterfactual explanations are subject to idealization failure and suggest remedies for ameliorating idealization failure.

Other Versions

No versions found

Links

PhilArchive

External links

  • This entry has no external links. Add one.
Setup an account with your affiliations in order to access resources via your University's proxy server

Through your library

Similar books and articles

Externalist XAI?Anders Søgaard - forthcoming - Theoria:e12581.
Causal Explanations and XAI.Sander Beckers - 2022 - Proceedings of the 1St Conference on Causal Learning and Reasoning, Pmlr.
The Structure of Idealization in Biological Theories: The Case of the Wright-Fisher Model.Donato Rodriguez Xavier & Arroyo-Santos Alfonso - 2012 - Journal for General Philosophy of Science / Zeitschrift für Allgemeine Wissenschaftstheorie 43 (1):11-27.
Defining Explanation and Explanatory Depth in XAI.Stefan Buijsman - 2022 - Minds and Machines 32 (3):563-584.

Analytics

Added to PP
2024-04-24

Downloads
319 (#81,620)

6 months
145 (#29,025)

Historical graph of downloads
How can I increase my downloads?

Author's Profile

Emily Sullivan
Utrecht University

Citations of this work

No citations found.

Add more citations

References found in this work

Depth: An Account of Scientific Explanation.Michael Strevens - 2008 - Cambridge: Harvard University Press.
Three Kinds of Idealization.Michael Weisberg - 2007 - Journal of Philosophy 104 (12):639-659.
Models Don’t Decompose That Way: A Holistic View of Idealized Models.Collin Rice - 2019 - British Journal for the Philosophy of Science 70 (1):179-208.

View all 11 references / Add more references