Attitudinal Tensions in the Joint Pursuit of Explainable and Trusted AI

Minds and Machines 33 (1):55-82 (2023)
  Copy   BIBTEX

Abstract

It is frequently demanded that AI-based Decision Support Tools (AI-DSTs) ought to be both explainable to, and trusted by, those who use them. The joint pursuit of these two principles is ordinarily believed to be uncontroversial. In fact, a common view is that AI systems should be made explainable so that they can be trusted, and in turn, accepted by decision-makers. However, the moral scope of these two principles extends far beyond this particular instrumental connection. This paper argues that if we were to account for the rich and diverse moral reasons that ground the call for explainable AI, and fully consider what it means to “trust” AI in a descriptively rich sense of the term, we would uncover a deep and persistent tension between the two principles. For explainable AI to usefully serve the pursuit of normatively desirable goals, decision-makers must carefully monitor and critically reflect on the content of an AI-DST’s explanation. This entails a deliberative attitude. Conversely, calls for trust in AI-DSTs imply the disposition to put questions about their reliability out of mind. This entails an unquestioning attitude. As such, the joint pursuit of explainable and trusted AI calls on decision-makers to simultaneously adopt incompatible attitudes towards their AI-DST, which leads to an intractable implementation gap. We analyze this gap and explore its broader implications: suggesting that we may need alternate theoretical conceptualizations of what explainability and trust entail, and/or alternate decision-making arrangements that separate the requirements for trust and deliberation to different parties.

Links

PhilArchive



    Upload a copy of this work     Papers currently archived: 91,571

External links

Setup an account with your affiliations in order to access resources via your University's proxy server

Through your library

Similar books and articles

Errata.[author unknown] - 1999 - Minds and Machines 9 (3):457-457.
Erratum.[author unknown] - 2004 - Minds and Machines 14 (2):279-279.
Editor’s Note.[author unknown] - 2003 - Minds and Machines 13 (3):337-337.
Book Reviews. [REVIEW][author unknown] - 1997 - Minds and Machines 7 (1):115-155.
Call for papers.[author unknown] - 1999 - Minds and Machines 9 (3):459-459.
Instructions for authors.[author unknown] - 1998 - Minds and Machines 8 (4):587-590.
Volume contents.[author unknown] - 1998 - Minds and Machines 8 (4):591-594.
Editor's Note.[author unknown] - 2001 - Minds and Machines 11 (1):1-1.
Book Reviews. [REVIEW][author unknown] - 1997 - Minds and Machines 7 (2):289-320.
Book Reviews. [REVIEW][author unknown] - 2004 - Minds and Machines 14 (2):241-278.
Erratum.[author unknown] - 1997 - Journal of Applied Non-Classical Logics 7 (3):473-473.
Correction to: What Might Machines Mean?Mitchell Green & Jan G. Michel - 2022 - Minds and Machines 32 (2):339-339.
Defining Explanation and Explanatory Depth in XAI.Stefan Buijsman - 2022 - Minds and Machines 32 (3):563-584.

Analytics

Added to PP
2023-02-28

Downloads
27 (#585,141)

6 months
11 (#231,656)

Historical graph of downloads
How can I increase my downloads?