Explaining in Time: Meeting Interactive Standards of Explanation for Robotic Systems

ACM Transactions on Human-Robot Interaction 10 (3):1-23 (2021)
  Copy   BIBTEX

Abstract

Explainability has emerged as a critical AI research objective, but the breadth of proposed methods and application domains suggest that criteria for explanation vary greatly. In particular, what counts as a good explanation, and what kinds of explanation are computationally feasible, has become trickier in light of oqaque “black box” systems such as deep neural networks. Explanation in such cases has drifted from what many philosophers stipulated as having to involve deductive and causal principles to mere “interpretation,” which approximates what happened in the target system to varying degrees. However, such post hoc constructed rationalizations are highly problematic for social robots that operate interactively in spaces shared with humans. For in such social contexts, explanations of behavior, and, in particular, justifications for violations of expected behavior, should make reference to socially accepted principles and norms. In this article, we show how a social robot’s actions can face explanatory demands for how it came to act on its decision, what goals, tasks, or purposes its design had those actions pursue and what norms or social constraints the system recognizes in the course of its action. As a result, we argue that explanations for social robots will need to be accurate representations of the system’s operation along causal, purposive, and justificatory lines. These explanations will need to generate appropriate references to principles and norms—explanations based on mere “interpretability” will ultimately fail to connect the robot’s behaviors to its appropriate determinants. We then lay out the foundations for a cognitive robotic architecture for HRI, together with particular component algorithms, for generating explanations and engaging in justificatory dialogues with human interactants. Such explanations track the robot’s actual decision-making and behavior, which themselves are determined by normative principles the robot can describe and use for justifications.

Links

PhilArchive



    Upload a copy of this work     Papers currently archived: 91,122

External links

Setup an account with your affiliations in order to access resources via your University's proxy server

Through your library

Similar books and articles

Robot companions for children with down syndrome: A case study.Hagen Lehmann, Iolanda Iacono, Kerstin Dautenhahn, Patrizia Marti & Ben Robins - 2014 - Interaction Studiesinteraction Studies Social Behaviour and Communication in Biological and Artificial Systems 15 (1):99-112.
Explaining the Brain.Carl F. Craver - 2007 - Oxford, GB: Oxford University Press.
Philosophy, Kairosophy and the Lesson of Time.Marianna Papastephanou - 2014 - Educational Philosophy and Theory 46 (7):718-734.
Robotic Performance: An Ecology of Response.Zornitsa Dimitrova - 2017 - Performance Philosophy 3 (1):162-177.
Neurophysics of the Flow of Time.Ronald Gruber - 2008 - Journal of Mind and Behavior 29 (3):241-255.

Analytics

Added to PP
2021-11-04

Downloads
9 (#1,154,504)

6 months
1 (#1,346,405)

Historical graph of downloads
How can I increase my downloads?

Author's Profile

Thomas Arnold
Tufts University

Citations of this work

No citations found.

Add more citations

References found in this work

No references found.

Add more references