Moral Responsibility for AI Systems

Advances in Neural Information Processing Systems 36 (Neurips 2023) (forthcoming)
  Copy   BIBTEX

Abstract

As more and more decisions that have a significant ethical dimension are being outsourced to AI systems, it is important to have a definition of moral responsibility that can be applied to AI systems. Moral responsibility for an outcome of an agent who performs some action is commonly taken to involve both a causal condition and an epistemic condition: the action should cause the outcome, and the agent should have been aware -- in some form or other -- of the possible moral consequences of their action. This paper presents a formal definition of both conditions within the framework of causal models. I compare my approach to the existing approaches of Braham and van Hees (BvH) and of Halpern and Kleiman-Weiner (HK). I then generalize my definition into a degree of responsibility.

Links

PhilArchive

External links

Setup an account with your affiliations in order to access resources via your University's proxy server

Through your library

Similar books and articles

Analytics

Added to PP
2022-12-05

Downloads
191 (#103,854)

6 months
47 (#90,662)

Historical graph of downloads
How can I increase my downloads?

Author's Profile

Sander Beckers
University of Amsterdam

Citations of this work

No citations found.

Add more citations

References found in this work

No references found.

Add more references