Abstract
The aim of this paper is to offer an analysis of the notion of artificial moral agent (AMA) and of its impact on human beings’ self-understanding as moral agents. Firstly, I introduce the topic by presenting what I call the Continuity Approach. Its main claim holds that AMAs and human moral agents exhibit no significant qualitative difference and, therefore, should be considered homogeneous entities. Secondly, I focus on the consequences this approach leads to. In order to do this I take into consideration the work of Bostrom and Dietrich, who have radically assumed this viewpoint and thoroughly explored its implications. Thirdly, I present an alternative approach to AMAs—the Discontinuity Approach—which underscores an essential difference between human moral agents and AMAs by tackling the matter from another angle. In this section I concentrate on the work of Johnson and Bryson and I highlight the link between their claims and Heidegger’s and Jonas’s suggestions concerning the relationship between human beings and technological products. In conclusion I argue that, although the Continuity Approach turns out to be a necessary postulate to the machine ethics project, the Discontinuity Approach highlights a relevant distinction between AMAs and human moral agents. On this account, the Discontinuity Approach generates a clearer understanding of what AMAs are, of how we should face the moral issues they pose, and, finally, of the difference that separates machine ethics from moral philosophy.
Similar content being viewed by others
Notes
This is why the authors spend many words explaining the level of abstraction (LoA) methodology, which allows them to achieve a particular outcome—the inclusion of AMAs in the category of moral agents—without at the same time reducing all moral agents to the basically technological model they resort to. In fact, it is possible to acknowledge the status of moral agents to machines or computer software only if a particular LoA is assumed. That is, the assumption of a specific LoA sets the conditions by which a particular claim makes sense. If these conditions change, then the same claim must be reassessed at a different LoA. Although highly controversial, the recourse to similar methodological precautions is in my opinion of utmost importance in the field of machine ethics, since it helps avoiding epistemological trespasses which lead to confusion and conceptual fuzziness.
In this paper, by “machine ethics” I mean the technological discipline the purpose of which is to design, build, and develop robots exhibiting some sort of moral behaviour—not the philosophical discussion about how such technological undertaking impinges on our traditional moral ideas. This helps to keep the technological perspective separate from the philosophical one, which is a very important distinction not to be forgotten. For similar concerns regarding the expression “machine ethics” see Anderson (2011), who distinguishes between machine ethics and machine metaethics; and Torrance (2011), who distinguishes between practical and philosophical machine ethics.
In distinguishing between an operative approach and, as it were, a “philosophical” or “definitional” approach to moral agency I follow Gunkel 2012, p. 74. Gunkel stresses the difference between, on the one hand, Floridi’s and Sanders’s “functional approach” to moral agency, which implies “engineering solutions” or “‘effective characterization(s)’ that can work” and, on the other hand, the discussion of the “fundamental philosophical problem” of moral agency, which entails “advancing and defending a decision concerning a definition”. Although the issue lurks at the back of this distinction, I do not intend to take any stance on the epistemological opposition between realism and constructivism. The only point I wish to highlight here is that operative approaches organize knowledge in accordance with a specific purpose they assume in advance and intend to achieve—in the case of machine ethics, the technological reproduction of human moral behaviour. On the contrary, “philosophical” approaches like those I have in mind here strive to understand things independently from any productive purpose—i.e., absolutely—although of course this may very well lead to practically relevant conclusions. This argument is based in the well-known Aristotelian distinction between theoria and poiesis.
Similarly, Hall claims that future “superethical machines” (Hall 2011a, p. 42) will render society a better place for human beings too, which makes the effort towards the manufacturing of conscious machines a moral duty.
To some extent, a similar role can be acknowledged to science fiction too. Sure enough, philosophical discussions and fictional writings are very different forms of human expression, which aim at different targets through different means. However, a very interesting intercourse between the two has always been occurring, since literary imagination can offer radical insight into the implication of concepts, which may have already found their way to common sense. In the field of machine ethics the most interesting example is that of Isaac Asimov’s well-known collection of stories I, Robot, where the author proposed his Three (or One plus Three) Laws of Robotics—a set of laws intended to serve as a moral guideline for future technological agents. The laws have been seriously taken into account since the problem of implementing moral constraints in machines passed from fictional imagination to engineering practice (Clarke 2011).
In philosophy of technology the instrumental approach is often criticized on the account that it supposedly supports the thesis that technological tools are neutral and transparent entities, which are simply used and by no means feedback on the users’ practices and relations to the world (Verbeek 2005; Kiran and Verbeek 2010). I cannot go deep into this now, but it must be said that neither Heidegger nor Jonas endorsed such a claim. In fact, they both submitted an opposite interpretation: tools, as simple or as advanced as they may be, do transform practices and contribute to shaping the human experience of the world by both modifying how ends are achieved and making new ends achievable, as Verbeek (2005) correctly points out. Even so, they remain tools, i.e., objects we primarily resort to in order to achieve our ends. So, in my opinion, it is not necessary to dismiss the instrumental approach to technology altogether. What is necessary, however, is to avoid any “externalist” oversimplification in defining what tools are and how they impact on human existence.
It is to be briefly noticed that, although learning machines may be thought of as self-determining entities, they still are built in order to serve some purpose set by human beings, as all machines are. So, their form of “self-determination”, being still of a functional kind, does not match the way human beings set purposes and values. In fact, machine “self-determination” is always embedded in practical contexts the general purposes and values of which are already set by human beings.
This tendency, however, is of course enforced by human-likeness, so that to a certain degree it is possible to exploit it in order to promote the social acceptability of artificial agents (Mori 1970; Duffy 2003, 2013; Fink 2012). Moreover, if machines are to be deployed in human contexts, they must be able to interact with objects which have been designed for human use, like stairs, doorknobs, or trays, and with us. So, machines inevitably feature cues which trigger our inclination to frame them in human terms.
References
Allen, C., Varner, G., & Zinser, J. (2000). Prolegomena to any future artificial moral agent. Journal of Experimental and Theoretical Artificial Intelligent, 12, 251–261.
Anderson, S. L. (2011). Machine metaethics. In M. Anderson & S. L. Anderson (Eds.), Machine ethics (pp. 21–27). Cambridge: Cambridge University Press.
Beavers, A. F. (2012). Moral machines and the threat of ethical nihilism. In P. Lin, K. Abney & G. A. Bekey (Eds.), Robot ethics. The ethical and social implications of robotics (pp. 333–344). Cambridge: The M.I.T. Press.
Bostrom, N. (2003). Ethical issues in advanced artificial intelligence. https://nickbostrom.com/ethics/ai.html. Accessed 22 Aug 2017.
Bostrom, N. (2014). Superintelligence. Paths, dangers, strategies. Oxford: Oxford University Press.
Bryson, J. J. (2010). Robots Should Be Slaves. In Y. Wilks (Ed.), Close engagements with artificial companions: Key social, psychological, ethical and design issues (pp. 63–74). Amsterdam: John Benjamins.
Bryson, J. J., & Kime, P. (2011). Just an artifact: Why machines are perceived as moral agents. https://www.cs.bath.ac.uk/~jjb/ftp/BrysonKime-IJCAI11.pdf. Accessed 22 Aug 2017.
Clarke, R. (2011). Asimov’s laws of robotics. Implications for information technology. In M. Anderson & S. L. Anderson (Eds.), Machine ethics (pp. 254–284). Cambridge: Cambridge University Press.
Dennett, D. C. (1997). When HAL kills, who’s to blame? Computer ethics. In D. G. Stork (Ed.), Hal’s legacy: 2001’s computer as dream and reality (pp. 351–366). Cambridge: The M.I.T. Press.
Dietrich, E. (2007). After humans are gone. Journal of Experimental and Theoretical Artificial Intelligence, 19(1), 55–67.
Dietrich, E. (2011). Homo Sapiens 2.0. Building the better robots of our nature. In M. Anderson & S. L. Anderson (Eds.), Machine ethics (pp. 531–538). Cambridge: Cambridge University Press.
Duffy, B. (2003). Anthropomorphism and the social robot. Robotic and Autonomous Systems, 42, 177–190.
Duffy, B. (2013). Anthropomorphism and robotics. http://medialabeurope.org/anthropos/publications/pubsIAISB02-Duffy.pdf. Accessed 28 Nov 2017.
Fabris, A. (2016). Philosophy, image and the mirror of machines. In Ž. Paić & K. Purgar (Eds.), Theorizing images (pp. 111–120). Newcastle upon Tyne: Cambridge Scholars.
Fink, J. (2012). Anthropomorphism and human likeness in the design of robots and human-robot interaction. In S. S. Ge et al. (Eds.), ICSR 2012, LNAI 7621, pp. 199–208.
Floridi, L., & Sanders, J. W. (2004). On the morality of artificial agents. Minds and Machine, 14, 349–379.
Franklin, S., & Graesser, A. (1996). Is it an agent, or just a program? A taxonomy for autonomous agents. In J. P. Müller, M. J. Wooldridge & N. R. Jennings (Eds.), Intelligent Agents III. Agent Theories, Architectures, and Languages. ATAL 1996. Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence), vol. 1193 (pp. 22–35). Berlin: Springer.
Friedman, B., & Kahn, P. H. (1992). Human agency and responsible computing: Implications for computer system design. Journal of Systems Software, 17(7), 7–14.
Fussel, S. R., Kiesler, S., Setlock, L. D., & Yew, V. (2008). How people anthropomorphize robots. In HRI’08 Proceedings of the 3rd ACM/IEEE International Conference on Human Robot Interaction (pp 145–152).
Gips, J. (1995). Towards the ethical robot. In G. K. Ford, C. Glymour & P. J. Hayes (Eds.), Android epistemology (pp. 243–252). Cambridge: The M.I.T. Press.
Grodzinsky, F. S., Miller, K. W., & Wolf, M. J. (2008). The ethics of designing artificial agents. Ethics and Information Technology, 10, 115–121.
Gunkel, D. J. (2012). The machine question. Critical perspectives on AI, robots and ethics. Cambridge: The M.I.T. Press.
Hall, J. S. (2011a). Ethics for machines. In M. Anderson & S. L. Anderson (Eds.), Machine ethics (pp. 28–44). Cambridge: Cambridge University Press.
Hall, J. S. (2011b). Ethics for self-improving machines. In M. Anderson & S. L. Anderson (Eds.), Machine ethics (pp. 512–523). Cambridge: Cambridge University Press.
Heidegger, M. (2010). Being and time. New York: State University of New York Press.
Heidegger, M. (2013). The question concerning technology and other essays. New York: Harper Perennial.
Henry, B. (2014). Imaginaries of the Global Age. “Golem and others” in the post-human condition. Politica e Società, 2/2014, 221–246.
Himma, K. E. (2009). Artificial agency, consciousness, and the criteria for moral agency: What properties must an artificial agent have to be a moral agent? Ethics and Information Technology, 11(1), 19–29.
Johnson, D. G. (2003). Computer ethics. In R. G. Frey & C. H. Wellman (Eds.), A companion to applied ethics (pp. 608–619). Malden-Oxford-Carlton: Blackwell.
Johnson, D. G. (2011). Computer systems. Moral entities, but not moral agents. In M. Anderson & S. L. Anderson (Eds.), Machine ethics (pp. 168–183). Cambridge: Cambridge University Press.
Jonas, H. (1953). Cybernetics and purpose: A critique. Social research, XX(2), pp. 172–192. Reprinted as § 5 in Id. (2001). The Phenomenon of Life. Toward a Philosophical Biology (pp. 108–127). Evanston: Northwestern University Press.
Jonas, H. (1959). The practical uses of theory. Social research, XXVI(2), pp. 151–166. Reprinted as § 8 in Id. (2001). The Phenomenon of Life. Toward a Philosophical Biology (pp. 188–210). Evanston: Northwestern University Press.
Kakoudaki, D. (2014). Anatomy of a robot. Literature, cinema, and the cultural work of artificial people. New Brunswick: Rutgers University Press.
Kiran, A. E., & Verbeek, P.-P. (2010). Trusting our selves to technology. Knowledge, Technology, and Policy, 23, 409–427.
Kurzweil, R. (2005). The singularity is near. When Humans transcend biology. New York: Viking.
Laukyte, M. (2017). Artificial agents among us. Should we recognize them as agents proper? Ethics and Information Technology, 19(1), 1–17.
Lemaignan, S., Fink, J., & Dillenbourg, P. (2014). The Dynamics of Anthropomorphism in Robotics. In HRI’14 Proceedings of the 2014 ACM/IEEE International Conference on Human-Robot Interaction (pp. 226–227).
McDermott, D. (2008). What matters to a machine? In M. Anderson & S. L. Anderson (Eds.), Machine ethics (pp. 88–114). Cambridge: Cambridge University Press.
Moor, J. H. (1995). Is ethics computable? Metaphilosophy, 26(1–2), 1–21.
Moor, J. H. (2006). The nature, importance, and difficulty of machine ethics. IEEE Intelligent Systems, 21(4), 18–21.
Moore, G. E. (1965). Cramming more components into integrated circuits. Electronics, 38(8), 114–117.
Mori, M. (1970). Bukimi no tani. Energy, 7, 33–35. English version: The Uncanny Valley. IEEE Robotics and Automation Magazine, June 2012, 98–100.
Nass, C., & Moon, Y. (2000). Machines and mindlessness: Social responses to computers. Journal of Social Issues, 56(1), 81–103.
Nissenbaum, H. (2001). How computer systems embody values. Computer, 34, 118–120.
Scheutz, M. (2012). The inherent dangers of unidirectional emotional bonds between humans and social robots. In P. Lin, K. Abney & G. A. Bekey (Eds.), Robot ethics. The ethical and social implications of robotics (pp. 205–222). Cambridge: The MIT Press.
Searle, J. R. (1980). Minds, brains, and programs. The Behavioral and Brain Sciences, 3, 417–424.
Sullins, J. P. (2011). When is a robot a moral agent? In M. Anderson & S. L. Anderson (Eds.), Machine ethics (pp. 151–161). Cambridge: Cambridge University Press.
Torrance, S. (2011). Machine ethics and the Idea of a more-than-human moral world. In M. Anderson & S. L. Anderson (Eds.), Machine ethics (pp. 115–137). Cambridge: Cambridge University Press.
Turing, A. M. (1950). Computing machinery and intelligence. Mind, LIX(236), 433–460.
Turkle, S. (2011). Authenticity in the age of digital companions. In M. Anderson & S. L. Anderson (Eds.), Machine ethics (pp. 62–76). Cambridge: Cambridge University Press.
Verbeek, P.-P. (2005). What Things Do. Philosophical Reflections on Technology, Agency, and Design. University Park: The Pennsylvania State University Press.
Vinge, V. (1993). The coming technological singularity: How to survive in the post-human era. Vision-21: Interdisciplinary Science and Engineering in the Era of Cyberspace (pp. 11–22). NASA Scientific and Technical Information Program.
Wallach, W. (2010). Robot minds and human ethics: the need for a comprehensive model of decision making. Ethics and Information Technology, 12(3), 243–250.
Wallach, W., & Allen, C. (2009). Moral machines. Teaching robots right from wrong. New York: Oxford University Press.
Wallach, W., Allen, C., & Smit, I. (2011). Why machine ethics? In M. Anderson & S. L. Anderson (Eds.), Machine ethics (pp. 51–61). Cambridge: Cambridge University Press.
Whitby, B. (2011). On computable morality: An examination of machines as moral advisors. In M. Anderson & S. L. Anderson (Eds.), Machine ethics (pp. 138–150). Cambridge: Cambridge University Press.
Yudkowsky, E. (2008). Artificial intelligence as a positive and negative factor in global risk. Machine Intelligence Research Institute. http://intelligence.org/files/AIPosNegFactor.pdf. Accessed online 22 Aug 2017.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Fossa, F. Artificial moral agents: moral mentors or sensible tools?. Ethics Inf Technol 20, 115–126 (2018). https://doi.org/10.1007/s10676-018-9451-y
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10676-018-9451-y