Abstract
The emergence of AI is posing serious challenges to standard conceptions of moral status. New non-biological entities are able to act and make decisions rationally. The question arises, in this regard, as to whether AI systems possess or can possess the necessary properties to be morally considerable. In this chapter, we have undertaken a systematic analysis of the various debates that are taking place about the moral status of AI. First, we have discussed the possibility that AI systems, by virtue of its new agential capabilities, can be understood as a moral agent. Discussions between those defending mentalist and anti-mentalist positions have revealed many nuances and particularly relevant theoretical aspects. Second, given that an AI system can hardly be an entity qualified to be responsible, we have delved into the responsibility gap and the different ways of understanding and addressing it. Third, we have provided an overview of the current and potential patientist capabilities that AI systems possess. This has led us to analyze the possibilities of AI possessing moral patiency. In addition, we have addressed the question of the moral and legal rights of AI. Finally, we have introduced the two most relevant authors of the relational turn on the moral status of AI, Mark Coeckelbergh and David Gunkel, who have been led to defend a relational approach to moral life as a result of the problems associated with the ontological understanding of moral status.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
- 2.
Animals and ecosystems had previously been morally considered, but instrumentally so. Examples are the doctrine of indirect duties (Kant 2017), which understands that we have duties towards animals because of the effects that cruel treatment of them could have towards the consideration of other human beings; and the conservationist doctrine that defended the preservation of natural spaces on account of the aesthetic pleasure that they produced in human beings (Callicott 1990).
- 3.
Some have argued that AI may one day become a superagent capable of greater capacity for moral agency than humans (Bostrom 2017).
- 4.
By standard conceptions of moral status, we refer to the most dominant accounts that have been developed to answer the question about the possible inclusion of different entities in the circle of moral consideration before the emergence of AI. Christian ethics, Kantianism or utilitarianism give different answers to the inquiry about the criteria for determining moral status. However, they all agree that, because artifacts lack such properties relevant to moral status as sentience or rationality, they cannot be morally considered in themselves.
- 5.
Utilitarianism defends an asymmetric conception of moral status, but only on the side of patiency. That is, it is not necessary to be both a moral agent and patient to have moral status, but it is sufficient to be solely the latter. This means that artifacts cannot have moral status either under this viewpoint.
- 6.
When we refer to a conscious mind, we are understanding it as a biological mind. Even though some proposals in the philosophy of mind envision a full mind without biological anchorage (Fodor 2000), the lack of agreement leads us to endorse this minimum definition.
- 7.
This is not to say that moral personhood is coincident with the human species. Human beings can have full moral status if they are moral persons, that is, if they have the properties required to have full moral status, such as conscience, rationality, etc. This should be kept in mind since being human is often confused with being a moral person. Thanks to an anonymous reviewer for pointing this out.
- 8.
The importance of the level of abstraction can be argued using a thought experiment developed by Nicholas Agar (2020). If you found out after years of marriage that your wife is a robot who has no mind, even though she always behaved as if she had one, would you stop considering her morally? Intuitive rejections of this question place serious limits on markedly internalist approaches.
- 9.
- 10.
This perspective can be denied from two different angles. On the one hand, although AI currently lacks the properties necessary to be a responsible moral agent, it is possible that in the future, due to technological progress, it may possess them. On the other hand, as we will see later in those who have advocated for a profound change in the concept of responsibility, AI can be responsible in the wake of totally new relations of responsibility.
- 11.
We distinguish between instrument and machine because of the argument offered by Gunkel (2020). Gunkel argues that the main reason why the responsibility gap occurs is that we try to respond to the advent and importance of AI from the instrumentalist paradigm. However, AI can be understood not only as an instrument, but also as a machine. However, this has the problem that a machine, although distinct from an instrument in the independence of its behavior, is not widely different in terms of autonomy.
- 12.
Strict liability is defined as an offense that, although a behavior punishable by the transgression of a norm, does not imply blameworthiness. By this we refer to the legal conditions on which the assumptions of strict liability are based. Kiener (2022) refers in particular to three: i) the offenses must be “not very serious”; ii) the ones responsible are usually those who benefit most from the damage produced; iii) these offenses do not carry any sufficiently serious stigma.
- 13.
- 14.
The interests of animals and ecosystems are not based on the same assumptions. The defense of animal rights presupposes the possibility of being able to subjectively experience a certain degree of well-being. In contrast, ecosystem interests are holistically understood as needs for a certain natural environment to maintain its ecological and/or biotic balance (Callicott 1980).
- 15.
By general level we refer to a specific part of Miller’s argument. One of the criticisms that can be made of Miller is that human beings can also be created for some purpose and that this does not make them lose their rights. However, Miller argues that being or not being produced according to a purpose should not be understood at the individual level but at the level of the existential origin of a species or typology of artifacts. The human species is the result of natural selection, a blind process devoid of teleology, quite the opposite of AI, a product of human purpose.
- 16.
We use the term quasi-ontological to express the possibility of partially relational approaches. There are patientist positions that emphasize relations; however, these relations are marked by certain ontological properties such as moral agency (in the case of virtue ethics, see Cappuccio et al. 2020).
References
Agar, N. 2020. How to treat machines that might have minds. Philosophy & Technology 33 (2): 269–282. https://doi.org/10.1007/s13347-019-00357-8.
Anderson, S.L. 2008. Asimov’s “three laws of robotics” and machine metaethics. AI & SOCIETY 22 (4): 477–493. https://doi.org/10.1007/s00146-007-0094-5.
———. 2011. Philosophical concerns with machine ethics. In Machine ethics, ed. M. Anderson and S.L. Anderson, 162–167. Cambridge: Cambridge University Press.
Anderson, M., S.L. Anderson, J.H. Moor, J. Stirrs, C. Allen, W. Wallach, I. Smit, et al. 2011. In Machine ethics, ed. M. Anderson and S.L. Anderson. Cambridge: Cambridge University Press. https://doi.org/10.1017/cbo9780511978036.
Andreotta, A.J. 2021. The hard problem of AI rights. AI & SOCIETY 36 (1): 19–32. https://doi.org/10.1007/s00146-020-00997-x.
Asaro, P.M. 2006. What should we want from a robot ethic? International Review of Information Ethics 6 (12): 9–16. https://doi.org/10.29173/irie134.
Asaro, P. 2012. On banning autonomous weapon systems: Human rights, automation, and the dehumanization of lethal decision-making. International review of the Red Cross 94 (886): 687–709.
Ashrafian, H. 2015. Artificial intelligence and robot responsibilities: Innovating beyond rights. Science and Engineering Ethics 21 (2): 317–326. https://doi.org/10.1007/s11948-014-9541-0.
Behdadi, D., and C. Munthe. 2020. A normative approach to artificial moral agency. Minds and Machines 30 (2): 195–218. https://doi.org/10.1007/s11023-020-09525-8.
Bostrom, N. 2003. Ethical issues in advanced artificial intelligence. In Science fiction and philosophy: From time travel to superintelligence, 277–284.
———. 2017. Superintelligence. Paris: Dunod.
Bostrom, N., and E. Yudkowsky. 2018. The ethics of artificial intelligence. In Artificial intelligence safety and security, ed. R.V. Yampolskiy, 57–69. London: Routledge.
Bringsjord, S. 1992. What robots can and can’t be. New York: Kluwer Academic.
———. 2007. Ethical robots: The future can heed us. AI & SOCIETY 22 (4): 539–550.
Bryson, J.J., and P.P. Kime. 2011. Just an artifact: Why machines are perceived as moral agents. In Twenty-second international joint conference on artificial intelligence, vol. 22, 1641.
Callicott, J.B. 1980. Animal liberation: A triangular affair. Environmental ethics 2 (4): 311–338.
———. 1990. Whither conservation ethics? Conservation Biology 4 (1): 15–20.
Cappuccio, M.L., A. Peeters, and W. McDonald. 2020. Sympathy for Dolores: Moral consideration for robots based on virtue and recognition. Philosophy & Technology 33 (1): 9–31. https://doi.org/10.1007/s13347-019-0341-y.
Champagne, M., and R. Tonkens. 2015. Bridging the responsibility gap in automated warfare. Philosophy & Technology 28 (1): 125–137. https://doi.org/10.1007/s13347-013-0138-3.
Chesterman, S. 2020. Artificial intelligence and the limits of legal personality. International & Comparative Law Quarterly 69 (4): 819–844. https://doi.org/10.1017/s0020589320000366.
Chomanski, B. 2020. Should moral machines be banned? A commentary on van Wynsberghe and Robbins ‘critiquing the reasons for making artificial moral agents’. Science and Engineering Ethics 26 (6): 3469–3481. https://doi.org/10.1007/s11948-020-00255-9.
Coeckelbergh, M. 2009. Virtual moral agency, virtual moral responsibility: On the moral significance of appearance, perception and performance of artificial agents. AI & SOCIETY 24 (2): 181–189. https://doi.org/10.1007/s00146-009-0208-3.
———. 2010. Moral appearances: Emotions, robots, and human morality. Ethics and Information Technology 12 (3): 235–241. https://doi.org/10.1007/s10676-010-9221-y.
———. 2012. Growing moral relations: Critique of moral status ascription. New York: Palgrave Macmillan.
———. 2014. The moral standing of machines: Towards a relational and non-Cartesian moral hermeneutics. Philosophy & Technology 27 (1): 61–77. https://doi.org/10.1007/s13347-013-0133-8.
———. 2020. Artificial intelligence, responsibility attribution, and a relational justification of explainability. Science and Engineering Ethics 26 (4): 2051–2068. https://doi.org/10.1007/s11948-019-00146-8.
Danaher, J. 2016. Robots, law and the retribution gap. Ethics and Information Technology 18 (4): 299–309. https://doi.org/10.1007/s10676-016-9403-3.
———. 2020. Welcoming robots into the moral circle: A defense of ethical behaviourism. Science and Engineering Ethics 26 (4): 2023–2049. https://doi.org/10.1007/s11948-019-00119-x.
De Jong, R. 2020. The retribution-gap and responsibility-loci related to robots and automated technologies: A reply to Nyholm. Science and Engineering Ethics 26 (2): 727–735. https://doi.org/10.1007/s11948-019-00120-4.
DeGrazia, D. 2008. Moral status as a matter of degree? The Southern Journal of Philosophy 46 (2): 181–198. https://doi.org/10.1111/j.2041-6962.2008.tb00075.x.
Dennett, D. 1997. When HAL kills, who’s to blame? Computer ethics. In HAL’s legacy: 2001’s computer as dream and reality, ed. D. Stork, 351–366. Cambridge, Massachusetts: MIT Press.
Dietrich, E. 2001. Homo sapiens 2.0: Why we should build the better robots of our nature. Journal of Experimental & Theoretical Artificial Intelligence: JETAI 13 (4): 323–328. https://doi.org/10.1080/09528130110100289.
Donath, J. 2020. Ethical issues in our relationship with artificial entities. In The Oxford handbook of ethics of AI, ed. M.D. Dubber, F. Pasquale, and S. Das, 53–73. Oxford: Oxford University Press.
Etzioni, A., and O. Etzioni. 2016. AI assisted ethics. Ethics and Information Technology 18 (2): 149–156. https://doi.org/10.1007/s10676-016-9400-6.
Floridi, L. 2010. Artificial companions and their philosophical challenges. In Close engagements with artificial companions, ed. J. Wilks, 23–28. Amsterdam: John Benjamins Publishing Company.
———. 2016. Faultless responsibility: On the nature and allocation of moral responsibility for distributed moral actions. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences 374 (2083): 20160112. https://doi.org/10.1098/rsta.2016.0112.
Floridi, L., and J.W. Sanders. 2004. On the morality of artificial agents. Minds and Machines 14 (3): 349–379.
Fodor, J. 2000. The mind doesn’t work that way: The scope and limits of computational psychology. Cambridge, Massachusetts: MIT Press.
Formosa, P., and M. Ryan. 2021. Making moral machines: Why we need artificial moral agents. AI & SOCIETY 36 (3): 839–851. https://doi.org/10.1007/s00146-020-01089-6.
Fossa, F. 2018. Artificial moral agents: Moral mentors or sensible tools? Ethics and Information Technology 20 (2): 115–126. https://doi.org/10.1007/s10676-018-9451-y.
Frankfurt, H.G. 1988. Freedom of the will and the concept of a person. In What is a person, ed. M.F. Goodman, 127–144. Totowa: Humana Press.
Friedman, B., and P.H. Kahn. 1992. Human agency and responsible computing: Implications for computer system design. Journal of Systems Software 17 (7): 7–14.
Gellers, J.C. 2020. Rights for robots: Artificial intelligence, animal and environmental law. London: Routledge.
Gerdes, A., and P. Øhrstrøm. 2015. Issues in robot ethics seen through the lens of a moral Turing test. Journal of Information, Communication and Ethics in Society 13 (2): 98–109. https://doi.org/10.1108/jices-09-2014-0038.
Gibert, M., and D. Martin. 2022. In search of the moral status of AI: Why sentience is a strong argument. AI & SOCIETY 37 (1): 319–330. https://doi.org/10.1007/s00146-021-01179-z.
Goldie, P. 2006. Anti-empathy: Against empathy as perspective-shifting. In Empathy: Philosophical and psychological perspectives, ed. P. Goldie and A. Coplan, 302–317. Oxford: Oxford University Press.
Gordon, J.S. 2020. What do we owe to intelligent robots? In Smart technologies and fundamental rights, ed. J.S. Gordon, 17–47. Leiden: Brill.
Gordon, J.S.D., and J. Gunkel. 2022. Moral status and intelligent robots. The Southern Journal of Philosophy 60 (1): 88–117. https://doi.org/10.1111/sjp.12450.
Gunkel, D. 2012. The machine question: Critical perspectives on AI, robots, and ethics. Cambridge, Massachusetts: MIT Press.
———. 2018. Robot rights. Cambridge, Massachusetts: MIT Press.
Gunkel, D.J. 2014. A vindication of the rights of machines. Philosophy & Technology 27 (1): 113–132. https://doi.org/10.1007/s13347-013-0121-z.
———. 2020. Mind the gap: Responsible robotics and the problem of responsibility. Ethics and Information Technology 22 (4): 307–320. https://doi.org/10.1007/s10676-017-9428-2.
Hall, J.S. 2011. Ethics for machines. In Machine ethics, ed. M. Anderson and S.L. Anderson, 28–44. Cambridge: Cambridge University Press.
Hanson, F.A. 2009. Beyond the skin bag: On the moral responsibility of extended agencies. Ethics and Information Technology 11 (1): 91–99. https://doi.org/10.1007/s10676-009-9184-z.
Harris, J., and J.R. Anthis. 2021. The moral consideration of artificial entities: A literature review. Science and Engineering Ethics 27 (4): 1–95. https://doi.org/10.1007/s11948-021-00331-8.
Heider, F., and M. Simmel. 1944. An experimental study of apparent behavior. The American Journal of Psychology 57 (2): 243–259.
Himma, K. 2009. Artificial agency, consciousness, and the criteria for moral agency: What properties must an artificial agent have to be a moral agent? Ethics and Information Technology 11 (1): 19–29. https://doi.org/10.1007/s10676-008-9167-5.
Hogan, K. 2017. Is the machine question the same question as the animal question? Ethics and Information Technology 19 (1): 29–38. https://doi.org/10.1007/s10676-017-9418-4.
Hohfeld, W.N. 1919. Fundamental legal conceptions as applied in judicial reasoning. New Haven: Yale University Press.
Hursthouse, R. 2013. Moral status. In International encyclopedia of ethics, ed. H. LaFollette. New York: John Wiley & Sons. https://onlinelibrary.wiley.com/doi/10.1002/9781444367072.wbiee076.
Ihde, D. 1990. Technology and the lifeworld. Bloomington: Indiana University Press.
Inayatullah, S. 2001. The rights of robot: Inclusion, courts and unexpected futures. Journal of Future Studies 6 (2): 93–102.
Jaworska, A., and J. Tannenbaum. 2013. The grounds of moral status. In The Stanford encyclopedia of philosophy, ed. E.N. Zalta. https://plato.stanford.edu/entries/grounds-moral-status/.
Johnson, D.G. 2006. Computer systems: Moral entities but not moral agents. Ethics and Information Technology 8 (4): 195–204. https://doi.org/10.1007/s10676-006-9111-5.
Johnson, D.G., and K.W. Miller. 2008. Un-making artificial moral agents. Ethics and Information Technology 10 (2): 123–133. https://doi.org/10.1007/s10676-008-9174-6.
Johnson, D.G., and T.M. Powers. 2006. Computer systems and responsibility: A normative look at technological complexity. Ethics and Information Technology 7 (2): 99–107. https://doi.org/10.1007/s10676-005-4585-0.
Johnson, D.G., and M. Verdicchio. 2018. Why robots should not be treated like animals. Ethics and Information Technology 20 (4): 291–301. https://doi.org/10.1007/s10676-018-9481-5.
Kant, I. 2017. Kant: The metaphysics of morals. Cambridge: Cambridge University Press.
Kiener, M. 2022. Can we bridge AI’s responsibility gap at will? Ethical Theory and Moral Practice 25 (4): 575–593. https://doi.org/10.1007/s10677-022-10313-9.
Kolodny, N., and B. John. 2016. Instrumental rationality. In The Stanford encyclopedia of philosophy, ed. E.N. Zalta. http://plato.stanford.edu/archives/spr2016/entries/rationality-instrumental/.
Koops, B.J., M. Hildebrandt, and D.O. Jaquet-Chiffelle. 2010. Bridging the accountability gap: Rights for new entities in the information society. Minnesota Journal of Law, Science & Technology 11 (2): 497. https://scholarship.law.umn.edu/mjlst/vol11/iss2/4/.
LaChat, M.R. 1986. Artificial intelligence and ethics: An exercise in the moral imagination. AI Magazine 7 (2): 70–79.
Latour, B. 1999. Pandora’s hope: Essays on the reality of science studies. Cambridge: Harvard University Press.
Leopold, A. 2020. A Sand County almanac: And sketches here and there. Oxford: Oxford University Press.
Levy, D. 2009. The ethical treatment of artificially conscious robots. International Journal of Social Robotics 1 (3): 209–216. https://doi.org/10.1007/s12369-009-0022-6.
Llorca Albareda, J. 2022. Agencia (y) moral en la era de la inteligencia artificial. In Filosofía, tecnopolítica y otras ciencias sociales: nuevas formas de revisión y análisis del humanismo, ed. M. Bermúdez and A. Sánchez Cotta, 127–147. Madrid: Dykinson.
———. 2023. El estatus moral de las entidades de inteligencia artificial. Disputatio. Philosophical Research Bulletin 12 (24): 241–249.
Llorca-Albareda, J., and G. Díaz-Cobacho. 2023. Contesting the consciousness criterion: A more radical approach to the moral status of non-humans. AJOB Neuroscience 14 (2): 158–160. https://doi.org/10.1080/21507740.2023.2188280.
Loh, J. 2019. Responsibility and robot ethics: A critical overview. Philosophies 4 (4): 58. https://doi.org/10.3390/philosophies4040058.
Macnamara, C. 2015. Blame, communication, and morally responsible agency. In The nature of moral responsibility: New essays, ed. R. Clarke, M. McKenna, and A.M. Smith, 211–236. Oxford: Oxford University Press.
Maibom, H. 2009. Feeling for others: Empathy, sympathy, and morality. Inquiry 52: 483–499.
Marino, D., and G. Tamburrini. 2020. Learning robots and human responsibility. In Machine ethics and robot ethics, ed. W. Wallach and P. Asaro, 377–382. London: Routledge.
Matheson, B. 2012. Manipulation, moral responsibility, and machines. In The machine question: AI, ethics and moral responsibility, ed. D. Gunkel, J. Bryson, and S. Torrance, 25–29. The Society for the Study of Artificial Intelligence and Simulation of Behaviour.
Matthias, A. 2004. The responsibility gap: Ascribing responsibility for the actions of learning automata. Ethics and Information Technology 6 (3): 175–183. https://doi.org/10.1007/s10676-004-3422-1.
McKenna, M.A.C., and D. Justin. 2015. Compatibilism. In The Stanford encyclopedia of philosophy, ed. E.N. Zalta. http://plato.stanford.edu/archives/sum2015/entries/compatibilism/.
McNally, P., and S. Inayatullah. 1988. The rights of robots: Technology, culture and law in the 21st century. Futures 20 (2): 119–136.
Miller, L.F. 2015. Granting automata human rights: Challenge to a basis of full-rights privilege. Human Rights Review 16 (4): 369–391. https://doi.org/10.1007/s12142-015-0387-x.
Moor, J.H. 1985. What is computer ethics? Metaphilosophy 16 (4): 266–275.
Mosakas, K. 2021. On the moral status of social robots: Considering the consciousness criterion. AI & SOCIETY 36 (2): 429–443. https://doi.org/10.1007/s00146-020-01002-1.
Müller, V.C. 2021. Is it time for robot rights? Moral status in artificial entities. Ethics and Information Technology 23 (4): 579–587. https://doi.org/10.1007/s10676-021-09596-w.
Nadeau, J.E. 2006. Only androids can be ethical. In Thinking about android epistemology, ed. K.M. Ford, C. Glymour, and P. Hayes, 241–248. Palo Alto: AAAI Press.
Nagel, T. 1974. What is it like to be a bat? The Philosophical Review 83 (4): 435–450.
Nagenborg, M. 2007. Artificial moral agents: An intercultural perspective. International Review of Information Ethics 7 (9): 129–133. https://doi.org/10.29173/irie14.
Neely, E. 2014. Machines and the moral community. Philosophy & Technology 27 (1): 97–111. https://doi.org/10.1007/s13347-013-0114-y.
Neuhäuser, C. 2015. Some Sceptical remarks regarding robot responsibility and a way forward. In Collective agency and cooperation in natural and artificial systems, ed. C. Misselhorn, 131–146. Cham: Springer.
Picard, R. 1997. Affective computing. Cambridge, MA: MIT Press.
Powers, T.M. 2013. On the moral agency of computers. Topoi 32 (2): 227–236. https://doi.org/10.1007/s11245-012-9149-4.
Purves, D., R. Jenkins, and B.J. Strawser. 2015. Autonomous machines, moral judgment, and acting for the right reasons. Ethical Theory and Moral Practice 18 (4): 851–872. https://doi.org/10.1007/s10677-015-9563-y.
Raz, J. 1984. On the nature of rights. Mind 93 (370): 194–214.
Regan, T. 2004. The case for animal rights. Oakland: University of California Press.
Rueda, J., and F. Lara. 2020. Virtual reality and empathy enhancement. Ethical aspects. Frontiers in Robotics and AI 7: 506984. https://doi.org/10.3389/frobt.2020.506984.
Russell, S., and P. Norvig. 2005. AI: A modern approach. Learning 2 (3): 4.
Schwitzgebel, E., and M. Garza. 2015. A defense of the rights of artificial intelligences. Midwest Studies in Philosophy 39: 98–119. https://doi.org/10.1111/misp.12032.
Searle, John. 1980. Minds, brains, and programs. Behavioral and Brain Sciences 3 (3): 417–424.
Shen, S. 2011. The curious case of human-robot morality. In Proceedings of the 6th international conference on human-robot interaction, 249–250. New York: Association for Computer Machinery.
Shepherd, J. 2018. Consciousness and moral status. London: Routledge.
Shevlin, H. 2021. How could we know when a robot was a moral patient? Cambridge Quarterly of Healthcare Ethics 30 (3): 459–471. https://doi.org/10.1017/S0963180120001012.
Shoemaker, D. 2015. Responsibility from the margins. Oxford: Oxford University Press.
Singer, P. 1981. The expanding circle: Ethics and sociobiology. Oxford: Clarendon Press.
———. 2011. Practical ethics. Cambridge: Cambridge University Press.
Sliwa, P. 2015. Moral worth and moral knowledge. Philosophy and Phenomenological Research 93 (2): 393–418. https://doi.org/10.1111/phpr.12195.
Solum, L.B. 1992. Legal personhood for artificial intelligences. North Carolina Law Review 70: 1231.
Søraker, J.H. 2014. Continuities and discontinuities between humans, intelligent machines, and other entities. Philosophy & Technology 27 (1): 31–46. https://doi.org/10.1007/s13347-013-0132-9.
Sparrow, R. 2007. Killer robots. Journal of Applied Philosophy 24 (1): 62–77. https://doi.org/10.1111/j.1468-5930.2007.00346.x.
Strawson, P.F. 1962. Freedom and resentment and other essays. Proceedings of the British Academy 48: 1–25.
Sullins, J.P. 2011. When is a robot a moral agent? In Machine ethics, ed. M. Anderson and S.L. Anderson, 151–161. Cambridge: Cambridge University Press.
Tigard, D.W. 2021a. There is no techno-responsibility gap. Philosophy & Technology 34 (3): 589–607. https://doi.org/10.1007/s13347-020-00414-7.
———. 2021b. Artificial moral responsibility: How we can and cannot hold machines responsible. Cambridge Quarterly of Healthcare Ethics 30 (3): 435–447. https://doi.org/10.1017/S0963180120000985.
———. 2021c. Responsible AI and moral responsibility: A common appreciation. AI and Ethics 1 (2): 113–117. https://doi.org/10.1007/s43681-020-00009-0.
Torrance, S. 2013. Artificial agents and the expanding ethical circle. AI & SOCIETY 28 (4): 399–414. https://doi.org/10.1007/s00146-012-0422-2.
van Wynsberghe, A. 2022. Social robots and the risks to reciprocity. AI & SOCIETY 37 (2): 479–485. https://doi.org/10.1007/s00146-021-01207-y.
van Wynsberghe, A., and S. Robbins. 2019. Critiquing the reasons for making artificial moral agents. Science and Engineering Ethics 25 (3): 719–735. https://doi.org/10.1007/s11948-018-0030-8.
Verbeek, P.P. 2005. What things do? Philosophical reflections on technology, agency, and design. Pennsylvania: Pennsylvania State University Press.
———. 2011. Moralizing technology: Understanding and designing the morality of things. Chicago: University of Chicago Press.
Veruggio, G., and F. Operto. 2008. Roboethics: Social and ethical implications of robotics. In Springer handbook of robotics, ed. B. Siciliano and O. Khatib, 1499–1524. Berlin: Springer.
Véliz, C. 2021. Moral zombies: why algorithms are not moral agents. AI & SOCIETY 36: 487–497. https://doi.org/10.1007/s00146-021-01189-x.
Wallace, R.J. 2014. Practical reason. In The Stanford encyclopedia of philosophy, ed. E.N. Zalta. http://plato.stanford.edu/archives/sum2014/entries/practical-reason/.
Wallach, W. 2010. Robot minds and human ethics: The need for a comprehensive model of moral decision making. Ethics and Information Technology 12 (3): 243–250. https://doi.org/10.1007/s10676-010-9232-8.
Wallach, W., and C. Allen. 2008. Moral machines: Teaching robots right from wrong. Oxford: Oxford University Press.
Wareham, C.S. 2021. Artificial intelligence and African conceptions of personhood. Ethics and Information Technology 23 (2): 127–136. https://doi.org/10.1007/s10676-020-09541-3.
Warren, M.A. 1997. Moral status: Obligations to persons and other living things. Oxford: Clarendon Press.
Watson, G. 1996. Two faces of responsibility. Philosophical Topics 24 (2): 227–248.
Acknowledgements
This chapter was written as a part of the research projects Digital Ethics. Moral Enhancement through an Interactive Use of Artificial Intelligence (PID2019-104943RB-I00), funded by the State Research Agency of the Spanish Government, and Moral enhancement and artificial intelligence. Ethical aspects of a Socratic virtual assistant (B-HUM-64-UGR20), funded by FEDER/ Junta de Andalucía—Consejería de Transformación Económica, Industria, Conocimiento y Universidades. The authors are also grateful for the insightful comments of Jan Deckers on a previous version of this chapter.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this chapter
Cite this chapter
Llorca Albareda, J., García, P., Lara, F. (2023). The Moral Status of AI Entities. In: Lara, F., Deckers, J. (eds) Ethics of Artificial Intelligence. The International Library of Ethics, Law and Technology, vol 41. Springer, Cham. https://doi.org/10.1007/978-3-031-48135-2_4
Download citation
DOI: https://doi.org/10.1007/978-3-031-48135-2_4
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-48134-5
Online ISBN: 978-3-031-48135-2
eBook Packages: Religion and PhilosophyPhilosophy and Religion (R0)