Abstract
The question whether AI systems have agency is gaining increasing importance in discussions of responsibility for AI behavior. This paper argues that an approach to artificial agency needs to be teleological, and consider the role of human goals in particular if it is to adequately address the issue of responsibility. I will defend the view that while AI systems can be viewed as autonomous in the sense of identifying or pursuing goals, they rely on human goals and other values incorporated into their design, and are, as such, dependent on human agents. As a consequence, AI systems cannot be held morally responsible, and responsibility attributions should take into account normative and social aspects involved in the design and deployment of the said AI. My argument falls in line with approaches critical of attributing moral agency to artificial agents, but draws from the philosophy of action, highlighting further philosophical underpinnings of current debates on artificial agency.
Similar content being viewed by others
Notes
See Burr and Morley (2020) for a discussion of digital health technologies along the lines of the software mentioned.
See Tan and Lim (2018) for a review of current advances in AI, including machine learning.
This is a view held by Bryson and Kime (2011), in whose account the responsibility for AI systems should rest with the developer. While moral responsibility is not my main focus here, my view looks at human goals more broadly (including, for instance, the possibility of deliberation, or the involvement of institutions).
Another example would be certain human actions as well, like skilled action (see Clark 2010). I will not discuss this here.
I am grateful to an anonymous referee for bringing this problem to my attention.
This also helps set apart AI systems from other artifacts; see van de Poel (2020a: 399–400) for a comparison.
See D’Oro and Sandis (2013) for a historical overview.
Though I should note that they would not count as absolute presuppositions, which in Collingwood’s view do not have truth values. Rather, I take the falsity of these assumptions to lead to a worse account of AI behavior than approaches relying on different assumptions.
See Cardoso et al. (2020) on autonomy and the Curiosity rover.
I am grateful to an anonymous referee for this point.
Also see Elliott (2017) for a discussion of transparency about values in a general philosophy of science context.
I am grateful to an anonymous referee for raising this point.
References
Anscombe, G. E. M. (1957). Intention. Basil Blackwell.
Asilomar AI Principles (2017). Principles developed in conjunction with the 2017 Asilomar conference [Benevolent AI 2017],https://futureoflife.org/ai-principles/.
Barandiaran, X. E., Di Paolo, E., & Rohde, M. (2009). Defining agency: Individuality, normativity, asymmetry, and spatio-temporality in action. Adaptive Behavior, 17(5), 367–386.
Bedau, M. (1992). Goal-directed systems and the good. The Monist, 75, 34–49.
Beer, R. D. (1995). A dynamical systems perspective on agent-environment interaction. Artificial Intelligence, 72(1–2), 173–215.
Behdadi, D., & Munthe, C. (2020). A normative approach to artificial moral agency. Minds & Machines, 30, 195–218.
Brooks, R.A. (1991). Intelligence without reason. In J. Myopoulos & R. Reiter (Eds.), Proceedings of the 12th International Joint Conference on Artificial Intelligence (pp. 569–595). San Mateo: Morgan Kaufmann.
Bryson, J. J., & Kime, P. P. (2011). Just an artifact: Why machines are perceived as moral agents. In Twenty-second international joint conference on artificial intelligence.
Burr, C., & Morley, J. (2020). Empowerment or engagement? Digital health technologies for mental healthcare. In The 2019 Yearbook of the Digital Ethics Lab (pp. 67–88). Springer, Cham.
Cardoso, R. C., Farrell, M., Luckcuck, M., Ferrando, A., & Fisher, M. (2020). Heterogeneous verification of an autonomous Curiosity rover. In NASA Formal Methods Symposium (pp. 353–360). Springer, Cham.
Clark, R. (2010). Skilled activity and the causal theory of action. Philosophy and Phenomenological Research, 80(3), 523–555.
Collingwood, R. G. (2001). An essay on metaphysics. Oxford University Press.
Davidson, D. (1963). Actions, reasons, and causes. The Journal of Philosophy, 60(23), 685–700.
de Miranda, L. (2020). Artificial intelligence and philosophical creativity: From analytics to crealectics. Human Affairs, 30(4), 597–607.
Dennett, D. C. (1987). The Intentional Stance. MIT Press.
Dennett, D. C. (1988). Précis of the intentional stance. Behavioral and Brain Sciences, 11(3), 495–505.
Dignum, V. (2017). Responsible autonomy. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence (IJCAI’2017), pp. 4698–4704.
D’Oro, G., & Sandis, C. (2013). From anti-causalism to causalism and back: A century of the reasons/causes debate. Reasons and Causes: Causalism and Non-causalism in the Philosophy of Action, 1–47.
Elliott, K. C. (2017). A tapestry of values: An introduction to values in science. Oxford University Press.
Ezenkwu, C. P., & Starkey, A. (2019). Machine autonomy: Definition, approaches, challenges and research gaps. In Intelligent Computing-Proceedings of the Computing Conference (pp. 335–358). Springer, Cham.
Floridi, L., & Sanders, J. W. (2004). On the morality of artificial agents. Minds and Machines, 14(3), 349–379.
Floridi, L. (2008). The method of levels of abstraction. Minds and Machines, 18(3), 303–329.
Friedman, B., Kahn, P., & Borning, A. (2006). Value sensitive design and information systems. Advances in Management Information Systems, 6, 348–372.
Froese, T., Virgo, N., & Izquierdo, E. (2007). Autonomy: A review and a reappraisal. In European Conference on Artificial Life (pp. 455–464). Springer, Berlin, Heidelberg.
Gabriel, I. (2020). Artificial intelligence, values, and alignment. Minds & Machines, 30, 411–437.
Grodzinsky, F. S., Miller, K. W., & Wolf, M. J. (2008). The ethics of designing artificial agents. Ethics and Information Technology, 10(2–3), 115–121.
Gunkel, D. J. (2012). The machine question: Critical perspectives on AI, robots, and ethics. MIT Press.
Hacker, P. M. S. (2019). Wittgenstein: Meaning and mind (Volume 3 of an Analytical Commentary on the Philosophical Investigations), Part 1: Essays. John Wiley & Sons.
Hanfling, O. (2003). Wittgenstein and the human form of life. Routledge.
Illies, C., & Meijers, A. (2009). Artefacts without agency. The Monist, 92(3), 420–440.
Johansson, L. (2010). The functional morality of robots. International Journal of Technoethics, 1(4), 65–73.
Johnson, D. (2006). Computer systems: Moral entities but not moral agents. Ethics and Information Technology, 8(4), 195–204.
Johnson, D. G., & Miller, K. W. (2008). Un-making artificial moral agents. Ethics and Information Technology, 10(2–3), 123–133.
Johnson, D. G., & Verdicchio, M. (2017). Reframing AI discourse. Minds and Machines, 27(4), 575–590.
Johnson, D. G., & Verdicchio, M. (2018). Why robots should not be treated like animals. Ethics and Information Technology, 20(4), 291–301.
Johnson, D. G., & Verdicchio, M. (2019). AI, agency and responsibility: The VW fraud case and beyond. Ai & Society, 34(3), 639–647.
Kitcher, P. (2011a). Science in a democratic society. Prometheus Books.
Kitcher, P. (2011b). The ethical project. Harvard University Press.
Matthias, A. (2004). The responsibility gap: Ascribing responsibility for the actions of learning automata. Ethics and Information Technology, 6(3), 175–183.
Maturana, H.R., & Varela, F.J. (1980). Autopoiesis and cognition: The realization of the living. Boston Studies in the Philosophy and History of Science, 42. Dordrecht: Springer Netherlands.
Nolfi, S., & Floreano, D. (2000). Evolutionary robotics: The biology, intelligence, and technology of self-organizing machines. MIT Press.
Papineau, D. (2020). Naturalism, The Stanford encyclopedia of philosophy, Edward N. Zalta (ed.), https://plato.stanford.edu/archives/sum2020/entries/naturalism/
Peterson, M., & Spahn, A. (2011). Can technological artefacts be moral agents? Science and Engineering Ethics, 17(3), 411–424.
Popa, E. (2020). Artificial life and ‘nature’s purposes’: The question of behavioral autonomy. Human Affairs, 30(4), 587–596.
Schlosser, M. (2019), “Agency”, The Stanford encyclopedia of philosophy (Winter 2019 Edition), Edward N. Zalta (ed.),https://plato.stanford.edu/archives/win2019/entries/agency/.
Schumann, G. (Ed.). (2019). Explanation in action theory and historiography: Causal and teleological approaches. Routledge.
Sehon, S. (2010). Teleological explanation. A companion to the philosophy of action. Blackwell.
Tan, K. H., & Lim, B. P. (2018). The artificial intelligence renaissance: Deep learning and the road to human-level machine intelligence. APSIPA Transactions on Signal and Information Processing, 7.
Tigard, D. W. (2020). There is no techno-responsibility gap. Philosophy & Technology, 1–19.
van de Poel, I. (2020a). Embedding values in artificial intelligence (AI) systems. Minds and Machines, 30(3), 385–409.
van de Poel, I. (2020b). Three philosophical perspectives on the relation between technology and society, and how they affect the current debate about artificial intelligence. Human Affairs, 30(4), 499–511.
van den Hoven, J. (2005). Design for values and values for design. Information Age, 4, 4–7.
Von Wright, G. H. (1971). Explanation and understanding. Routledge & Kegan Paul.
Wiener, N. (1960). Some moral and technical consequences of automation. Science, 131, 1355–1358.
Wittgenstein, L. (1958). The blue and brown books. Blackwell.
Wittgenstein, L. (2009). Philosophical investigations, 4th edition, P.M.S. Hacker and Joachim Schulte (eds. and trans.), Oxford: Wiley-Blackwell.
Acknowledgements
I would like to thank Daniel Kodaj, Gunnar Schumann, László Bernáth, and the anonymous referees for their feedback, which has helped improve this article. This paper has been presented at the seminar “New Work on the Metaphysics of Teleology”, held at Central European University.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Popa, E. Human Goals Are Constitutive of Agency in Artificial Intelligence (AI). Philos. Technol. 34, 1731–1750 (2021). https://doi.org/10.1007/s13347-021-00483-2
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s13347-021-00483-2