Abstract
The dominant methodological assumptions in climate ethical debates are rational-individualistic. The aim of this paper is to examine whether the rational-individualistic methodological framework is compatible with a theory of moral responsibility for climate change. I employ three fitness criteria of moral agency: (1) a normatively significant choice, (2) sufficient knowledge and (3) control. I demonstrate that the rational-individualistic methodology does not provide a framework in which rational agents meet the three criteria. I conclude that rational-individualistic agents are not fit to be held morally responsible for climate change. The paper demonstrates that the dominant climate-ethical view rests on a methodology that does not allow for a conceptualisation of a moral agent of climate change.
Similar content being viewed by others
Notes
The discussion of control touches upon the philosophical debate on determinism and free will (Fischer and Ravizza 1998). In the present context, control refers to whether relevant choices are available within what is referred to later in this paper institutional behavioural spaces.
A theory is indirectly self-defeating ‘when it is true that, if someone tries to achieve his theory-given aims, these aims will be, on the whole, worse achieved’ (Parfit 1984, 5). The theory is not directly self-defeating if it remains rational for one to aim to achieve the theory-given aims (Parfit 1984, 13). Furthermore, self-interest theories are directly collectively self-defeating ‘when it is true that, if all of us successfully follow the self-interest theory, we will thereby cause our theory-given aims to be worse achieved than they would have been if none of us had successfully followed the self-interest-theory’ (Parfit 1984, 53).
A critical point of the rational-individualistic methodology should be added. It can be questioned whether it makes sense to discuss the moral significance of the future in economic discount rates. Why should the moral significance of the future and future people be dependent on an economic and rational theory of valuing? Most commonly, discount rates differentiate events temporarily. However, as Parfit rightly points out, it is a mistake to discount values for pure time. The fact that negative consequences may be temporarily (or geographically) remote does not make them less likely or less important. Indeed, when applied farther to the future, many predictions become more likely. Alternatively, people could discount for probability. Again, as Parfit explains, it may be rational to be less concerned about the more remote effects of our current acts and behaviour, but this would never occur because these effects are more remote. Rather, the reason for less concern would be because they are less likely to occur. In other words, if effects are likely to occur, it is rational to discount for them. Following this line of argument, we ought to be equally concerned about the predictable effects of our acts, whether these will occur in one, one hundred, or a 1000 years (Parfit 1984, 483–486).
The prisoner’s dilemma explains the occurrence of social coordination problems when (1) agents have complete information, e.g. they know the full structure of the game and the plausible outcomes, and (2) communication between the agents is forbidden or impossible. Consider the example illustrated in Matrix X below. Two persons are questioned separately about a joint crime. The attorney suggests two alternatives to each prisoner: confess to the crime or not confess. If neither confesses, they will both receive 2 years’ punishment for some minor crime [B, B].
Matrix X
Agent II
Agent I
A (confess)
B (not-confess)
A (confess)
10, 10
12, 0
B (not-confess)
0, 12
2, 2
If the persons both confess, they will be receive less than the most severe sentence [A, A] (see Parfit 1984, 56–60; Ostrom 1990, 3–28, 217). The initial preference ordering is [B, A], [B, B], [A, A] and [A, B] for both agents. This prisoner’s dilemma game illustrates the problem that it is impossible for the agent to choose the decision with the best outcome if she is unable to take into account the decisions of the other players. The agent’s dominant strategy is to play not-confess, but the outcome of the game depends on the choice of Agent II. Given that neither Agent I nor II knows what the other Agent plays, the initial preference ordering is unreliable and may be self-defeating.
The value of 20 units may be lower due to a decrease in marginal utility of the last 5–10 units.
An exchange situation in which the highest level of collective utility is not reached is called a Pareto-inferior exchange.
Following Ostrom, the preference function for Agent B can be represented by: u2 = π2 − δs, where π2 is payoff obtained and δs is the decrease in the value of π2 for violating the social norm (Ostrom 2005, 122).
Note that the snatch dilemma is not a test of the prisoner’s dilemma but emphasises instead the limited use of the prisoner’s dilemma in social dilemmas with common-pool resources (Ostrom 1990, 183–184).
Questions of distribution are set aside here.
By applying this institutional analysis to several empirical studies, Ostrom demonstrates that governing the commons sometimes succeeds and sometimes fails. (Ostrom 1990, 29). The explanation of success and failure can be found in the particular institutional arrangements.
Note that it is uncertain to what extent guilt and shame are effective in motivating people to engage in environmentally friendly behaviour. For further discussion (Swim et al. 2009).
To avoid discussion about to what extent it is fair to give priority to human entities over non-human entities, π(nature) and π(animals) are not taken into consideration. The subject of possible possitive effects of climate change is also set aside. Some evolutionary biologists argue that changes in the planet’s climate and atmosphere have positive consequences for new species and better living conditions for certain types of animals (Jablonski 2001, 5393–5398; de Perthuis 2011, 36).
Whereas the first two outcomes are examples of environmental pollution, the latter is an example of spatially and temporally diffused climate change; I distinguish between the two later in this paper.
By contrast, others believe that people have at least two sets of motives: one related to economics and self-interest and the other related to moral questions. If we accept this, one important challenge is to establish institutional arrangements that appeal to both sets of motives (Frey and Jegen 2000; Le Grand 2006).
To what extent it is impossible or too expensive for someone to buy an environmentally friendly car depends, of course, on her wealth. Such an expense is less expensive for wealthy people compared to poor people. This issue of financial inequality is set aside here.
References
Arrow, K. J. (1983). Behavior under uncertainty and its implications for policy. In B. P. Stigum & F. Wenstøp (Eds.), Foundations of utility and risk theory with applications. Dordrecht: D. Reidel Publishing Company.
Barry, B. (2005). Why social justice matters. Cambridge: Polity Press.
Bennett, J. (1995). The act itself. Oxford: Oxford University Press.
Caney, S. (2005). Justice beyond borders: A global political theory. Oxford: Oxford University Press.
Cohen, G. A. (1997). Where the action is: On the site of distributive justice. Philosophy & Public Affairs, 26(1), 3–20.
Cox, G. W. (2004). Lies, damned lies, and rational choice analyses. In G. W. Cox (Ed.), Problems and methods in the study of politics (Ch. 8, pp. 167–185). Cambridge: Cambridge University Press.
Dawes, R. M. (1980). Social dilemmas. Annual Review of Psychology, 31, 169–193.
de Perthuis, C. (2011). Economic choices in a warming world. Cambridge: Cambridge University Press.
Dietz, T., Ostrom, E., & Stern, P. C. (2003). The struggle to govern the commons. Science, 302(5652), 1907–1912.
Elster, J. (1982). Sour grapes—Utilitarianism and the genesis of wants. In A. Sen & B. Williams (Eds.), Utilitarianism and beyond (pp. 219–238). Cambridge: Cambridge University Press.
Fahlquist, J. N. (2009). Moral responsibility for environmental problems—Individual or institutional? Journal of Agricultural Environmental Ethics, 22, 109–124.
Fischer, J. M., & Ravizza, M. (1998). Responsibility and control: A theory of moral responsibility. Cambridge: Cambridge University Press.
Frey, B. S., & Jegen, R. (2000). Motivation crowding theory: A survey of empirical evidence. Zurich IEER working paper 26.
Hardin, G. (1968). The tragedy of the commons. Science, 162(3859), 1243–1248.
Hobbes, T. (1991/1651). Leviathan. Cambridge: Cambridge University Press.
Hurley, S. (2012). The public ecology of responsibility. In C. Knight & Z. Stemplowska (Eds.), Responsibility and distributive justice (pp. 187–215). Oxford: Oxford University Press.
Jablonski, D. (2001). Lessons from the past: Evolutionary impacts of mass extinctions. PNAS, 98(10), 5393–5398.
Jamieson, D. (2010). Ethics, public policy, and global warming. In S. Gardiner, S. Caney, D. Jamieson, & H. Shue (Eds.), Climate ethics: Essential readings. Oxford: Oxford University Press.
Le Grand, J. (2006). Motivation, agency, and public policy: Of knights and knaves, pawns and queens. New York: Oxford University Press.
List, C., & Pettit, P. (2011). Group agency: The possibility, design, and status of corporate agents. Oxford: Oxford University Press.
Lovelock, J. (2009). The vanishing face of Gaia: A final warning. New York: Basic Books.
Miller, S. (2010). The moral foundations of social institutions: A philosophical study. Cambridge: Cambridge University Press.
Ostrom, E. (1990). Governing the commons: The evolution of institutions for collective action. Cambridge: Cambridge University Press.
Ostrom, E. (2005). Understanding institutional diversity. Princeton: Princeton University Press.
Parfit, D. (1984). Reasons and persons. Oxford: Clarendon Press.
Plant, R. (2009). The neoliberal state. Oxford: Oxford University Press.
Prior, A. N. (1956). The consequences of actions. Proceedings of the Aristotelian Society, Supplementary, 30, 91–119.
Rosa, H., & Trejo-Mathys, J. (2013). Social acceleration: A new theory of modernity. New York: Columbia University Press.
Singer, P. (2010). One atmosphere. In S. Gardiner, S. Caney, D. Jamieson, & H. Shue (Eds.), Climate ethics: Essential readings. Oxford: Oxford University Press.
Stehr, N. (2008). Moral markets: How knowledge and affluence change consumers and products. Boulder, CO: Paradigm Publishers.
Stern, N. (2006). Stern review on the economics of climate change. Executive summary. London: HM Treasury.
Stern, N. (2010). The economics of climate change. In S. Gardiner, S. Caney, D. Jamieson, & H. Shue (Eds.), Climate ethics: Essential readings. Oxford: Oxford University Press.
Swim, J., et al. (2009). Psychology and global climate change: Addressing a multi-faceted phenomenon and set of challenges. A report by the American Psychological Association’s task force on the interface between psychology and global climate change. Washington: American Psychological Society.
Taylor, M. (1982). Community, anarchy and liberty. Cambridge: Cambridge University Press.
Acknowledgements
I am grateful to Simon Caney, Eva Erman and Christian Rostbøll for their comments on an earlier draft.
Author information
Authors and Affiliations
Corresponding author
Additional information
I am grateful to Simon Caney, Eva Erman, Anders Berg-Sørensen, Christian Rostbøll and Fabian Schuppert for their comments on an earlier draft.
Rights and permissions
About this article
Cite this article
Scavenius, T. The Issue of No Moral Agency in Climate Change. J Agric Environ Ethics 30, 225–240 (2017). https://doi.org/10.1007/s10806-017-9663-x
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10806-017-9663-x