1 Introduction

The critics of rational choice theory (henceforth, RCT) frequently claim that RCT is self-defeating in the sense that agents who abide by RCT’s prescriptions are less successful in satisfying their preferences than they would be if they abided by some normative theory of choice other than RCT (e.g. Bratman 1999, 2000; Gauthier 1984, 1997; Kavka 1978, 1983; McClennen 1990, 1997). The idea is that abiding by RCT’s prescriptions hampers (rather than enhances) agents’ ability to satisfy their preferences, and that an agent who abides by RCT’s prescriptions will often “end up satisfying his preferences less well than he would have done, had he [abided by] some other [theory]” (Sugden 1991, 752; also Bradley 2007 and 2017; Broome 2007a and 2007b; Dietrich et al. 2013 and 2019; Rabinowicz 1995 and 2019; Spohn 2009 and 2012, for recent discussions). In this paper, I combine insights from philosophy of action, philosophy of mind and the normative foundations of RCT to rebut this often-made criticism. I then explicate the implications of my thesis for the wider philosophical debate concerning the normativity of RCT for both ideal agents who can form and revise their intentions instantly without cognitive costs and real-life agents who have limited control over the formation and the dynamics of their own intentions.Footnote 1

The paper is organized as follows. In Sect. 2, I examine one issue that figures centrally in the debate as to whether RCT is self-defeating, namely whether an agent who abides by RCT’s prescriptions can rationally form what I call anomalous intentions, i.e. intentions to perform actions that maximize the total stream of payoffs the agent can get over the entire course of a decision problem, but fail to maximize the payoffs the agent can get from some subsequent choice nodes onwards (e.g. Bratman 1999, 2000; Gauthier 1984, 1997; Kavka 1978, 1983; McClennen 1990, 1997). I then explicate my thesis that, despite prominent criticisms of RCT, agents can rationally form anomalous intentions, and therefore prominent attempts to demonstrate that RCT is self-defeating do not withstand scrutiny. In Sects. 36, I defend my thesis that agents can rationally form anomalous intentions against four major objections put forward in the specialized literature, namely: the objection from temporal situatedness (e.g. Bratman 1998; Mintoff 1997); the objection from bootstrapping (e.g. Bratman 2009; Broome 2013); the objection from psycho-physical inability (e.g. Farrell 1989; Shah 2009); and the overdemandingness objection (e.g. Mongin 2000; Steele 2006).Footnote 2

My thesis that agents can rationally form anomalous intentions has at least three implications of general interest for the wider philosophical debate concerning the normativity of RCT (e.g. Bradley 2007, 2017; Broome 2007a, b; Dietrich et al. 2013, 2019; Rabinowicz 1995, 2019; Spohn 2009, 2012). First, anomalous intentions figure in a vast range of decision problems where the payoffs agents can get if they form the intention to perform an action are at least partly independent of the payoffs agents can get if they actually perform such action (e.g. Andreou 2006; Clarke 2007; van Hees and Roy 2009). Hence, showing that agents can rationally form anomalous intentions would have significant bearing on the normativity of RCT across several decision problems. Second, the claim that agents cannot rationally form anomalous intentions is commonly premised on the assumption that instrumental rationality requires agents to maximize the payoffs they can get from the choice nodes they face onwards irrespective of whether maximizing these payoffs maximizes the total stream of payoffs agents can get over the entire course of a decision problem (e.g. Bratman 1998, 62–66, Mintoff 1997, 624–5, Williams 1981, 35). Below I critically examine this widely endorsed conception of instrumental rationality and argue that in presence of anomalous intentions such conception has implications that contrast with independently plausible requirements of payoff maximization. And third, several authors build on the claim that RCT is self-defeating to argue that this theory must be revised or even rejected (e.g. Bratman 1999, 2000; Gauthier 1984, 1997; Kavka 1978, 1983; McClennen 1990, 1997). If my thesis that agents can rationally form anomalous intentions is correct, prominent attempts to demonstrate that RCT is self-defeating do not withstand scrutiny. This result does not per se vindicate RCT as our best available normative theory of choice. Still, it challenges RCT’s critics to put forward more convincing reasons and evidence to support their claim that RCT is self-defeating.Footnote 3

Before proceeding, one preliminary remark is in order. Various characterizations of preferences and intentions have been advocated in the economic and philosophical literatures (e.g. Cozic and Hill 2015; Dietrich and List 2016; Guala 2019; Hausman 2011; Jeffrey 1965; Savage 1954; Thoma 2017, on preferences; Anscombe 1963; Bratman 1987; Davidson 1978; Holton 2009; Roy 2009; Searle 1983; Tenenbaum 2018, on intentions). I shall expand on these characterizations wherever my evaluation directly rests on those characterizations (e.g. footnote no.7 on the so-called belief constraint on intending; also footnote no.17 for a comparison between intentions and beliefs). For now, it suffices to note that although intentions do not figure in all applications of RCT, many applications of RCT model decision-making in terms of the formation and the dynamics of intentions (e.g. Audi 1991; Bales 2020; Cullity 2008; Mele 2000; Pink 1991), and the debate concerning the putative self-defeating character of RCT often targets such applications (Sects. 26).

2 RCT and anomalous intentions

In this section, I examine the issue whether an agent who abides by RCT’s prescriptions can rationally form what I call anomalous intentions, i.e. intentions to perform actions that maximize the total stream of payoffs the agent can get over the entire course of a decision problem, but fail to maximize the payoffs the agent can get from some subsequent choice nodes onwards. To clarify this issue, I focus on a putative paradox that is commonly taken to indicate that RCT is self-defeating, namely Kavka’s (1983) ‘toxin puzzle’. I then explicate my thesis that, despite prominent criticisms of RCT, agents can rationally form anomalous intentions even in this putative paradox, and therefore prominent attempts to demonstrate that RCT is self-defeating do not withstand scrutiny.

Kavka explicates his toxin puzzle as follows:

“An eccentric billionaire [offers] you the following deal. He places before you a vial of toxin that, if you drink it, will make you painfully ill for a day, but will not threaten your life or have any lasting effects. […] The billionaire will pay you one million dollars tomorrow morning if, at midnight tonight, you intend to drink the toxin tomorrow afternoon. […] You need not drink the toxin to receive the money; in fact, the money will already be in your bank account hours before the time for drinking it arrives, if you succeed. […] All you have to do is sign the agreement and then intend at midnight tonight to drink the stuff tomorrow afternoon. You are perfectly free to change your mind after receiving the money and not drink the toxin. [However, arranging] external incentives is ruled out, as are such alternative gimmicks as hiring a hypnotist to implant the intention, forgetting the main relevant facts of the situation, and so forth. […] The presence or absence of the intention is to be determined by the latest 'mind-reading' brain scanner [which] will correctly detect the presence or absence of the relevant intention. [The question is whether today you can rationally form] the intention to drink the toxin [tomorrow]” (1983, 33–34).

Kavka’s puzzle is often represented by the following decision tree with two choice nodes (e.g. Van Hees and Roy 2009, Fig. 4), one before midnight where the agent decides whether or not to form the intention to drink the toxin (assuming the agent is psycho-physically able to form such intention) and the other tomorrow afternoon where the agent decides whether or not to drink the toxin (assuming the agent has formed the intention to drink the toxin)Footnote 4:

figure 1

Before assessing the proposed solutions of this puzzle, let us distinguish between what I call global payoffs (henceforth, GP), i.e. the total stream of (expected or actual) payoffs an agent can get over the entire course of a decision problem, and subsequent payoffs (henceforth, SP), i.e. the stream of (expected or actual) payoffs an agent can get from the choice nodes she faces onwards.Footnote 5 At the initial choice node of a decision problem, the set of actions that maximize one’s GP and the set of actions that maximize one’s SP coincide. At later choice nodes, the two sets of actions may significantly differ, and maximizing one’s GP may require one to “perform an action other than the one that at the time of performance would [maximize her SP]” (Gauthier 1994, 697).Footnote 6

To illustrate this, consider again the decision problem envisaged by Kavka. In this decision problem, maximizing GP requires the agent to form the intention to drink the toxin despite anticipating that, when it comes to drinking the toxin, drinking it will fail to maximize her SP. In particular, forming the intention to drink the toxin and drinking the toxin is the course of action that maximizes the agent’s GP. To be sure, the agent could conceivably get even higher GP if she succeeded in forming the intention to drink the toxin today and later revised such intention before drinking the toxin. However, under Kavka’s (1983, 34) assumptions, the agent anticipates since the first choice node whether she will revise the intention to drink the toxin before drinking the toxin. Moreover, if the agent anticipates that she will revise the intention to drink the toxin, she will not be able to form such intention and she will fail to get the million (ibid., 34). Hence, of the courses of action that the agent can successfully plan to perform, forming the intention to drink the toxin and drinking it are part of the course of action that maximizes the agent’s GP.Footnote 7

Is it rational for an agent who abides by RCT’s prescriptions to form the intention to drink the toxin in the decision problem envisaged by Kavka? Two sets of solutions are prominent in the literature. On the one hand, the globalists hold that an agent can always rationally form the intention to drink the toxin because, of the courses of action that the agent can successfully plan to perform, forming the intention to drink the toxin and drinking it are part of the course of action that maximizes the agent’s GP (e.g. Gauthier 1994, 721, McClennen 1990, 230–1). The globalists’ reasoning proceeds as follows. One can rationally form an intention if one has reason to expect that forming and acting on this intention are part of the course of action that maximizes her GP (e.g. Gauthier 1998, 50). Of the courses of action that the agent can successfully plan to perform in the toxin puzzle, forming the intention to drink the toxin and drinking it are “part of the best course of action that [the agent] could embrace as a whole” (Gauthier 1998, 48). For these actions are the only way for the agent to get the million, which by assumption makes her better off even at the cost of drinking the toxin (Gauthier 1994, 702–9; also McClennen 1990, 230–1). Therefore, an agent can always rationally form the intention to drink the toxin.Footnote 8

On the other hand, the localists hold that an agent cannot rationally form the intention to drink the toxin in the decision problem envisaged by Kavka because the agent knows that, when it comes to drinking the toxin, drinking it will fail to maximize her SP, and one cannot rationally form the intention to perform an action that she knows will fail to maximize her SP (e.g. Kavka 1983, 34–5; also Bratman 1998, 62 and 72–73; Quinn 1985, 371). The localists’ reasoning proceeds as follows. One can rationally form an intention if one has reason to expect that forming and acting on this intention are part of the course of action that maximizes her SP. In the toxin puzzle, the million gives the agent reason to form the intention to drink the toxin because, by assumption, the agent will be better off if she gets the million and drinks the toxin, than if she does not form the intention to drink the toxin.Footnote 9 Yet, when it comes to drinking the toxin, the agent will know whether or not she has got the million. And at that point, it will be irrational for the agent to drink the toxin. For drinking the toxin will bring the agent a day of illness without yielding any apparent benefit. Moreover, the agent is assumed to know all these facts today. Hence, today the agent cannot rationally form the intention to drink the toxin tomorrow, and she fails to get the million.Footnote 10

The globalist and the localist solutions are grounded on different conceptions of instrumental rationality, which take actions (e.g. forming specific intentions and acting on such intentions) to be instrumentally rational to the extent that these actions maximize agents’ GP and SP, respectively. These two solutions concur in cases where the set of actions that maximize SP and the set of actions that maximize GP coincide, but diverge in cases where these sets of actions differ. In the remainder of this paper, I critically examine those two solutions and argue that neither solution is sufficiently sensitive to the extent RCT’s prescriptions vary depending on agents’ ability to control the formation and the dynamics of their own intentions. More specifically, I shall argue that: (1) contrary to the localist solution, “an action may be rational even though at the time of performance it [does] not, and is not believed to [maximize SP]” (Gauthier 1994, 701), and so it can be rational for agents to form anomalous intentions such as the intention to drink the toxin in the decision problem envisaged by Kavka; and (2) contrary to the globalist solution, whether agents can rationally form anomalous intentions crucially depends on the extent to which agents are able to control the formation and the dynamics of their own intentions, and so it is not the case that agents can always rationally form anomalous intentions such as the intention to drink the toxin in the decision problem envisaged by Kavka.

My claims (1) and (2) agree with the globalists that the rationality of forming intentions is to be assessed in terms of whether forming and acting on such intentions are part of the course of action that maximizes agents’ GP. However, they reject the globalist solution that agents can always rationally form anomalous intentions. In particular, they hold that whether it is rational for agents to form such intentions crucially depends on the extent to which these agents are able to control the formation and the dynamics of their own intentions. In Sects. 36, I defend my claims (1) and (2) against four major objections put forward in the specialized literature, namely: the objection from temporal situatedness (e.g. Bratman 1998; Mintoff 1997); the objection from bootstrapping (e.g. Bratman 2009; Broome 2013); the objection from psycho-physical inability (e.g. Farrell 1989; Shah 2009); and the overdemandingness objection (e.g. Mongin 2000; Steele 2006). In doing so, I explicate the implications of my claims for the wider philosophical debate concerning the normativity of RCT for both ideal agents who can form and revise their intentions instantly without cognitive costs and real-life agents who have limited control over the formation and the dynamics of their own intentions.Footnote 11

3 Objection from temporal situatedness

The objection from temporal situatedness holds that agents cannot rationally form anomalous intentions because instrumental rationality requires agents to maximize their SP irrespective of whether maximizing SP enables agents to maximize their GP (e.g. Bratman 1998, 62–66, Mintoff 1997, 624–5). The objection proceeds as follows. Forming an intention is instrumentally rational for an agent only if forming and acting on this intention enables the agent to maximize her SP (e.g. Williams 1981, 35). However, anomalous intentions are defined as intentions to perform actions that—while maximizing one’s GP—fail to maximize one’s SP. Therefore, agents cannot rationally form anomalous intentions. Paraphrasing Bratman’s remarks about instrumentally rational choice in sequential decision problems, “the agent may well rank her alternatives with respect to past [intentions, but] what is now under her control are her alternatives from now on [and one ought to base her decisions on] one's ranking of options that are at that time in one’s control” (1998, 66).

This objection points to a widely endorsed localist conception of instrumental rationality, which takes actions (e.g. forming specific intentions and acting on such intentions) to be instrumentally rational to the extent that these actions maximize agents’ SP. This localist conception of instrumental rationality has plausible implications in decision problems where the set of actions that maximize one’s SP and the set of actions that maximize one’s GP coincide. In presence of anomalous intentions, however, these two sets of actions differ. In these situations, it would be implausible to hold that instrumental rationality invariably requires one to maximize SP (rather than GP). For in presence of anomalous intentions, agents who aim to maximize their SP often obtain lower (expected or actual) payoffs than the (expected or actual) payoffs they would obtain if they aimed to maximize their GP (e.g. DeHelian and McClennen 1993; Machina 1989). To be sure, the fact that agents who aim to maximize their SP often obtain lower (expected or actual) payoffs than the (expected or actual) payoffs they would obtain if they aimed to maximize their GP does not imply that agents can always rationally form the intention to perform actions that maximize their GP (e.g. Sect. 5 on cases where agents know that they are psycho-physically unable to form such intentions). Yet, it indicates that one can rationally form anomalous intentions.

To illustrate this, consider the following transfer variants of Kavka’s toxin puzzle, where the agent gets the million tomorrow if today she forms the intention to give up (e.g. give back to the billionaire) a specified part of the million upon receiving the million the day after. These variants retain the assumptions of Kavka’s original puzzle (e.g. if the agent forms the relevant intention, the million will be in her bank account before the time of acting on such intention; the agent is free to change her mind after receiving the million and before acting on the relevant intention; irreversible arrangements that bind the agent to act on the relevant intention are ruled out by assumption), but eliminate reference to the toxin and its illness-related effects. In this way, they bypass various tangential issues affecting Kavka’s original puzzle (e.g. commensurability of money prizes and states of illness) and make it easier to assess the merits of the proposed solutions of such puzzle.Footnote 12

Is it rational for an agent who abides by RCT’s prescriptions to form the intention to give up a specified part of the million upon receiving the million the day after? According to the localist solution, an agent cannot rationally form the intention to give up any positive part of the million upon receiving the million the day after, since giving up any positive part of the million would fail to maximize the agent’s SP. This solution might seem plausible in transfer variants of the puzzle where the agent is required to form the intention to give up the whole (or most of the) million, but is far less plausible in variants where the agent is required to form the intention to give up a small part of the million. In fact, it would be rather implausible to claim that an agent cannot rationally form the intention to give up any positive part of the million upon receiving the million the day after.

To see this, consider a minimum transfer variant of the puzzle, where the agent gets the million tomorrow if today she forms the intention to give up 1 dollar upon receiving the million. In this context, the mere fact that giving up 1 dollar will fail to maximize the agent’s SP falls short of implying that the agent cannot rationally form the intention to give up such dollar. For such dollar will be of extremely low value to the agent upon receiving the million, and the agent can gain 999,999$ by forming the intention to give up such dollar compared to a situation where she does not form such intention. This point holds not merely for a conveniently selected subset of transfer variants of the toxin puzzle, but generalizes across a wide subset of such variants. That is to say, given an arbitrarily small amount of money ε > 0 that (while being lower than 1 million) is large enough to make a difference to the agent’s payoff valuations, one can construct a transfer variant of the puzzle where the agent gets the million if today she forms the intention to give up ε. To claim that the agent cannot rationally form such intention irrespective of how small ε is amounts to a reductio ad absurdum of the localist solution. To put it differently, one can rationally form the intention to perform actions that maximize her GP but fail to maximize her SP, and the claim that instrumental rationality invariably requires one to maximize her SP (rather than GP) does not withstand scrutiny.Footnote 13

4 Objection from bootstrapping

The objection from bootstrapping holds that agents cannot rationally form anomalous intentions because intending to perform an action cannot per se “bootstrap a new reason into existence” that makes it rational to perform such action (Broome 2007b, 354; also Bratman 1987, 24–27, 2009, 415–6). The objection proceeds as follows. Forming the intention to perform an action can indirectly create reasons to perform this action (e.g. Sobel 1994, on cases where forming an intention makes one’s choice situation change in ways that make it rational to act on such intention). However, forming the intention to perform an action cannot per se create any new reason to perform this action, i.e. any reason to perform such action that one did not have before forming the intention. For “if it did, we could give ourselves a reason to [perform an action] just by intending to [perform] it; and that cannot be right” (Holton 2004, 513; also Broome 2013, ch.5). Now, the objection goes, the fact that forming the intention to perform an action cannot per se create any new reasons to perform this action, together with the fact that agents have no reason to perform actions that fail to maximize their SP, implies that agents have no reason to form the intention to perform such actions. Hence, agents cannot rationally form anomalous intentions.Footnote 14

This objection correctly notes that in many decision problems, the mere fact that one forms the intention to perform an action does not per se make it rational for her to perform such action. However, this does not exclude that “[one’s] reasons for performing an action can derive from her reasons for forming the preceding intention” (Gauthier 1994, 709; also Harman 1998, 84). Moreover, realizing that forming and acting on an anomalous intention are part of the course of action that maximizes one’s GP provides one with reason to form and act on such intention (e.g. Gauthier 1994, 721, Smith 2016, 2263–4). This reason, despite the globalists, is not always strong enough to make it rational for one to form anomalous intentions (e.g. Holton 2004).Footnote 15 Yet, despite the localists, in cases where aiming to maximize GP enables one to obtain higher (expected or actual) payoffs than the (expected or actual) payoffs she would obtain if she aimed to maximize SP, such reason can be strong enough to make it rational for an agent to form anomalous intentions. In fact, one may envision several decision problems where very large discrepancies between the (expected or actual) payoffs one would obtain by aiming to maximize her GP and the (expected or actual) payoffs one would obtain by aiming to maximize her SP cast doubt on the localists’ claim that instrumental rationality invariably requires agents to maximize their SP rather than GP (e.g. Sect. 5 below on large prize variants of the toxin puzzle).Footnote 16

A proponent of the objection from bootstrapping may concede that forming anomalous intentions may yield significant benefits to an agent (e.g. Bratman 2000, on coordination with one’s future selves). However, she may object that these benefits fall short of licensing the claim that agents can rationally form anomalous intentions. The objection proceeds as follows. Whether an agent can rationally form specific intentions is “a function of the […] rationality of performing the actions that they are intentions to perform” rather than the benefits derivable from forming such intentions (Farrell 1989, 293; also Parfit 2001, 21–22, on the contrast between object-given reasons for attitudes, which derive from facts about the objects of agents’ propositional attitudes, and putative state-given reasons for attitudes, which derive from facts about agents’ having such propositional attitudes). Hence, to settle the question whether it is rational for her to intend to drink the toxin, an agent must settle the question whether it is rational for her to drink the toxin (e.g. Goetz 1998; Shah 2009). Unfortunately, the agent anticipates that, when it comes to drinking the toxin, it will be irrational for her to drink the toxin because drinking the toxin fails to maximize her SP (e.g. Mintoff 1997; Quinn 1985). Therefore, the agent cannot rationally form the intention to drink the toxin.

This objection invites two rejoinders. First, the mere fact that an action such as drinking the toxin fails to maximize an agent’s SP does not per se indicate that such action is irrational unless one already presupposes a localist conception of instrumental rationality. Hence, reiterating that drinking the toxin fails to maximize an agent’s SP does not per se provide any independent reasons to think that the agent cannot rationally form the intention to drink the toxin. And second, in presence of anomalous intentions, an agent may have reasons to form an intention (e.g. maximizing one’s GP) which make it rational to form and act on this intention even if an agent’s reasons to form an intention do not generally coincide with the agent’s reasons to act on such intention (e.g. Clarke 2008; Pink 1998). More specifically, realizing that forming and acting on the intention to drink the toxin are part of the course of action that maximizes the agent’s GP provides the agent with reason to form and act on such intention. This reason, despite the globalists, is not always strong enough to make it rational for an agent to form the intention to drink the toxin (e.g. Sobel 1994). Yet, despite the localists, such reason can make it rational for the agent to form the intention to drink the toxin even if the agent knows that drinking the toxin fails to maximize her SP.

To illustrate this, consider the following psychological variant of the toxin puzzle, where the billionaire offers the toxin deal to a confident agent whose psychology is so constituted that she believes that she will drink the toxin tomorrow irrespective of whether today she forms the intention to drink the toxin (e.g. Mele 1992). Today, the confident agent can rationally form the intention to drink the toxin tomorrow even if she knows that tomorrow drinking the toxin will fail to maximize her SP. For today this agent has strong reasons both to form the intention to drink the toxin tomorrow and to believe that she will drink the toxin. In this respect, a localist may well object that the confident agent is irrational on the alleged ground that a rational agent, knowing that drinking the toxin will fail to maximize her SP, would regard drinking the toxin as irrational and would believe that she will not drink the toxin. However, as noted in the previous paragraph, the mere fact that an action such as drinking the toxin fails to maximize an agent’s SP does not per se indicate that such action is irrational unless one already presupposes a localist conception of instrumental rationality. Hence, reiterating that drinking the toxin fails to maximize an agent’s SP does not per se provide any independent reasons to think that a rational agent would regard drinking the toxin as irrational and would believe that she will not drink the toxin. This point holds not merely for the aforementioned confident agent, but generalizes for several agents whose degree of belief that they will drink the toxin is lower than 1. That is to say, given an arbitrarily small probability p > 0 that an agent thinks she will drink the toxin, one can construct a variant of the toxin puzzle where the (expected or actual) payoffs the agent can get by forming and acting on the intention to drink the toxin are large enough to license the claim that forming such intention maximizes the agent’s GP. To claim that the agent cannot rationally form such intention irrespective of how large these payoffs are amounts to a reductio ad absurdum of the localist solution (e.g. Sect. 5 below for similar remarks concerning large prize variants of the toxin puzzle).Footnote 17

5 Objection from psycho-physical inability

The objection from psycho-physical inability holds that real-life agents cannot rationally form anomalous intentions because they lack sufficient control over the formation and the dynamics of their own intentions to be able to reliably form anomalous intentions (e.g. Farrell 1989; also Shah 2009, on cases where one would benefit greatly from forming an intention, yet is unable to form it because she doubts that she will be able to act on it). The objection proceeds as follows. Consider an anomalous intention such as the intention to drink the toxin in the decision problem envisaged by Kavka. Suppose some real-life agent realizes that forming and acting on this anomalous intention is part of the course of action that maximizes her GP. By itself, this realization does not enable the agent to form such anomalous intention. For the agent knows that, when it comes to performing the relevant action (i.e. drinking the toxin), performing this action will fail to maximize her SP. And this knowledge, in turn, undermines the agent’s ability to form the intention to perform such action. For real-life agents are psycho-physically unable to form intentions to perform actions that they know will fail to maximize their SP. Paraphrasing Farrell, an agent’s intention to perform an action that she knows will fail to maximize her SP “would necessarily be unstable [since] reflection on the fact that [this intention] is directed towards an action which [fails to maximize SP] will inevitably undermine it” (1989, 288).

This objection correctly notes that the realization that performing a particular action fails to maximize one’s SP may reduce (or even undermine) one’s ability to form the intention to perform such action. Still, it is an open empirical question how often (and to what extent) this realization reduces real-life agents’ ability to form anomalous intentions (e.g. Holton 2004). In fact, variations in the payoffs involved in specific variants of the toxin puzzle may significantly affect real-life agents’ ability to form anomalous intentions in such variants. For instance, in transfer variants of the toxin puzzle, relatively few agents would presumably be able to form today the intention to give up the specified part of the million if this part amounted to 999,999$. Yet, comparatively more agents would be able to form this intention for lower amounts of money, and even more agents would be able to form such intention if they were required to give up just 1 dollar upon receiving the million. For as noted in Sects. 23, such dollar will be of extremely low value to agents upon receiving the million, and agents can gain 999,999$ by forming the intention to give up such dollar compared to a situation where they do not form such intention.

More generally, the point remains that real-life agents’ purported inability to form specific anomalous intentions constitutes an empirical limitation in agents’ ability to control their own intentions, but does not directly bear against the normative claim that real-life agents can rationally form anomalous intentions. To be sure, whether real-life agents can rationally form specific anomalous intentions may significantly depend on their ability to form such intentions (e.g. footnote no.20 on putative cases where agents cannot rationally form specific anomalous intentions because they know that it is psycho-physically impossible for them to form such intentions). Moreover, real-life agents are often unable to form the intention to perform an action for the sole reason that forming and acting on this intention maximizes their GP (e.g. Shah 2009). Still, these limitations concern the empirical issue whether real-life agents are psycho-physically able to form specific anomalous intentions, and do not directly bear against the normative claim that real-life agents can rationally form anomalous intentions.Footnote 18

A proponent of the objection from psycho-physical inability may object that real-life agents’ control over their own intentions is so limited that it is rather unlikely that these agents are able to form anomalous intentions for the sole reason that forming and acting on these intentions maximizes their GP (e.g. Farrell 1989). Suppose, for the sake of argument, that this objection is correct. Assume further that real-life agents are aware of this limitation on their ability to control their own intentions. Even this does not bear against the claim that real-life agents can rationally form anomalous intentions. To see this, consider situations where an agent knows that it is rather unlikely that she will be able to form an anomalous intention (e.g. the intention to drink the toxin). In some of these situations, the agent cannot rationally try to form this anomalous intention because the probability that she will be able to form such intention is overly low to license the claim that trying to form such intention maximizes her GP.Footnote 19 In other such situations, instead, the agent can rationally try to form this anomalous intention because the (expected or actual) payoffs she can get by forming and acting on the intention are large enough to license the claim that trying to form such intention maximizes her GP (e.g. McCann 1986). That is to say, given an arbitrarily small probability p > 0 that the agent is able to form a given anomalous intention (e.g. the intention to drink the toxin), one can construct a large prize variant of the toxin puzzle where the (expected or actual) payoffs the agent can get by forming and acting on the intention are large enough to license the claim that trying to form such intention maximizes the agent’s GP. To claim that the agent cannot rationally try to form such intention irrespective of how large these payoffs are amounts to a reductio ad absurdum of the localist solution.Footnote 20

6 Overdemandingness objection

The overdemandingness objection holds that real-life agents cannot rationally form anomalous intentions because calculating whether forming specific anomalous intentions maximizes their GP is cognitively too costly for them or simply beyond the epistemic access they have to earlier choice nodes of sequential decision problems (e.g. Mongin 2000; Steele 2006). The objection proceeds as follows. The mere fact that anomalous intentions are intentions to perform actions that maximize agents’ GP does not per se imply that real-life agents can calculate whether forming specific anomalous intentions maximizes their GP in the sequential decision problems they face. In particular, calculating whether forming specific anomalous intentions enables one to maximize her GP in the sequential decision problems she faces would require one to “keep track of earlier [choice nodes of such problems] that are no longer possible” (Steele 2006, 9). The cognitive costs of keeping track of these choice nodes might be assumed away in idealized decision problems such as the toxin puzzle (Sect. 2). However, for real-life agents it is often cognitively too costly to keep track of earlier choice nodes of the sequential decision problems they face (e.g. Gilboa 2009, ch.11–12). In fact, real-life agents frequently lack the epistemic access to earlier choice nodes required to calculate whether forming specific anomalous intentions enables them to maximize their GP in the sequential decision problems they face (e.g. Mongin 2000). These limitations, in turn, make it dubious that real-life agents can rationally form anomalous intentions.Footnote 21

This objection correctly notes that calculating whether forming specific anomalous intentions enables real-life agents to maximize their GP in the sequential decision problems they face often involves high cognitive costs for such agents. Still, it is doubtful that these cognitive costs are generally so high that they prevent real-life agents who perform such calculations from maximizing their GP. Moreover, one may consistently hold that in many sequential decision problems keeping track of earlier choice nodes is ‘burdensome’ (Steele 2006, 19), yet enables real-life agents to obtain higher (expected or actual) payoffs than the (expected or actual) payoffs they would obtain if they did not keep track of such choice nodes. In fact, one can envision several sequential decision problems where the benefits derivable from keeping track of earlier choice nodes outweigh the cognitive costs involved in keeping track of such choice nodes (e.g. Sect. 5 above on large prize variants of the toxin puzzle). In this respect, it would be of little import to object that in several sequential decision problems, it is cognitively too costly for real-life agents to calculate whether keeping track of earlier choice nodes enables them to maximize their GP. For substantiating the claim that real-life agents can rationally keep track of earlier choice nodes only requires one to show that there are some sequential decision problems where the benefits derivable from keeping track of these choice nodes outweigh the cognitive costs involved in keeping track of such choice nodes. That is to say, both GP maximization and SP maximization may be more or less cognitively demanding depending on what sequential decision problems are faced by real-life agents (e.g. footnote no.21). This variability makes it dubious that real-life agents are always able to maximize their (expected or actual) payoffs in sequential decision problems (e.g. think of highly extended sequential decision problems with more stages and nodes than the involved real-life agents are able to keep track of). However, it does not provide any reason to think that, in general, real-life agents cannot rationally form anomalous intentions (e.g. think of less extended sequential decision problems like Kavka’s puzzle).

A proponent of the overdemandingness objection may concede that keeping track of earlier choice nodes of sequential decision problems frequently enables real-life agents to obtain higher (expected or actual) payoffs than the (expected or actual) payoffs they would obtain if they did not keep track of such choice nodes. Still, she may object that assessing the rationality of forming anomalous intentions requires agents to calculate the cognitive costs and benefits involved in forming such intentions, and that the rationality of performing these calculations itself depends on further calculations, leading to a regress (e.g. Mongin 2000, 95). Suppose, for the sake of argument, that assessing the rationality of forming anomalous intentions requires agents to calculate the cognitive costs and benefits involved in forming such intentions. This does not per se imply that the rationality of performing these calculations itself depends on further calculations. Moreover, even if the rationality of performing these calculations itself depended on further calculations, this would not per se entail a regress. For there are empirical limitations on how many orders of calculations real-life agents can perform (e.g. van Hees and Roy 2009, on agents’ cognitive and computational limitations). These empirical limitations make it dubious that real-life agents are always able to maximize their (expected or actual) payoffs in sequential decision problems (e.g. Mongin 2000, 102). However, they do not provide any reason to think that, in general, real-life agents cannot rationally form anomalous intentions. To put it differently, regress-based considerations cast doubt on the claim that real-life agents can always rationally form anomalous intentions, but do not bear against the thesis that real-life agents can rationally form anomalous intentions.

7 Conclusion

Over the last few decades, several prominent authors have built on the claim that agents cannot rationally form anomalous intentions to criticize RCT for being self-defeating. In this paper, I argued that despite these prominent criticisms of RCT, both ideal agents who can form and revise their intentions instantly without cognitive costs and real-life agents who have limited control over the formation and the dynamics of their own intentions can rationally form anomalous intentions. If my thesis is correct, prominent attempts to demonstrate that RCT is self-defeating do not withstand scrutiny. This result does not per se vindicate RCT as our best available normative theory of choice. Still, it challenges RCT’s critics to put forward more convincing reasons and evidence to support their claim that RCT is self-defeating.