1 Introduction

In the sixties and seventies of last century, the discovery of, what we call, ‘the paradox of predictability’ triggered a lively debate on determinism, predictability and free will, mainly in British philosophy. Although it ended without any clear upshot and with a lot of loose ends, we have three reasons to revisit and reopen this debate. First, the introduction of an explicit distinction between two different kinds of predictability, allows us to reformulate the paradox in a way that transparently reveals its underlying structure (Sect. 2). As a result, and this is our second and main point, we are able to show why the earlier explanations of the paradox failed (Sect. 3) and, subsequently, to give a new alternative explanation of its origin (Sect. 4). Third, we believe that the paradox of predictability might have important as yet unnoticed consequences for the contemporary compatibilism–incompatibilism debate on free will. In this paper, we elaborate the first two points, while leaving the third—apart from a promissory note (Sect. 5)—for another occasion.

2 The Paradox of Predictability

Before formulating the paradox of predictability, we introduce some basic terminology. We assume, first, that a universe U is deterministic when, for any arbitrarily chosen time t 0, there exists a law-like function f L which maps the initial state of the universe U 0 at time t 0 in a unique manner onto the state of the universe U t at any arbitrarily chosen later time t:Footnote 1

$$ U_{t} = f_{L} (U_{0} ) $$
(1)

We assume, furthermore, that determinism implies external (or Laplacean) predictability. This kind of predictability is defined as the possibility of a (God-like) external observer, not part of the universe U, to make predictions of all the future events in U on the basis of its perfect knowledge of the initial conditions U 0 and the law-like function f L .Footnote 2 For our purposes, it is, finally, crucial to distinguish this notion of external predictability from that of embedded predictability:Footnote 3

Embedded predictability holds in a universe, U, if there exists a subsystem, S, embedded in U—for instance, a highly intelligent (human) being or a very powerful computer—that is able to predict all the future events in U.

In the past, several authors argued that, especially but not exclusively with regard to non-linear systems, there are serious epistemic limitations on the ability of an embedded subsystem of a deterministic universe to make predictions of future events (see, for example, Suppes 1985; Bishop 2003). These epistemic limitations arise, for instance, because a finite subsystem of the universe is unable to make the required accurate measurements of the initial conditions of the universe, or because such a subsystem is unable to represent fully accurately the initial value of a variable, if this value is a (non-computable) irrational number (requiring an infinite string of digits). Though we believe that these arguments show that epistemic restrictions in fact or in practice preclude embedded predictability from holding in a deterministic universe, we want to argue for the stronger claim that embedded predictability cannot obtain in a deterministic universe even on the idealizing assumption that the epistemic limitations at issue are removed.Footnote 4 Therefore, we propose to lift these restrictions and to assume that the subsystem S in our definition of embedded predictability has infinite possibilities of representing information, is able to gather infinitely precise information about the initial conditions of the universe and is able to make calculations with any desired accuracy in a finite amount of time. We accept, of course, that it is physically impossible to realize these infinite capabilities in a finite subsystem of the universe. However, this reference to the infinite epistemic capabilities of subsystem S is not problematic because we simply use it as a shorthand for saying that the predicting subsystem can have arbitrarily strong computing capabilities. Authors who argue for unpredictability in terms of the epistemic limitations of predicting subsystems typically start from a given subsystem with finite capabilities and then construct a corresponding future event that remains beyond the predictive scope of this given subsystem. In contrast, we will identify a single future event that remains unpredictable for all possible predicting subsystems, no matter how strong their computing capabilities.

Already in 1950, Karl Popper has made an attempt to make this stronger case against embedded predictability. Although we think that his highly complex and at times overly technical arguments are of genuine interest, we believe that they are ultimately, as Popper himself seems to admit, not fully convincing (Popper 1950, p. 193).Footnote 5 A more successful attempt to make the stronger claim can be found implicitly in work of Donald MacKay (1960, 1961, 1967, 1971, 1973) and in a more explicit way in that of Scriven (1965) and P.T. Landsberg and D.A. Evans (1970, Evans and Landsberg 1972). These authors argued for the thesis that, conditional on some minimal assumptions about the universe U, there are fundamental limitations of a non-epistemic nature on the ability of any subsystem S to make predictions and that, therefore, even in the deterministic universe U, embedded predictability does not hold. They established this conclusion by constructing a paradoxical situation in which a subsystem S makes a prediction which is unavoidably self-defeating. Though we believe that their arguments provide a convincing example of a situation in which embedded predictability is impossible, we also believe that their accounts of this paradoxical situation are unwieldy and that, consequently, these authors failed to fully grasp the real origin of the paradox and the general nature of the unpredictability involved. Therefore, we proceed, first, by presenting our reconstruction of the paradox which allows us to lay out its underlying structure.Footnote 6

Suppose that, at an initial time t 0, the subsystem S 1 is asked to predict the future action, A, at a later time t 2, of another subsystem of the universe U, S 2. Assume that this action is that of making a choice between 0 and 1, such that either A = 0 or A = 1. Imagine, for instance, that subsystem S 2 is a human being who has to cast a vote in a referendum at time t 2. Assume, further, that S 1 has to make its prediction, call it P, at a specific time t 1, whereby t 0 < t 1 < t 2. Importantly, P is an identifiable physical event: S 1 is requested, for instance, to write down or print out the number 0 or 1 on a slip of paper. S 1’s predictive task can be formulated as

$$ P = A. $$
(2)

The paradox of predictability arises when the following two conditions obtain. First, there is some kind of interaction between the two subsystems, S 1 and S 2, such that the prediction P is revealed to the subsystem S 2. Imagine, for instance, that the human being is informed about, or secretly finds out the prediction of its future behaviour. Call this the revelation condition:

$$ P\,{\text{is}}\,{\text{revealed}}. $$
(3)

Second, assume that the predicted subsystem S 2 operates on the basis of, what Scriven calls, a contrapredictive mechanism.Footnote 7 This means that the subsystem S 2 always aims to defy any prediction made of its future actions and, as a consequence, always does the exact opposite of what is predicted. Imagine, for instance, that the human being has a recalcitrant personality and that it always does the opposite of what people expect. Formalized, the contrapredictivity condition is:

$$ {\text{If}}\,P\,{\text{is}}\,{\text{revealed}},\,{\text{then}}\,A = not\hbox{-}P. $$
(4)

Putting both conditions, (4) and (3) in a modus ponens format directly generates the paradox. Given that the actual behaviour of the (contrapredictive) subsystem S 2 (after revelation) is exactly the opposite of the predicted behaviour P, that is,

$$ A = not\hbox{-}P, $$
(5)

S 1’s predictive task, from (2) and (5), must be described by the unsolvable equation:

$$ P = not\hbox{-}P. $$
(6)

When the revelation and contrapredictivity conditions obtain, if S 1’s prediction at t 1 is P = 0, then S 2’s action at t 2 will be A = 1; if S 1’s prediction at t 1 is P = 1, then S 2’s action at t 2 will be A = 0. The paradoxical nature of the situation thus resides in the fact that any prediction made by S 1 is inevitably self-defeating.

3 How Not to Explain the Paradox (Away)

We briefly elucidate our reconstruction of the paradox of predictability against the backdrop of the earlier 1960–1970 discussion. We believe that, until now, no convincing explanation of the underlying structure of the paradox has been given. To dispel some of the misunderstandings which have clouded the debate, four important clarifications are in order.

First, the self-defeating structure does not, as some have mistakenly thought (DeWitt 1973), depend on some supposedly unique ontological characteristics of cognitive events, or on the presence of some special kind of free will. There is nothing extraordinary about a contrapredictive mechanism. We can safely assume that such a mechanism can be instantiated by a variety of biological, psychological or electronic mechanisms. Contrapredictivity could, for instance, easily be programmed into a computer, requiring the generation of output 1 if and only if receiving input 0, and vice versa (Landsberg and Evans 1970, pp. 356–357; Good 1971).

Second, the self-defeating structure is not, as some authors wrongly assume (Goldman 1968, pp. 148–149), attributable to some lack of knowledge on the part of the predictor S 1. Additional knowledge about the physical details of the interaction or the workings of the subsystem S 2 will not change the logical outcome that the predictive task yields the same unsolvable equation P = not-P. The fact that this equation is unsolvable does not, however, generate some kind of indeterminacy. We assumed universe U to be deterministic and, therefore, what will happen at t 1 and later at t 2 is uniquely determined. In order to know what will happen, we could make the relevant knowledge complete and add the assumption that if the predictor S 1 fails to arrive at a unique solution for P, it will make the ‘default’ prediction

$$ P = 0. $$
(7)

In this case, the prediction at t 1 will be P = 0 and, therefore, the action at t 2 will be A = 1. This uniquely determined course of events will be clear to everybody, even to the predictor S 1. However, this knowledge does not help the subsystem S 1: it also knows that it is unable to generate the correct output P = 1, for this prediction would again be self-defeating and bring about action A = 0.

Third, it is not the case, contrary to what several authors assume (Lewis and Richardson 1966; Good 1971; also Roberts 1975), that the paradox has something to do with limitations on the speed with which the predictor and the predicted system can make calculations concerning each other’s behaviour. These authors seem to have in mind a scenario in which the predictor is able to make a second and secret prediction, P′, which would be correct (say, P′ = 1). Yet, when this secret prediction is revealed, the predicted system will decide to act contrapredictively as regards this second prediction, which will require a third secret prediction, P″ (=0), which in its turn could be revealed, etc. This to-and-fro scenario generates a series of predictions and contrapredictive intentions, such that the subsystem that calculates the fastest will outsmart the other subsystem at the time of action, t 2.

However, this kind of consideration does not genuinely alleviate the paradox. There are three possibilities. We could, first, suspend our assumption that the predicting subsystem has infinite calculating capabilities and interpret this to-and-fro scenario as an actual to-and-fro process, whereby each prediction and each contrapredictive calculation takes a finite amount of time. Whether, in this case, the final prediction made by predictor S 1 before time t 2 will be correct, depends entirely on the contingent characteristics of the calculating and interacting mechanisms. Although, in this scenario, the stronger claim that we can identify a single future event which is unpredictable for all possible predicting subsystems no longer holds, the argument against embedded predictability as such remains. Remember that to demonstrate that embedded predictability cannot obtain, it suffices, strictly speaking, to show that, for every would-be predictor S 1, we can think of at least one situation in which it will fail to make a correct prediction. In this scenario, we can, therefore, secure the paradoxical result simply by stipulating that the predicted subsystem S 2 calculates faster than the predicting subsystem S 1.

The second interpretation is also straightforward. If we consider the to-and-fro scenario as an infinite series in a finite amount of time (by assuming, for instance, that both subsystems are able to make infinitely fast calculations), then it simply captures the typically oscillating nature of the paradoxical equation P = not-P (Wormell 1958). We are again just saying that this equation has no solution and that the predictor will have to resort to the default prediction P = 0, which is self-defeating.

On a third interpretation, we could make the asymmetrical assumption that the predicting subsystem S 1 is capable of making infinitely fast calculations, whereas the predicted subsystem S 2 is only capable of finitely fast ones. Whilst this third scenario is the only one in which the proposed ‘solution’ of the paradox actually works, and S 1 indeed always outsmarts S 2, this resolution comes at a high price. Not only does the asymmetry of the idealizing assumption seem arbitrary, it also remains a fact that by choosing an arbitrarily fast predicted subsystem S 2, we can make the time-gap ε which remains between prediction P and action A arbitrarily small. The claim that S 1 can actually ‘predict’ the behaviour of S 2, given a very small value for ε, reduces then to the claim that S 1 can fulfil its predictive task simply by waiting until S 2 is no longer capable of reacting to any relevant information from its environment, due to a lack of time. Since this scenario makes the idea of a prediction rather meaningless, we have introduced in our version of the paradox the plausible requirement that the prediction has to be produced at a specific time t 1 (<t 2). This assumption simply captures the idea that a real prediction needs to refer to an event that is in some significant sense situated in a finitely (as opposed to infinitesimally) remote future.

Importantly, the determination of a specific time t 1 (<t 2) at which the prediction has to be made immediately blocks all three versions of the to-and-fro scenario. The determinate, finite lapse of time between the prediction and the action allows the predicted subsystem to respond to P and the unpredictability of A by S 1 at time t 1 is restored. Since A is thus an event which remains unpredictable, independently from the specific characteristics of the predicting subsystem, our stronger claim that embedded predictability cannot obtain even on the idealizing assumption that S 1 has infinite calculating capabilities still stands in our version of the paradox.

Four, and in line with our first two clarifications, the paradox does not reveal some kind of ‘essential unpredictability in human behaviour’, as Michael Scriven (1965) wants us to belief. To begin with, neither his version of the paradox, nor ours essentially presupposes that the nature of the predicted subsystem S 2 is human. Neither of both versions makes a distinction between human and non-human agency. In addition, and more importantly, the situation we have sketched remains throughout fully determined and, therefore, no essential unpredictability is implied at all (Suppes 1964, p. 144). As we will show, Scriven’s claim is based on a serious flaw in his own argument. Close scrutiny of where it goes wrong is worthwhile, since it helps to reveal the real origin of the paradox.

Scriven’s version of the paradox differs from ours in one design feature. Although he too presupposes a contrapredictive mechanism, he rejects the idea of an actual interaction between the predicting and the predicted system. Instead, he assumes that the predicted system has access to the same information I and has the same predictive capacities as the predicting system. In this way, the predicted system can itself make the very same prediction P (without the need for interaction or revelation) and, subsequently, act contrapredictively. As we will explain in the next section, we believe that this difference in the design of the assumptions is in itself not really significant and does not actually change the structure of the paradoxical situation.

Vital to our purposes is that Scriven fails to make the crucial distinction between embedded and external predictability. Scriven proceeds in his argument e contrario ad absurdum. He adopts the assumption that action A can be logically derived from the available information I and then argues that the contrapredictive behaviour of the predicted system leads to an immediate falsification of this assumption:

It immediately follows that I either implies an incorrect prediction as to which A will be chosen by S 2 (that is, contains false information) or none: hence S 2’s choice cannot be rationally predicted by S 1 from I. (Scriven 1965, p. 415)Footnote 8

Though the second part of this crucial passage—S 2’s choice cannot be predicted by S 1 from I—is correct, the first part is not and, pace Scriven, it does not follow from the contrapredictive behaviour of S 2. Scriven’s reasoning goes wrong because he fails to distinguish between two different kinds of predictability. On the one hand, Scriven talks about the implication of A by I, thereby referring to the logical inference from I to A. On the other hand, Scriven talks about the actual deduction of A from I by S 1 or S 2, which is a physical event. As will become clear in the next section, the impossibility of the actual deduction of A by S 1 does not necessarily imply that the available information is insufficient to logically infer A. On the contrary, even in the paradoxical situation at issue, the deterministic nature of universe U guarantees, contrary to what Scriven supposes, that it must be possible to provide sufficient information from which to make the correct inference to A. Scriven conflates the external kind of predictability with the embedded one. From the standpoint of external predictability there is no mystery at all. On the basis of the relevant information about the initial conditions and the laws of universe U, an external observer can predict that S 1 will make a certain prediction P at time t 1 (in our default case, P = 0) and can also unequivocally predict that the action A at time t 2 will be the exact opposite of what has been predicted (A = 1). From the external standpoint, human behaviour in U remains inferentially predictable in principle. So, the first part of the quoted passage above—I either implies an incorrect prediction as to which A will be chosen by S 2 (…) or none—is false.

The paradox of predictability does not arise because action A cannot be logically inferred from the information available, but because the embedded subsystem S 1 cannot deduce its prediction without thereby invalidating it. More specifically, S 1’s prediction is not an external logical inference. The paradox, we conjecture, arises because S 1 is itself an embedded part of universe U and, therefore, its prediction a physical event which is itself part of the causal chain of events.

4 Predictions as Physical Events

Exploring the distinction between external and embedded predictability somewhat further, we can clarify why and how the paradox of predictability is inextricably connected with the embeddedness of S 1 as a subsystem of universe U. This furnishes the material for our alternative explanation of the paradox’s origin and its underlying structure.

Our paradox is a specific instantiation of a general situation of embedded prediction in which a subsystem S 1 makes a prediction P at time t 1 about a future action A at time t 2 of another subsystem S 2. To fully describe this general situation from the standpoint of external predictability, we introduce some additional equations. Recall equation (1): if the universe U is deterministic, U’s state at a certain time, t, is a law-like function of U’s initial state at t 0, whereby this law-like function defines a necessary and thus unique chain of events. Given that prediction P is an element of U 1’s state and that action A is an element of U 2’s state, we can, therefore, introduce two law-like functions, g and h, which describe how P and A, respectively, are uniquely determined by the initial conditions of the universe at time t 0. Using, in addition, the symbol P* to refer to the prediction made by the external observer, the situation for this external observer is captured by the following set of three equations.

$$ P = g(U_{0} ) $$
(8)
$$ A = h(U_{0} ) $$
(9)
$$ P^{*} = A $$
(10)

These equations mean the following. Equation (8) describes how prediction P, as a physical event at time t 1, is uniquely determined by the state of U at time t 0 and the law-like function g. Equation (9) describes, similarly, how action A, as a physical event at time t 2, is uniquely determined by the state of U at time t 0 and the law-like function h. Equation (10) describes the predictive task of the external predictor, who has to make sure that its prediction P* adequately predicts action A. Importantly, equation (10) has no physical interpretation because the prediction P* is not itself a physical event, but just the name for the prediction made by the external (and thus non-physical) observer.

As this set of equations makes clear, an external predictor has no problem whatsoever in understanding and describing the chain of events in the paradoxical situation at hand. On the basis of its knowledge of g, h and U 0 (and its calculating abilities), this observer is perfectly capable of determining the values of P and A and, thus, to fulfil its predictive task (10). If we apply the general description to the specific case of the paradox, equation (8), in our default case, simply becomes the default prediction P = 0 and equation (9) on the revelation and contrapredictivity conditions becomes equation A = not-P. Clearly, these equations allow the external observer to make correctly (10) the prediction P* = 1.

Turning to the standpoint of embedded predictability, the more general situation can be described by a similar yet different set of three equations.

$$ P = g(U_{0} ) $$
(8’)
$$ A = h(U_{0} ) $$
(9’)
$$ P = A $$
(11)

The first two equations (8′) and (9′), are identical to the first two equations of the previous set, (8) and (9), and mean the same. They describe how prediction P and action A, as physical events, are uniquely determined by the initial state of U at time t 0 and the laws describing its evolution. The third equation (11), however, is different and describes the predictive task of the embedded predictor S 1. In contrast to equation (10), and this is a point of great consequence, equation (11) has a physical interpretation. Prediction P at time t 1 can only be successful if it adequately describes the future action A at time t 2. In our case, this means that the number produced by the predictor S 1 at time t 1 has to be the same as the number chosen by S 2 at time t 2.

As the equations make clear, the fact that P, in contrast to P*, is a physical event has a dramatic impact on the ability of the embedded predictor to fulfil its task. Whereas the first set of equations (8)–(10) provides three equations from which to determine the value of three variables (P, A and P*), the set of equations (8′)–(9′)–(11) provides three equations from which to determine the value of only two variables (P and A). In the second set of equations, the variables are thus overdetermined and, consequently, a simultaneous solution of all the equations is no longer assured. Overdetermination possibly implies unsolvability.

The explanation of this essential difference is the following. The deterministic nature of the universe implies that the first two equations of each set already uniquely determine the value of P and A. The third equation of the second set, (11), imposes an additional physical constraint on an already fully determined universe. Equation (11) thus overdetermines the variables P and A. Prediction P is successful only if a certain relation exists between prediction P as a physical event and the predicted action A as a physical event (whereby P has to provide an adequate representation of the future action A). Whether this relation actually obtains and, thus, whether a successful prediction (P = A) is possible, cannot be guaranteed a priori but depends on the specifics of the laws and conditions of the universe and of the situation at hand.

By contrast, in the case of an external predictor, this uncertainty does not arise. In this case, the deterministic nature of the universe guarantees a priori that a successful prediction P* is always possible. The first two equations (8) and (9) allow the external predictor to determine the values of P and A. Although it is thereby perfectly possible that these two values turn out to be different (P ≠ A), this only means that the external predictor observes that the embedded predictor is unsuccessful. This observation, however, in no way affects the ability of the external predictor to make its own successful prediction P* = A.

Although we have no general theory to offer about the conditions under which the set of equations (8′)–(9′)–(11) is unsolvable and, therefore, an embedded prediction impossible, we would like to draw attention to a specific category of situations which seem particularly prone to overdetermination. We have in mind situations in which action A becomes an explicit function of prediction P. By this we mean that function h(U 0) can be written as function f(g(U 0)), so that equation (9′) becomes

$$ A = f(P). $$
(12)

When this new equation for A is combined with the additional physical constraint, (11), it becomes apparent that this category of situations is captured by the self-referential equation (compare Landsberg and Evans 1970, pp. 348–350):

$$ P = f(P). $$
(13)

To be sure, there are cases in which even equation (13) remains solvable. Imagine, for instance, a situation similar to the paradoxical one, apart from the fact that the predicted subsystem S 2 is no longer characterized by a contrapredictive mechanism but by a fatalist mechanism which always acts in accordance with a revealed prediction:

$$ {\text{If}}\,P\,{\text{is}}\,{\text{revealed}},\,{\text{then}}\,\;A = P. $$
(14)

In this case, the actual revelation of P generates a situation in which f becomes the relation of identity and equation (13) straightforwardly reduces to the (solvable) equation

$$ P = P. $$
(15)

Nevertheless, it is clear that this is an exceptional case and that, generally, self-referential situations as expressed by (13) will not yield solutions.Footnote 9 The paradoxical situation itself provides a case in point. The presence of the contrapredictive mechanism, as described by

$$ {\text{If}}\,P\,{\text{is}}\,{\text{revealed}},\,{\text{then}}\, \, A = not\hbox{-}P $$
(4)

implies that the functional dependence f of action A on prediction P is the one expressed by negation. Hence, the self-referential equation (13) becomes the unsolvable equation described earlier as

$$ P = not\hbox{-}P. $$
(6)

As these two examples make clear, self-referentiality, as functionally expressed by (13), can arise as the result of a direct causal interaction between the predicting subsystem S 1 and the predicted subsystem S 2. In our two examples, we have assumed that the revelation condition, (3), obtains and, thus, that S 2 somehow receives information about the prediction P after it has been made. In these situations, the functional dependence f of action A on prediction P can properly be understood as a description of the causal impact the prediction has as a physical event. In the paradoxical case, the causal impact of the prediction as a physical event is such that it inevitably falsifies the prediction itself.

However, such a direct causal interaction is not a necessary condition for the functional dependence, (12), to arise. As Scriven’s version of the paradox makes clear (see Sect. 3), the existence of a common causal history can suffice. Remember that Scriven drops the revelation condition, (3), and replaces it with the assumption that S 2 is itself capable of calculating the prediction that is made by S 1. In our description, this means that S 2 is capable of solving equation (8′) and to use this knowledge of P to act contrapredictively. This version of the paradox thus generates the same functional dependence of A on P as in our version:

$$ A = not\hbox{-}P. $$
(5)

While in this case the functional dependence of A on P arises without any actual causal interaction after time t 1 between the prediction as a physical event and the action itself, this dependence now crucially turns upon the presence of a common causal history. In our account, this common history is encapsulated by the initial conditions U 0 such that both prediction P and action A causally depend on these conditions as expressed by equations (8′) and (9′), respectively.Footnote 10

The presence of a common causal history is a necessary condition for Scriven’s version of the paradox to work. This can be demonstrated by taking a look at the hypothetical situation in which Scriven’s contrapredictive subsystem S 2 were to attempt to outsmart the external predictor by trying to calculate the prediction P* and then act contrapredictively with regard to this prediction as follows:

$$ A = not\hbox{-}P^{ *} . $$
(16)

In this case, the externality of the predictor implies that his prediction is not a physical event which causally depends on the initial conditions U 0 and that there is, consequently, no equivalent for equation (8′) from which S 2 could try to calculate P*. Instead, the only information about P* available to S 2 is contained in the predictive task of the external predictor, captured by the combination of equations (9) and (10) as

$$ P^{*} = h(U_{0} ). $$
(17)

However, since h(U 0) is, by equation (9), identical to action A, the hypothetical scenario in which S 2 were to use (17) to determine P* and, subsequently, were to use this knowledge to act contrapredictively, as specified by (16), would as a result have an action which is the negation of itself:

$$ A = not\hbox{-}A. $$
(18)

The fact that such an action cannot exist demonstrates that subsystem S 2 is unable to act contrapredictively with regard to an external observer with which it does not share a common causal history.

Although our account of the overdetermination involved in a case of embedded prediction has focused, thus far, on the category of more tractable situations in which action A is directly functionally dependent on prediction P (equation 12), the generality of the three equations (8′)–(9′)–(11) indicates that the problem of overdetermination might also be much more general. The laws of our universe seem to be such that all sorts of interactions and functional dependencies between two embedded subsystems unavoidably exist. Physics tells us, for instance, that two of the fundamental forces in the universe (gravity and the electromagnetic force) have an infinite scope. This means that the actual physical workings of the predicting subsystem (containing biological or electronic devices which calculate and store the prediction) generate causal factors which will inevitably affect the future evolution of the predicted subsystem. Owing to this causal interconnectedness, one can expect that the set of equations (8′)–(9′)–(11) will often generate problems of self-referentiality similar to those of equation (13).

In fact, the ubiquity of causal interactions seems to imply a reversal of our initial problem. Given that functional dependencies are almost inescapable, how then is it at all possible that we are actually capable of making embedded predictions of future events? The answer is, of course, that in many situations when we want to make predictions we are able to discount or shield of our own past and present interactions with the predicted system at hand. For instance, if we want to calculate the orbit of Jupiter and predict its position at a specific time in the future, we may safely discount the gravitational impact of our calculations (as a physical activity) on its orbit. In such a situation, the future event still remains functionally dependent on the present state of the universe, but will be, for all practical purposes, sufficiently causally uncoupled from our predicting activities. This uncoupling eliminates the self-referentiality of the situation and allows us to find a single solution that describes and predicts the future event.

Needless to say, however, that such an independence is never complete and only possible to a sufficient extent for some (but not all) specific ‘local’ problems and purposes. A general and complete causal uncoupling of our past and future activities from the predicted subsystem, would require, among other things, that we were capable of obtaining all the information needed for our predictions without actually disturbing the predicted system. This would require, what Popper (1950, p. 129) called, a ‘one-way membrane’ between the predictor and the predicted system. Because such a membrane cannot physically exist, the possibility of complete causal uncoupling is only available to a Laplacean predictor as a non-physical observer outside the universe.Footnote 11

5 Take-It-or-Leave-It Control

In the 1960–1970 debate, it was widely assumed that the paradox of predictability was relevant for the debate on freedom and determinism. Especially Donald MacKay (1960, 1961, 1967, 1971, 1973) tried to use the paradox as the starting point for a compatibilist theory, according to which, even in a deterministic universe, human brain-states remain ‘logically indeterminate’ until the moment of ‘free’ choice. Although MacKay’s approach is interesting in many respects, we believe that the works of Landsberg and Evans (1970), Evans and Landsberg (1972) and Grünbaum (1971) have convincingly shown that MacKay ultimately failed in developing a plausible compatibilist position.Footnote 12

We believe that this failure is not accidental but rather the effect of an underlying assumption which MacKay and most other authors in the earlier debate seemed to have shared. They all assumed a direct conceptual connection between freedom of human actions and their (embedded) unpredictability. We believe, on the contrary, that this direct connection fails to hold and that (embedded) unpredictability is neither a sufficient nor a necessary condition for human freedom. It is not sufficient, since, as we have noted, it is perfectly possible that ‘soulless’ machines programmed in a contrapredictive manner remain unpredictable. It also seems unreasonable to suggest that unpredictability is a necessary condition for freedom. If someone or something predicts that tomorrow I am going to help my neighbour fix the shed in the garden and if, as a matter of fact, I see no good reason not to do so, it seems implausible to suggest that my freedom to choose is undermined simply because the prediction is indeed in line with my personality and my reasons for action.

While these considerations imply that no direct inference from the lack of embedded predictability to the possibility of human freedom can be drawn, we conjecture that the paradox of predictability leads to results which might be relevant for the compatibilism-incompatibilism debate on free will. Introducing some basic terminology will be helpful.Footnote 13 If determinism, as characterized by (1) in Sect. 2, is true, the facts of the past, together with the laws of nature, entail all facts of the present and future. Indeterminism is the denial of determinism. Compatibilism is the view that free will, free action and moral responsibility are compatible with determinism; incompatibilism is the denial of compatibilism. Libertarians are incompatibilists (and indeterminists) who believe that at least some of us, at times, perform free actions for which we are responsible. Hard determinists are incompatibilists who deny that we have the sort of free will required for moral responsibility; hard incompatibilists deny the same, irrespective of accepting determinism or not. Incompatibilists standardly hold that responsibility-relevant freedom can be nothing less than regulative control (the sort of control which involves the existence of genuinely open alternative possibilities), whereas compatibilists argue that such freedom only requires guidance control. At a minimum, our conjecture pertains to guidance control.Footnote 14

To briefly substantiate our suggestion that the paradox of predictability might (at least) be relevant for the compatibilist notion of guidance control, we propose to return for a moment to the self-referential situations discussed in the previous section. Remember that we introduced both a contrapredictive mechanism, characterized by (4), which always goes against revealed predictions and whose outcome is therefore always unpredictable, and a fatalist mechanism, characterized by (14), which always conforms to a revealed prediction and whose outcome is therefore always predictable. Relevant for our purposes is that it is furthermore possible to consider a series of more complex intermediary situations with mechanisms which reject a revealed prediction when other sufficient conditions for rejection (SCR) hold true and are also revealed:

$$ {\text{If}}\,P\,{\text{is}}\,{\text{revealed}}\,{\text{and if}}\,{\text{SCR}}\,{\text{are}}\,{\text{revealed}},\,{\text{then}}\,A = not\hbox{-}P. $$
(19)

Situations in which this type of condition-responsive mechanism operates in a subsystem S 2, the predictability of action A depends on the additional SCR-conditions. If these are in fact revealed to hold, then the embedded predictor S 1 faces the by now familiar contrapredictive self-referential and unsolvable equation

$$ P = not\hbox{-}P. $$
(6)

As explained, this means that S 1’s embedded prediction will be inevitably self-defeating. Yet, in case the SCR-conditions are not revealed or fail to obtain, a solution for the pertinent equations is not necessarily excluded and, depending on the details of the situation, an adequate prediction might remain possible.

The presupposition that we actually live in a deterministic universe is compatible with the additional and, in our view, plausible assumption that human beings are in possession of (deterministic) psychological mechanisms of the condition-responsive kind, indicated by (19). This means that, even in a deterministic universe, a human being who is confronted with a prediction about its own future behaviour might consider several other conditions and, depending on whether or not these hold, decide to go against or, alternatively, decide to act in accordance with the prediction made. We illustrate the types of situation we have in mind by considering two possible sufficient conditions for rejection. Suppose, first, that the subsystem which predicts our actions does so on the basis of its foreknowledge for the purpose of manipulating us. In this case, if the prediction and the manipulative intentions of the predictor are revealed, we could be motivated to act contrapredictively. Similarly, it could turn out that a prediction of my future behaviour is based on the discovery of some kind of biological or sociological mechanism which determines my actions. In this second case, such a discovery could lead me to reassess what I previously considered as my authentic preferences and my authentic reasons for making my decisions. As a result, I could decide, on the basis of the more extensive information now available to me, that I no longer wish to pursue my biologically or socially determined urges and desires.Footnote 15 If, as in these two examples, manipulation or the presence of inauthentic preferences is interpreted as a SCR-condition, and if this condition is revealed, then the working of (deterministically) operating condition-responsive psychological mechanisms warrants that the predictions made about our future behaviour will inevitably be self-defeating.

Our analysis of the paradox demonstrates that embedded subsystems in a deterministic universe retain, what we call, a take-it-or-leave-it control as regards revealed predictions. The specifics of the operating mechanisms of such an embedded subsystem establish the conditions under which it will accept or reject predictions about its future behaviour made by an other embedded subsystem in the same universe. Whether or not this consequence will turn out to be important for the contemporary debate on the compatibility of human freedom with determinism depends on the additional assumption that take-it-or-leave-it control is an essential part of a convincing analysis of free human agency. We conjecture that this assumption holds water, but a further elaboration of such an analysis is beyond this paper’s scope.