1 Introduction

You are as certain as you can be that the stone paperweight on your desk does not feel pain even though you have no evidence of this. Indeed, the fact that there is no evidence is precisely the reason why you are so certain. Each of us is certain of our own capacity for pain and pretty nearly certain of the capacity of other human beings to feel pain. On the spectrum between stones and human beings lie creatures about which it is harder to judge with near certainty but about which we hope at least to be able to make informed judgements. As we approach single-cell organisms, we will inevitably have to rely on absence of evidence reasoning in ruling against a species’ capacity for pain. Null hypotheses might seem dull in comparison to hypotheses claiming that protozoa or plants or, if one is inclined to panpsychism, every particle of matter, is conscious, but they play an important scientific role. In inferential statistics, an absence of evidence provides evidentiary support for the null hypothesis. Sometimes, and the science of consciousness is a case in point, an absence of evidence might well be evidence of absence.

In what follows, we focus our attention on pain not only for illustrative purposes but because pain is robustly coupled to emotion, motivation and behaviour in humans, and thus, in comparative studies, invites arguments from analogy based on multiple criteria. With respect to pain, there seems to be a standard argumentative approach that has been adopted for either accepting or rejecting the pervasive existence of pain in phylogenetically-diverse organisms. On the negative side, it is argued first that motor behaviours do not necessarily indicate feeling states such as pain, and second, that defined neural circuitry is necessary for such states, and in its absence, there can be no pain. These ideas, which we endorse, are closely linked to the well-established evolutionary and bioengineering principle that structure-determines-function. On the positive side, it is counterargued that the absence of evidence for a certain kind of neural circuitry is not evidence of an absence of pain. It exists and is somehow—although unknown—multiply realized in different species. The dual concepts of “absence of evidence” and “multiple realizability” have consequently become closely intertwined in these arguments. In this paper, we focus on both of these issues and their intersection with the aim of moving the debate over the pervasiveness of pain in animals closer towards some form of closure.

Pain research is notorious for making absence of evidence reasoning based on specific neurological criteria appear less conclusive. Point out that a species does not have a neocortex (necessary for pain in humansFootnote 1) and researchers will often reject the implication and look instead for some other criterion—neural or behavioural—as indicative of the species’ capacity for pain. Different argumentative strategies are deployed at that point—some leaning toward the conclusion that pain is “ambiguous” (Klein 2008); others toward the conclusion that pain has a different realizer in different species (“realizer functionalism”; McLaughlin 2006; Armstrong 1968; Lewis 1970, 1972; Kim 1998, 2005); and others towards the conclusion that pain is a higher-order functional property—the property of having some first-order realizer that fulfils the functional role of pain (“role functionalism”; McLaughlin (2006)). What motivates such moves is an interwoven set of assumptions: that all genuinely causal or explanatory properties are functional (Shoemaker 1984, 1998, 2001); that statements about unobservable psychological events need to be “Ramseyfied” to be rendered scientific—i.e., translated into sentences existentially quantifying only over observable properties (typically observable causes and effects of the psychological event) and individuals (Lewis 1970, 1972); and the idea that, for any function, there is an open disjunction of actual or possible realizers of that function—the “multiple realizability” assumption (Putnam 1967). Viewed through the lens of comparative neurobiology, these assumptions, if true, rule out an absence of evidence hardly ever lending support to the null hypothesis about a species’ capacity for pain so long as there is some observable mapping of inputs (noxious stimuli) to outputs (nocifensive behaviour). This, we shall endeavour to show, is deeply problematic.

In Sect. 1, we defend absence of evidence reasoning drawing on, among other things, results from informal logic. These results delineate the kinds of epistemic contexts in which absence of evidence arguments can be cogent. This gives us a prima facie reason for thinking that we should be able to utilize absence of evidence arguments in the animal consciousness debate. In Sect. 2, we show how these epistemic conditions could be met in practice, drawing on extant investigations in comparative neurobiology. In Sect. 3, we draw out the implications of this investigation for the widely held assumption that psychological states like pain are multiply realizable. We thus seek to add to the voices of dissent advocating a revision of the multiple realizability assumption. In Sect. 4, we develop the idea of a common neural architecture across extant species that could serve as a minimal condition for a certain kind of conscious experience. We advocate theorizing about multiple realizability in observance of the foundational axiom of biology that structure-determines-function. Paying close attention to this axiom forces us to revise whether ‘realization’ is the right way to characterise the relationship between entities at different levels of analysis. Overall, if we are right, an absence of evidence of a certain kind of neural architecture could well be evidence of an absence of consciousness.

Let us be clear about our intentions: We do not assume that the example we offer below of an absence of evidence argument—‘FISH’—is sound; only that it is a kind of argument that could be cogent if the right epistemic conditions were met. We defend only the right to use the form of argument instantiated in FISH when reasoning about a given species’ capacity for conscious experience. Whether arguments of this form yield conclusions that are likely to be true is an empirical matter, and while we cite here some evidence for (and against!) FISH, this is only to provide a sense of what it would be like to go through the motions of evaluating an argument like FISH on the basis of scientific evidence (or a lack of it), rather than simply dismissing the argument for ideological reasons. The point of the exercise is to establish the acceptability of the form of the argument, not, on this occasion, its content.

2 Absence of evidence arguments and the science of consciousness

FISH: Animals that feel pain possess neural circuitry enabling them to execute the neural computations that lead to pain. There is no evidence that fish possess such circuitry. Hence, fish do not feel pain.

FISH is an instance of a form of argument that appears in neuroscientific debates about the capacity for conscious experience across phylogenetically-diverse species. In practice, the neuroscientist purports to have discovered a kind of circuitry that they take to be strongly correlated with or explanatory of a certain kind of conscious experience in humans, positively verified through first-person reports, and then reasons inductively to other species’ capacity for that experience based on the presence or absence of homologies of that circuitry. This is standard evolutionary biological practice (Bock and Von Wahlert 1965; Wake 1992), and yet while many are inclined to accept positive evidence as indicative of the capacity for consciousness, few are inclined to accept negative findings as evidence of a lack of consciousness. FISH is an ‘absence of evidence’ argument. It reasons from negative findings to the conclusion that fish are incapable of feeling pain. Why should there be such asymmetry in the way positive and negative findings are regarded?

Among some researchers, the battle cry ‘Absence of evidence is not evidence of absence!’ has taken on the status of an axiomatic truth. Brown (2016b, p.2) writes: “It is true that “human-like” nociceptors have not been identified in elasmobranchs yet, but that is not to say that they do not exist. Science 101: Absence of evidence is not evidence of absence.” In an article entitled “Assessing negative and positive evidence for animal pain”, Elwood (2017) similarly objects to anyone’s relying on negative evidence to conclude that a species is incapable of pain. While he may be correct that negative evidence resulting from an insufficiently extensive search is not reliable, his own reliance on the slogan “absence of evidence is not evidence of absence” (2017, p. 1) goes beyond this otherwise minor methodological cautionary tale. Manzotti (2018, p. 2) offers a stronger blanket prohibition of the use of absence of evidence arguments, stating that “there is no evidence that an animal that shows pain-avoidance does not feel pain, no matter what neural structure is involved”, meaning by “no evidence” here an absence of evidence and assuming that no such negative findings could ever be conclusive of an absence of pain. Such aversion to absence of evidence reasoning is not restricted, moreover, to individual researchers. A report of the National Research Council of Canada, (2009, p. 23) states that:

Although definitive evidence is often unavailable, this report does not treat the absence of evidence as evidence of absence. Instead, the consensus of the committee is that all vertebrates should be considered capable of experiencing pain.Footnote 2

Our task here is try to move this debate beyond its uncritical reliance on this slogan. Grounds for assuming that the slogan is true are rarely cited, but we can make some educated guesses. Some might fear that absence of evidence reasoning is logically illicit. Those operating within a Bayesian framework might, for example, find it questionable to update beliefs on the basis of an absence of evidence. Some have thought that an absence of evidence is at best “weak” (Sober 2009) or “subjective” (Strevens 2009) evidence. It is true that an absence of evidence becomes weaker the more that finding evidence given the truth of a hypothesis is left to chance. If, for example, the probability of finding an intermediate fossil (Sober’s interest case) is low, then the absence of evidence will not warrant believing that intermediate fossils do not exist or in any way vindicate Creationism. Generally, however, there is no barrier within a probability framework to updating on the basis of an absence of evidence any more than on the basis of positive findings. An absence of evidence can form part of the ‘total evidence’ to be considered. It can tip the balance in favour of one hypothesis over a competing one, especially if the probability of finding evidence if a certain hypothesis is true is reasonably high and the evidence is not forthcoming. And an absence of evidence itself could be something that stands in need of a scientific explanation. Indeed, it could be the very thing that calls for a replacement or revision of an accepted theory or hypothesis. An absence of evidence should not, in other words, be thought of as necessarily epistemically inert.

FISH might appear logically spurious for other reasons, however. Superficially, it looks like an argumentum ad ignorantiam (argument from ignorance), which, in classical logic textbooks, would standardly be classified as a fallacy. Arguments from ignorance are bad because for any proposition, p, not knowing that p is true is not a reason for inferring that not-p is true (Walton 1996, p. 143). But FISH need not be so flimsy. If FISH were based on a complete lack of information about the neuroanatomy of fish, then it would be a bad argument. In that situation, negative results would at best provide a reason to suspend judgement or prompt further investigation, but never to draw a definitive conclusion in the way that FISH does. Not accepting the proposition that fish feel pain on the basis of an absence of evidence is not the same thing as accepting the proposition that fish do not feel pain, and, you might think, we should be careful to avoid conflating the two.

The central premise of Douglas Walton’s (1996) book, Arguments from Ignorance, is that arguments from ignorance are presumptive arguments—arguments that may be fallacious in certain epistemic contexts, but cogent in others. Consider the following two acceptable reasoning scenarios:

KEYS: You are looking for your keys and think you might have left them on the bookshelf, but when you look, you don’t see them. A natural conclusion to draw is that the keys are not there.

VULCAN: You’re an early twentieth century astrophysicist seeking to test Le Verrier’s hypothesis that there is a planet (Vulcan) causing perturbations in Mercury’s orbit. You keep looking but find nothing. You conclude that Vulcan does not exist. (If you happen to be Albert Einstein, you devise an alternative explanation—General Relativity theory—live long, and prosper.)

These are absence of evidence arguments, and yet are cogent.Footnote 3

Ultimately, what all these considerations suggest is that absence of evidence arguments are neither cogent nor fallacious in themselves but only in relation to their contexts of use. In KEYS and VULCAN, what makes the arguments cogent is that in each context the agent could reasonably have expected to find evidence if the hypothesis were true, and yet did not. This is sometimes referred to as the epistemic closure principle. Put simply: ‘If it were true, I would know it’ (de Cornulier 1988, p. 182; Walton 1996, p. 147).Footnote 4

Armed with the epistemic closure principle, we can see an obvious path to making FISH not just a rationally acceptable form of argument but one that is, even by the standards of classical logic, valid. Indeed, it would be difficult to dismiss FISH as illogical for the simple reason that it can be easily transformed into a logically valid form of argument, modus tollens:

Modus Tollens:

(P1) If p, then q.

(P2) Not-q

(C) Therefore, not-p.

Adding a tailored version of the epistemic closure principle (P1* below), we can transform FISH into FISH*, a modus tollens argument:

FISH*:

(P1*) If the hypothesis that fish feel pain were true, diligent investigation would identify neural circuitry enabling fish to execute the neural computations that lead to pain.

(P2*) Diligent investigation does not identify neural circuitry enabling fish to execute the neural computations that lead to pain.

(C*) Therefore, the hypothesis that fish feel pain is not true.

We have chosen the non-probabilistic form (FISH*) for ease of exposition. Because FISH* is a presumptive, scientific argument, however, it might be better to construe it probabilistically, as in FISH**:

FISH**:

(P1) If the hypothesis that fish feel pain is true, then it is highly probable that diligent investigation would identify neural circuitry enabling fish to execute the neural computations that lead to pain.

(P2) We are (reasonably) certain that diligent investigation does not identify neural circuitry enabling fish to execute the neural computations that lead to pain.

(C) Therefore, it is highly probable that the hypothesis that fish feel pain is not true.Footnote 5

As Sober (2009, p. 64) observes, one problem with adopting this kind of strategy to justify absence of evidence reasoning is that in some cases the epistemic closure principle is true; in other cases, not. Sober’s counterexample involves the following scenario. Suppose it happens to be true that there are storms now on Jupiter (but we don’t know it). The likelihood of finding evidence of storms on Jupiter right now is, however, low. In this case, we could not reasonably expect to find evidence right now even if the hypothesis were true, and so the fact that we do not have evidence in support of the hypothesis is no reason to reject it. (Of course, it is no reason to accept it either!)

This tells us two things. As Sober (ibid) points out, an absence of evidence doesn’t logically entail an absence—it is evidence of absence except when it isn’t! But it also tells us something else about the epistemic conditions under which an absence of evidence is evidence of absence. Besides closure, we need sufficient “epistemic coverage” (Goldberg 2011; Pedersen and Kallestrup 2013). KEYS and VULCAN work because in both contexts our investigations are extensive and the contexts are restricted enough for us to draw a ‘positive- negative’ conclusion (i.e. that what we are looking for is not there). Epistemic coverage is sometimes construed in reliabilist terms—i.e., agents have good epistemic coverage of the evidence when they are connected to a source that feeds them reliable information about the domain and where they are receptive to being “hooked in” to information from that domain (Pedersen and Kallestrup 2013, p. 2576). But whether one needs to meet reliabilist standards in order to claim epistemic coverage is a matter which need not concern us here. The important point for our purposes is that in order to judge the cogency of an absence of evidence argument, there must be some mechanism available for evaluating the degree of epistemic coverage. We cannot assume right now to have good epistemic coverage of the surface of Jupiter, but the neuroanatomy of fish is a different kettle of fish. Whether FISH* is sound or not would depend on whether the epistemic closure principle (P1*) and the epistemic coverage claim (P2*) are true. So far though, nothing about the form of the argument prevents it from being an argument we could rationally accept.

3 Absence of evidence of neural circuitry in fish

What we learned from the previous section is that where absence of evidence arguments work, they do so on the basis of a mix of negative findings and positive knowledge (Copi 1982; p. 102; Walton 1996, pp. 135–138). The epistemic closure principle is a positive claim, which, when supported by a true premise asserting an absence of evidence, provides grounds for inferring that the hypothesis that something exists is false. What would it be like to have such grounds for accepting the conclusion of an absence of evidence argument in relation to the science of consciousness?

Let’s go through the exercise of evaluating FISH (or its more formal variations, FISH* or FISH**). The first task would be to establish that the epistemic closure principle (P1*) in FISH* is true.

We start by defining pain as the conscious (felt) neural processing of noxious stimuli, in line with the standard neuroscientific definition (Anand and Craig 1996; Williams and Craig 2016) that emerged from the experimental work of C.S. Sherrington in the late 1800s (Woodworth and Sherrington 1904). This definition implies a distinction between pain and nociception, the non-conscious (unfelt) processing of noxious stimuli and associated nocifensive behaviours (e.g. the flexion withdrawal reflex). Given the overwhelming evidence in favour of the dissociation between pain and nociception,Footnote 6 and their being seated in different neural circuitry (Sherrington 1947; Geuter et al. 2020; Tracey 2005), we would expect this distinction to be uncontroversial. But it is not, creating enormous confusion.

It is the lack of attention to this distinction that lends more credence than deserved to the alternative functionalist solutions outlined earlier. Functionalists typically allow that fish feel pain even though they lack the neural circuitry evident in humans and other mammals in one of three ways: (1) fish have their own kind of pain and ‘pain’ is ambiguous (Huntingford et al. 2006; Sneddon 2015; Seth 2016; Manzotti 2018); (2) pain in fish is the same kind of pain as that experienced by humans but has a different neural realizer (Lewis 1972, 1980); or (3) ‘pain’ refers to a relational, higher-order functional role property shared by fish and humans but executed by different mechanisms in each species. None of these options does much to preserve a robust distinction between pain and nociception. (1) makes it hard to see why any kind of nocifensive behaviour, including that exhibited by single-celled organisms, would not count as a ‘sort of pain’.Footnote 7 (2), the “Ramsey-Lewis” solution, prioritizes behavioural criteria resulting in a kind of ‘black box functionalism’ that renders the approach insufficient to dismiss the null hypothesis. There is nothing, for example, in Lewis’ story of Martian pain, based at it is on observations of behaviour, that favours the hypothesis that the Martian is in pain over the null hypothesis that the Martian merely has nociception. (3) assumes that the functional role is itself causally explanatory, leaving it open to the objection that what is really doing the causal work is the particular neural realizer (Block 1990, p. 155; Kim 1998, p. 51). Since (3) places no constraints on what kind of thing is required to perform the specific function of pain, it too is insufficient to maintain the distinction between pain and nociception.

The costs of conflating pain and nociception are high. Consider the first option (1)—that ‘pain’ is ambiguous. Let’s call the kind of pain fish have on this view ‘painǂ’ to distinguish it from human pain. The problem is that we do not know what painǂ is like and so do not know what, beyond crude similarities in behaviour, to look for as the structure that explains painǂ. We could not assume that what-it-is-like (Nagel 1974) for fish to feel painǂ, the “quale” or “phenomenal” aspect (Block 1980; Jackson 1982) of their conscious experience, would be the same as pain experienced by humans. If we could, ‘pain’ wouldn’t be ambiguous. Strictly speaking, it is a category mistake to say that painǂ is painful, since the predicate ‘is painful’ inherits its meaning from ‘pain’ not ‘painǂ’. But now it is not clear what we are talking about when we refer to ‘fish painǂ’ or how we could establish that painǂ is even a feeling rather than just another name for nociception. If the criteria for both are the same, any painǂ hypothesis would be unverifiable. We could instead regard pain and painǂ as sharing the same epiphenomenal, emergent property, which, as Kim (1998 p. 18) notes, alleviates the problem by allowing type-identical epiphenomenal properties across different realizers with different causal powers. But this seems to us to make the question of realization moot, since what makes each a kind of pain is the shared epiphenomenal property and not what physically causes or constitutes pain. This has to be the counsel of despair.

(2) and (3) do not fair much better. The problem with the Ramseyfication in (2) or focusing on what the psychological state does (observationally) at the expense of what it is, as in (3), is that we lose sight of the intermediate, computational level of analysis. This level is couched in irreducibly theoretical language and involves abstracting from descriptions of lower-level constitutive properties. Without access to this level, it will always be tempting to obfuscate the distinction between pain and nociception, rendering any hypothesis that a creature has pain unverifiable as the criteria for being in pain become indistinguishable from the criteria for exhibiting nociception.

An alternative to all three options outlined above is to suppose that ‘pain’ is unambiguous and that it indeed refers to a “nomically projectable kind” (Kim 1998, p. 109). On this approach, we would seek to identify the neural architecture that explains why an organism has the phenomenal experience it does and whatever functional role that experience typically (but perhaps not universally) performs. Such an approach would yield conditionals of the form ‘If x is in pain, then x has neural circuitry, N’. The consequent would describe a necessary condition for pain that could serve to ground predictions of different species’ capacity for pain. If fish and humans both experience pain (i.e. both feel the same negative affect valence), then we could expect that discoveries about how pain is manifested in humans would inform decisions about whether fish feel pain. If we found similar circuitry, we could then explain why both fish and humans writhe about when administered electric shocks and how they use their pain to modify their behaviour so as to avoid noxious stimuli. However this investigation pans out, we could assume all the while that pain constitutes a single, unified, scientific kind while preserving the distinction between pain and nociception, on the one hand, and avoiding epiphenomenalism, on the other. Happy land!

This outcome would find considerable support within the community of neuroscientists claiming to have found evidence of nociception in fish but no pain (e.g., Adamo 2018; Derbyshire 2016; Favela 2017; Gamez 2018; Gerlai 2017; Gutfreund 2017; Rose et al. 2014; Stauffer 2017; Key 2015, 2016; Key and Brown 2018). It is well known that pain in humans is dependent on neural activity in the cerebral cortex of the dorsal telencephalon (Key and Brown 2018). On this basis, we could reasonably expect to find structurally homologous brain regions in fish that are also necessary for pain. However, surgical removal of the whole dorsal telencephalon (called the ‘pallium’) in fish fails to abolish the avoidance behavioural responses to noxious stimuli (in this case, electric shock; Portavella et al. 2004). Whatever is happening in the dorsal telencephalon of fish then is not responsible for its nocifensive behaviour and thus does not support the presence of pain. The absence of any signature pain-relevant neural architecture in fish would lend credence to the second premise of FISH*, P2*. If the neuroanatomy of fish were well-mapped, (as we have argued to be the case; Key 2015, 2016), indicating good epistemic coverage of the domain, by modus tollens, the conclusion that fish lack the capacity to feel pain would go through.

This is what it would be like to conclude that a species is not capable of pain on the basis of an absence of neurological evidence and plausible functionalist assumptions. But, of course, the elephant in the room is the prospect of multiple realizability, which we have conveniently so far ignored. The multiple realizability principle can always be trotted out to defeat an absence of evidence argument. Whether it should is another matter.

4 Addressing multiple realizability at the correct level of abstraction

It is important not to underestimate the role that the multiple realizability assumption plays in blocking absence of evidence arguments in the animal consciousness debate. When challenged to explain how a creature is able to feel pain when it lacks the neural architecture responsible for pain in humans, scientists and philosophers alike often defer to the concept of multiple realizability (Demski 2013; Sneddon 2015; Broom 2016; Brown 2016a; Merker 2016; Seth 2016; Striedter 2016; Woodruff 2017, 2018; Michel 2019). Pain, it is said, is a function and functions are essentially multiply realizable (Putnam 1967). Just as chairs involve many different kinds of structures and materials, so too it is argued that pain is realized in different kinds of neural circuitry, neural configurations, or the green goo of the inflated feet of Martians (Lewis 1980). There is no monolithic attachment to multiple realizability, however. Polger (2004, p.11) points out that we either have to reject multiple realizability or deny that we have psychological experiences in common with other species. He speculates that the identity theorist might indeed find commonalities between all critters who experience pain, noting, however, its status as an “ambitious hypothesis” (Polger 2015, p. 874). When Haikonen (2016) asks whether we should not just assume that the neural realizer for pain is a “simple one” that could fit into either “the diminutive fish brain or a complex one calling for larger cortical resources?”, he is swallowing the “ambitious hypothesis” hook, line and sinker. No one has any idea what this “simple realizer” might be.

One source of support for the multiple realizability assumption are folk arguments from analogy—jet planes fly without propellers (Ng 2016), crickets hear without human ears (Dinets 2016), teeth and gastric acid perform the “same function” (Jones 2016), and bats fly without feathers (Manzotti 2016). Sameness of function is possible in each of these examples only because of the indeterminate way in which the function is defined. For the concept of flying, for example, it is possible to imagine many different objects ‘flying’, including tennis balls, arrows, rockets and birds. If, instead, we are more discerning and consider flying at a more discriminatory and explanatory level—such as one involving aerodynamics—the argument for multiple realizability soon breaks down. At this level, the forces of lift, gravity, thrust and drag are markedly different between disparate structures. Honing in on a more determinate function—flying based on the capacity for aerodynamic lift—structures begin to exhibit more commonalities than differences. Despite variation in the shape and form of wings or wing-like appendages, all species capable of performing this function exhibit a common wing design that explains their capacity for aerodynamic lift (Lindhe Norberg 2002). There are clear explanatory benefits of this more fine-grained approach to functional categorization as it allows us to see what it is about the structure of the wing that explains its capacity for aerodynamic lift. Of course, one can swing too far in the opposite direction—categorising the function of being consciously aware of a red light in the left side of the visual field as being distinct from that of being aware of the same light coming from the right. Choosing the appropriate level of analysis and description for both structure and function is thus critical to the evaluation of whether there is multiple realizability of a function or not. When either the structural level is too granular or the functional level too indiscriminate, multiple realizability is assured—and its argumentative value is exhausted.

To get the structure and the function, and indeed, the level of analysis, aligned is not easy. Citing numerous studies revealing common neural substrates for visual processing in the brains of humans and monkeys, Bechtel and Mundale (1999) reject multiple realizability. But, of course, within any brain region there may be much variation. Beyond establishing correlations between cognitive functions and brain regions, neuroscience endeavours to decipher what it is about the structures that distinct species have in common that explains their ability to perform the cognitive function under examination. This imposes a tighter constraint on the attribution of a cognitive function than establishing correlations to brain regions. Specifically, neuroscience looks beyond homologies of brain regions to homologies of information processing tasks. This requires attention to the neural architecture or precise circuitry that makes possible the performance of a function.

This insistence on paying attention to intermediate levels of analysis and structure is not new (Nelson 1969; Piccinini and Maley 2014; Polger and Shapiro 2016; Bickle 2003; Wimsatt 2007; Craver 2007). We agree with Piccinini and Maley (2014, p. 131) that one needs to isolate the “relevant causal mechanisms” in judging whether there is multiple realization or not and with Polger and Shapiro (2016, p.98) that “sameness of function owes to sameness of structure.” In regard to neuroscience, however, a common problem is that theorists tend to operate at too high a level of generality. Polger and Shapiro, for example, never interrogate what kinds of circuitry underlie and explain the functions of memory, pain, trichromacy, etc., that they describe. In the end, they too fall back on behavioural criteria in making the call about species with neural circuitry different from that presented by humans—e.g. when considering whether octopi feel pain, they write that “the right conclusion is that…their pain experiences differ from human experiences” (2016, p. 113; cf: Key and Brown 2018). The result is some unfortunate mixed messaging. On the other side, Gillett (2002; 2007, 2010) and Aizawa and Gillett (2009), defend a “dimensioned” framework for attributions of multiple realizability that attempts to move away from reliance on intuitions. Multiple realizability applies only in contexts of hierarchically and compositionally organised entities, their properties (especially, “causal powers”) and relations. To avoid triviality, examples of multiple realizability must be drawn from the same level of analysis. But in the end, multiple realizability is guaranteed since one will always find some differences among the realizers. As Piccinini and Maley (2014, p. 128) point out, not every difference in the realizer interferes with its capacity to realize a given property, or hence, counts as a case of multiple realization.Footnote 8 In neuroscience, different protein structures and different sizes and numbers of subcellular compartments like dendritic spines that in turn create fluctuations in the excitatory electrical signalling of neurons, show how neurons multiply realize “many higher-level psychological properties” (Aizawa and Gillett 2009, pp. 200–201). This, we agree, is multiple realizability; it is just that it does not exclude non-multiple realizability at a level which is more explanatorily relevant.

Assessing multiple realizability in neuroscience requires paying attention to the correct level of abstraction. Structural differences at the levels of gene expression, subcellular compartments (e.g. synapse density, dendritic tree branching) and neuron population number are too granular, whereas the mere presence of a forebrain is too gross for drawing meaningful conclusions. Piccinini and Maley (2014, p. 144) touch on this idea with respect to action potentials in the nervous system, acknowledging that multiple realizability occurs only at some levels of organisation. They claim that a neuron’s capacity to fire an action potential is not multiply realized even though the ion channels producing the action potential could be. That is, action potentials could only be realised by the flow of ions across the plasma membrane, and while there may be various kinds of ion channels with relevantly different molecular properties, each of these kinds could still contribute to generating ion flow. We highlight this problem by examining a simple neural circuit—the central pattern generator (CPG)—that drives rhythmic locomotion in bilaterally symmetrical animals. The function of the CPG is to generate rhythmical bursts of action potentials that are causally responsible for cyclical movements of either body segments or limbs on either side of the body, as in stepping or swimming. The structure and function of CPGs have been well studied in Mollusca and, in particular, in two closely related nudibranchs—Melibe leonina and Dendronotus iris. The gross swim behaviour of these two species appear very similar and is described as stereotyped rhythmic left–right movements consisting of simple C-shaped or lateral bends of the body occurring at a frequency of three cycles per second (Sakurai et al. 2011). If we assume that this function is identical between species, nudibranchs provide a convenient model for assessing whether the CPGs have similar neural architectures, especially since these closely related species have many anatomically homologous neurons.

The CPGs in these two species share similarities but are not identical (Sakurai and Katz 2019; Fig. 1). In Dendronotus, there are two swim interneurons—Si2 and Si3—that are present on each side of the body that constitute the CPG. In contrast, Melibe has the same two homologous neurons as well as an additional two pairs—Si1 and Si4—that form the CPG. In addition to the extra swim interneurons and their interconnectivity, the synaptic connectivity between the common neurons—Si2 and Si3—is disparate.

Fig. 1
figure 1

Simplified circuitry diagrams of central pattern generators in a Dendronotus and b Melibe. Circles represent neurons which are referred to as swim interneurons (Si) 1–4 on either the left (L) or right (R) sides of the midline. Red and green lines represent axons with either inhibitory or excitatory monosynaptic chemical synapses, respectively. (Color figure online)

Neither neuron number, synapse number nor specific axonal interconnectivity appear to be suitable structural parameters for characterizing CPG between these species. However, on closer inspection, it is apparent that both species share the common structural feature of reciprocal crossed inhibition of interneurons. The projectable kind here is determined not by the number of individual neurons, which varies (and so there is multiple realizability at this level), but by the simplified structure, which describes the minimal conditions for left–right locomotion. Reciprocal inhibition has long been considered to be essential for the normal execution of locomotor behaviours in bilaterally symmetrical animals. This can be demonstrated by experimentally inducing unilateral blockage of crossed inhibitory neurons (Moult et al. 2013). When ipsilateral inhibitory neurons are prevented from inhibiting their contralateral counterparts, there is loss of rhythmic locomotor activity. Thus, both species possess neurons that project to the opposite side of the body to inhibit the contralateral neurons in the CPG. This neural connectivity is essential to ensure that one side of the body is relaxed while the other side contracts to perform the rhythmical locomotion—in this way, the structure of the circuitry explains its functional role. To date, all species exhibiting rhythmical locomotion employ crossed inhibition using crossed axons that lead to contralateral inhibition of motoneurons in the context of the CPG (Katz 2016; Kiehn 2016; Harris-Warrick and Ramirez 2017). The presence of neurons performing this neurocomputation becomes a useful criterion for determining whether an animal has the potential for performing rhythmical locomotion. At the level of analysis involving the necessary neurocomputation of crossed inhibition—the explanatory level—multiple realizability is no longer applicable. This strongly suggests that multiple realizability does not automatically apply at all levels of analysis, and therefore must not be assumed uncritically in rejecting absence of evidence arguments like FISH*.

While we have argued here that choosing the correct level of abstraction is critical for determining whether a function is multiply realized, we have not, as yet, provided any clue as to how one goes about identifying this level—apart from suggesting that the analysis should be neither too granular nor too obtuse.Footnote 9 Are there any rules or heuristics that could inform this choice? In this matter, we can be guided by the earlier work of Marr (1982) on the visual system. Marr reasoned that to understand the function of a nervous system, one has to address three levels of analysis: Level 1—the computational level that describes the goal of the system; Level 2—the algorithmic level that details the process of performing specific computations to achieve the system’s goal; and Level 3—the implementation level that contains the basic components that execute the algorithm. Marr realized that a computational goal could possibly be achieved via different algorithms in the same system and that the same algorithm may be physically realised by different circuits in different systems. Thus, the algorithmic level is probably not the best initial choice for a level of abstraction. By ‘computational goal’, Marr was referring not to the overall goal of visual perception, but rather to specific computational tasks such as stereopsis and motion detection that are necessary for normal human vision. Perturbations to relevant brain regions executing motion detection lead to perturbed visual perception (Zeki and Bartels 1998). Visual motion detection is also an essential computational goal in many visual systems and is realized by discrete neural modules in widely diverse species such as humans, insects and crustaceans (Strausfeld 2005; Borst and Helmstaedter 2015; Strother et al. 2017). In the absence of these modules, motion detection and associated behaviours are compromised. High level computational goals (i.e. specific information processing tasks), therefore, appear to be an appropriate starting point for selecting the correct level of abstraction.

While computational goals can be elucidated by theoretical approaches, their necessity must subsequently be confirmed through experimental manipulation and/or clinical case reporting—as illustrated above for bilateral symmetrical locomotion and motion detection. A number of specific information processing tasks have been shown to be essential for conscious processing of noxious stimuli (Key 2015, 2016). This does not prescribe how conscious awareness is generated by these computational goals, but only that these processes form the essential building blocks for pain. For example, pain depends at least on parcellation of neural tissue into modules or domains that perform pain-dependent computations accounting for the intensity (Porro et al. 1998), quality (Maihöfner and Kaltenhäuser 2009), unpleasantness (Rainville et al. 1997), motivational value (Wager and Barrett 2017), and localisation (Baumgärtner et al. 2006) of a noxious stimulus. These regions have been shown to have strong reciprocal interactions involving inhibitory and excitatory interconnections to make this possible. Supposing that the presence of such discrete brain regions and their feedforward and feedback interconnections are essential for pain, it would be reasonable to investigate whether these neural regions and circuits were realised in the fish brain in any form (i.e. without being restricted by phylogenetic homologies of brain regions or without concern for low level structural features such as local microcircuitry, neuronal population size and subcellular features). There is evidence that most of the forebrain in fish is not necessary for behavioural responses to noxious stimuli, and that the sub-forebrain regions driving this behaviour (as determined by recent whole brain imaging and ablation studies in zebrafish; Lovett-Barron et al. 2019; Wee et al. 2019) lack the neural architectures considered necessary to execute the prerequisite information-processing tasks for pain (Key and Brown 2018).

By focusing on the necessity of specific information-processing tasks for perception rather than on multiple ways of organising processing steps within an algorithm, we can see how the debate could move beyond the simple assumption of multiple realization. While we may continue to debate whether a particular information-processing task is necessary or not for pain, this does not negate the idea that there are certain core computations that are essential for pain. It merely means that those neural computations become a matter for empirical verification. As recently demonstrated, this approach is providing important insights into the evolutionary origins of phenomenal consciousness (Key and Brown 2018).

5 Beyond multiple realizability

The aim in the previous section was to defend the legitimacy of absence of evidence arguments in comparative neurobiology by arguing against the multiple-realizability-at-every-level assumption. We recognise, however, that many theorists would die in a ditch before jettisoning multiple realizability. Some, (e.g., Allen 2013; Jones 2016), go so far as to claim that pain might be something better investigated at the “behavioural and cognitive levels”:

The possibility of conscious experience at the behavioural and cognitive levels despite morphological and anatomical differences at the neurological level makes fish an enormously interesting testing ground for ideas about multiple realizability of cognition. (Allen 2013, p. 36Footnote 10; cited in Jones 2016)

This is a strange call to arms. It entails that in investigating whether a species feels pain, we should ignore state-of-the-art neural recording and imaging, molecular genetic manipulations allowing for the dissection and identification of distinct neural circuits, electrophysiological and advanced microscopy techniques involving neural circuit manipulation and mapping—just to maintain an attachment to multiple realizability. How might we better align our philosophical intuitions with such state-of-the-art advances in neuroscience?

One diagnosis for this misalignment is that theorists have become overly fixated on the realization relation. Because they understand the realizer relation compositionally, Aizawa and Gillett (2009, p. 194) insist that it is transitive. If x realizes y and y realizes z, then x realizes z. The same, they note, is true if ‘realization’ is understood in terms of a determination relation (Ibid, n. 16). Fair enough. It is this transitivity, however, that lends support to the idea that what is causally explanatory, ultimately, is what lies at the bottom of the chain of realizations. When there is no unity at the x level, it can seem as if the only recourse for a science of conscious experience, the z level, is not neuroscience but a science that sticks to the “behavioural or conceptual” levels. Aside from the danger that this outcome forfeits the nociception/pain distinction, it is based on a further assumption that ‘realization’ is the right term to characterise the relationship between different levels of scientific explanation. This is far from obvious. Although Aizawa and Gillett’s paper purports to be about ‘realization’, it is really about ‘compositionality’, which they freely admit is ambiguous. Following Simons (1987), Aizawa and Gillett (2009), p. 182, n.2) use ‘composition’ to encompass a number of relational concepts featured in mechanistic explanations, including ‘realization’ (of properties), ‘constitution’ (of individuals), ‘implementation’ (of processes), and ‘comprising’ relations (e.g., of one power by another). Realization is only one relation among many, not all of which are transitive.

Consider the analysis of rhythmic locomotion offered above. The CPG is realized by different sets of neurons in different species, and it implements the neurocomputations necessary for rhythmic locomotion, but it’s a mistake to say that the set of neurons in each case implements the neurocomputations for locomotion—at least they won’t in any old configuration! While the CPG is necessary for rhythmic locomotion, it’s wrong to say that a determinate set of neurons is necessary, because the sets of neurons constituting the CPG of Dendronotus and Melibe, while sufficient in each case for the existence of a functioning CPG, are not necessary for rhythmic locomotion. Only the simplified circuit is necessary. Formally, that x (set of neurons) is sufficient for y (circuit) and y is necessary for z (rhythmic locomotion), does not entail that x is necessary for z. In this context, transitivity fails and the explanatory feature is not the one doing the ‘realizing’ as such, but the one ‘implementing’ the neurocomputation. By parity of reasoning, techniques enabling us to better understand the neural circuitry performing a psychological function like pain would be more than fully justified.

There is no need, therefore, to resort to “local” or species-specific reductions (cf., Kim 1998, pp. 19–20) either, and as there can be intraspecific variation, this will not help. Nor is it helpful to suppose that because the notion of ‘structure’ used in neuroscience abstracts from differences that obtain at the neuronal level, it denotes an “ideal type” (Klein 2008). If there is a common neural architecture across distinct species that unifies the type we refer to as ‘pain’, then while this common architecture is defined in abstraction from what happens at the level of individual neurons, it does not follow that the circuitry so defined is an abstract entity. It would involve actual neural circuitry that does real work controlling the survival-enhancing behaviour of the organism that possesses it. And were we to lack evidence of that circuitry in a given species, we would be able to reasonably infer that the capacity for pain does not there exist.

6 Conclusion

We stress again that we have not tried to establish here that the conclusion of FISH is true, but only to establish the kinds of evidential and epistemic conditions that if met, would allow the argument to be cogent. If the right epistemic conditions are met, an absence of evidence argument shifts the burden of proof onto those asserting that a given species is capable of feeling pain to provide evidence that disconfirms the null hypothesis that it is not. FISH* uses modus tollens (FISH**, probabilistic modus tollens) to conclude that it is not likely that fish have the capacity to feel pain.Footnote 11 Attention to the structure that explains the function of pain in humans is what provides the foundation for the projectability of neurological criteria for pain onto other species. This ‘partial reduction’ should be enough; no ‘global reduction’ specifying necessary and sufficient conditions for pain would be necessary or helpful for either explanatory or predictive purposes. Whether FISH or its variants are indeed sound is, however, a matter for empirical interrogation. We have not endeavored, therefore, to show how ‘the explanatory gap’ in the science of pain can be closed, but simply to move the debate about the prospects for a unified science of pain a little closer towards consensus.Footnote 12