Introduction

Calo (2016, p. 225) has tracked robots through American case law. He identified how in the judicial system, imaginary robots, as legal objects and curial metaphors, have appeared mechanistic, predictable, programmable machines. Calo raises a significant question: how can law think of robots elsewise?Footnote 1 He particularly identifies the idea of ‘emergent’ behavior where the frayed connection between human command and robot doing is problematic for the judicial imaginary (Calo 2016, p. 227). This article asks a similar question. However, its concern is not with the specific ‘mechanical men’ phantasies of judges but rather the scholarship to which Calo is a well-recognised contributor (Calo 2011, 2015).

Law and technology scholarship interacts with a speculative jurisdiction (Tranter 2011b). Hopes and fears of technological futures become translated into legal analysis. Rules and principles are assessed for adequacy, legal vacuums identified (Beebe 1999) and reforms suggested. There are two essential features to this ‘enterprise’ (Tranter 2011a). The first is an underlying positivist account of law. Law is a tool through which the present can safeguard the future (Tranter 2011a).Footnote 2 The second is that the legal scholar undertakes a process of imagining. It is how technologies might affect the future that is within the frame. To do this the law and technology scholar draws upon an ambit science fiction infused cultural imaginary of technological futures: This broader cultural imaginary establishes, circulates, and makes popular images, memes, and narratives that animate and give structure to the hopes and fears of technological change (Tranter 2011b). The specific legal analysis of law and technology scholarship often manifests and depends on the cultural imaginary technological change. For example first generation space law literature from the late-1950s-early-1960s embodied and projected the fantasies of 1930-40s space opera (Tranter 2011b, pp. 821-6) or the legal literature considering legal and regulatory responses to mammalian cloning in the late-1990s-early-2000s drew upon the dreams and horrors of cloning nurtured and made popular by science fiction’s ‘clone canon’ (Tranter 2018b, pp. 22 − 7).

Drawing upon this homology of the legal and cultural imageries this article argues that when lawyers think of robots in the 2010 and 2020 s, there is a foundational ‘human-centrism’ inherited from a cultural imaginary that has been nurtured by ‘mainstream’ twentieth-century science fiction. Importantly, the elements of this human-centrism—human futures, material bodies, and separation—are disrupting clearer thinking about the messy hybrid and fluxing future of life and living with the digital. This argument is in three parts.

The first part argues that mainstream twentieth-century science fiction has nurtured a cultural imaginary concerning artificial beings. Core to this imaginary is the ‘human paradigm.’ It begins by remembering the two foundational twentieth-century robot texts: Karel Čapek’s 1920 play R.U.R. (Rossum’s Universal Robots) (Čapek and Čapek 1961) and Isaac Asimov’s robot short stories and robot detective novels from 1940s to 60s (Asimov 1960, 1967, 1969, 1983, 2008). Čapek’s robots go on to obliterate and replace humanity, while Asimov’s serve (mostly) faithfully. Notwithstanding significant differences, together, these texts establish the core science fictional tropes concerning robots—anxiety about human futures, robots as located in discrete bodies, and a distinction between humans and robots—that across subsequent decades became reiterated and replayed particularly by popular screened science fiction. In short, science fiction might imagine robots but only sees human life as it has been known.

The second part argues that legal thinking about robots can be seen as grounded on this human paradigm. It takes as its sample contributions that expressly orientate the analysis on robot law scholarship. It will be shown that robot law scholarship notwithstanding the sophistication of their legal analysis, manifest an iteration of the human paradigm. Informing each are foundational concerns with human futures, that robots are identifiable distinct material object, and a quarantining between robots and humans as the users, consumers, or ‘victims’ of robots.

The third part shows why the manifestation of science fiction’s human paradigm is problematic. This is something more than the ‘android fantasy’ that Neil Richards and William Smart (2016) discuss. It disrupts because it obscures a clearer appreciation of the forms and trajectories of living in and with the digital. First, its emphasis on the specificity of automation located within a singular material entity in a specific time and place, seems less than certain. While a trait within Asimov was robots ascending to the godhead, actual robots, and automated digital technologies tend to be less constrained by a material body than imagined in science fiction but fall short of transcendent omnipresence. Second, it denies a vision of the posthuman, of the fusion and hybridity between humans and autonomous digital technologies generating new and different makers and subjects of law. Without clear mapping of the complex, fluxing hybridity of humans and the digital, current robot law scholarship speaks past its subject. It is looking at the silver screen’s robots and not the emergent entities fused to a device’s screen.

Science Fiction, Robots, and the Human Paradigm

This part argues that mainstream twentieth-century science fiction’s imagination of artificial beings is predicated on what will be identified as the ‘human paradigm’. Starting with Čapek and Asimov, it is shown mainstream twentieth-century science fiction has informed a cultural imaginary regarding robots with three dominant tropes: anxiety about human futures, robots as embodied, and the distinction between humans and robots. These tropes form the human paradigm, and the repetition and reiteration of the human paradigm by science fiction has meant that robots are viewed through a ‘scanner darkly’ (Dick 2011). Robots might be envisaged, but humans remain the focus.

There is a broad and methodologically complex discipline of science fiction studies, particularly in relation to science fiction film (Telotte 2001). It has been argued that science fiction as a genre comes into its own through visual mediums (Sobchack 1997). Further, there is now an emerging cultural legal studies tradition that is particularly focused on cultural intertwining of law and science fiction (Green et al. 2022). The themes and focuses within science fiction studies and cultural legal studies of science fiction are diverse. Science fiction texts (broadly conceived as novels. plays, films, television, video games, products of prosuming) provide expansive vistas for play, critique, dreaming and fearing (Csicsery-Ronay 2008). However, a consistent feature of genre definitions of science fiction, and in the critical examinations and studies of science fiction texts is technology and the consequences of technological change (Tranter, 2018b, p. 3). Since Mary Shelley feverishly penned Victor Frankenstein creating his monster in his top floor room/laboratory, science fiction and its scholarly traditions, have indelibly engaged with technology (Yaszek 2023, p. 4). A significant concept is that of the ‘cultural imaginary.’ Tomislav Z. Longinović defines the cultural imaginary as ‘a realm of phantasms tied to one’s collective being’ (2011, p. 48). Within mass technological society, science fiction has been a crucible forging popular tropes, images and memes about technology and technological change (Robb & Tranter, 2023). Thomas M. Disch observed a ‘persuasive example of the power of ‘creative visualization’ [of science fiction in] the way the rocket-ship daydreams of the early twentieth century evolved into NASA’s hardware’ (Disch 1998, p. 7).

It is this ‘creative visualisation’ by science fiction, through informing the cultural imaginary about robots that is the focus of this part. It maps the core tropes. This trope mapping of mainstream science fiction’s cultural imaginary of robots is not an overtly sophisticated theoretical or methodological task. There is no need for the more detailed nuances of Darko Suvin’s ‘cognitive estrangement’ (Suvin 1979) that has been so influential in science fiction studies (Parrinder 2000, p. 4), or William P. MacNeil’s ‘reading jurisprudentially’ (MacNeil 2007) that has underpinned much of the recent cultural legal studies of science fiction (Travis & Tranter, 2024).Footnote 3 It is a surface level examination of primary denotations. Similarly, this is not a detailed examination of the entirety of science fiction in its sheer diversity or through an aca-fan’s obsessions with the cult, marginal and avant-garde (Kader 2015, p. 143). The cultural imaginary is comprised of collective phantasms that are shared, every-day and well-known. It is pop twice over. Popular in the sense as mainstream and ‘democratic’ in that the tropes, images, and memes circulate within pop culture. As such our consideration of science fiction goes to the mainstream and extremely well-known texts. Further our focus is on the tropes about robots associated with those texts. Often these texts do contribute to other cultural imageries of technological change, dystopia and existential themes that are often explored in detailed critical scholarship. However, at this stage of mapping the central tropes within science fiction’s cultural imaginary of robots, this broader contextualisation is secondary.

Mapping often begins with an origin, and the unambiguously popularly identified origin of robots in science fiction is with Čapek and Asimov.

Čapek and Asimov

The origin of the word ‘robot’ in English is due to Czech writer Karol Čapek’s 1920 play R.U.R. (Rosumovi Uměí Roboti), which is usually translated as ‘Rossum’s Universal Robots’ (Roberts 2005; Richards and Smart 2016) first performed in English by the Reandean Company at St Martin’s Theatre, London, in 1923 (Čapek and Čapek 1961). The narrative is familiar to contemporary audiences even if the actual text is not. Set on an island factory in the South Pacific, the headquarters and chief manufacturing plant of the Rossum Corporation, a collection of corporate humans (managing director, accountant, engineers) witness, over a series of years, growing malfunction within robots leading to revolution and their (and the human race’s) extermination (Čapek and Čapek 1961). In the triumphal words of the robot representative, ‘Mankind is no more. Mankind gave us too little life. We wanted more life’ (Čapek and Čapek 1961, p. 90). The final Act set several years after the revolution shows a crisis for Earth’s new robot overlords; with representatives of the robots pleading with the last human, the clerk Alquist, to bestow on the robots the capacity to reproduce (Čapek and Čapek 1961). With the secret of robot animation having been lost in the revolution. Alquist detects that the ‘male’ and ‘female’ robots that he was instructed to experiment on show care and affect, traits not seen in the other automatist robots (Čapek and Čapek 1961). The play ends with Alquist declaring the new ‘Adam and Eve’ and that with these two, the robots have become human (Čapek and Čapek 1961, p. 104; Segel 1995).

In many respects, R.U.R. brings Mary Shelley’s Frankenstein to the twentieth century. In common with Victor’s monster, the robots of the Rossum Corporation are flesh and blood. However, unlike Victor, whose creating is a private affair driven by a compulsive passion, Rossum has made money. The opening scenes show Rossum Corporation at full height, with an excess of orders for robots for the armies and the factories of Europe and North America, and the everyday corporate concerns of manufacturing capacity, quality control, and accounting practices (Čapek and Čapek 1961). Unlike Frankenstein, however, where the monster emerges as a pitiful character, spurred by his creator and cast into the world alone (Botting 1991), the robots in R.U.R. remain, even the supposedly Adam and Eve, alien and clunky. The revolution happens without much communication as to its cause (Čapek and Čapek 1961), unlike Frankenstein, where the monster speaks of hateful revenge (Shelley [1818] 1965). The dispassionate robots replace humans because, in their dispassion, they perceive themselves better. What can be seen in Čapek’s narrative of technology, consumption, and bloody revolution are the core elements of the human paradigm. First, the play gestures towards the future. There are anxieties about declining human births in the robot-oriented pre-revolution world economy (Čapek and Čapek 1961), and the narrative, overall, plays out a future where there are human ends (Čapek and Čapek 1961). Further, there is the possible rebooting of a ‘human’ future in the conclusion with the robots evolving to be human (Čapek and Čapek 1961). The drama concerned the decline and salvation of forms of ‘humanity’. Second, R.U.R. projected robots as ‘embodied’, that is as discrete entites within an identifiable, material body.Footnote 4 The body was in focus; both in the claims about robots’ carelessness towards their own body as a mark of their less-than-human status early in the play (Čapek and Čapek 1961) and also in the final act with the ‘evolved’ robots manifesting concern for their body and for the bodily integrity of each other as the signification of their transition to human (Čapek and Čapek 1961). Finally, the story of the substitution and replacement of humans by robots maintains a separation of the robots from the human. Robots kill and replace humanity.

Isaac Asimov rewired the anxieties of R.U.R. to present a vision of robots as maintaining their ‘proper place’ in relation to humans (Hockstein et al. 2007). Asimov was concerned that writers like Čapek could imagine engineers building dangerous machines without safeguards (Asimov 1967). He takes from Čapek robots as massed-produced goods and imagines that safety protocols will be hardwired into the working of their ‘positronic brains.’ Referred to within his robot short stories from 1941 to 1960 as ‘potentialities’ understood by equations, translated into English as the ‘Three Laws of Robotics’ were presented as:

First Law: A robot may not injure a human being or, through inaction, allow a human being to come to harm.

Second Law: A robot must obey orders given to it by human beings except where such orders would conflict with the First Law.

Third Law: A robot must protect its own existence as long as such protection does not conflict with the First or Second Law.Footnote 5

Asimov’s stories tell of the adventures of various employees—field testers, the head of Research and Development, and, especially, Susan Calvin, chief ‘robopsychologist‘—of US Robots and Mechanical Men Inc. (Portelli 1980; Leslie-McCarthy 2007). Each short story—and the more full-length robot detective novels featuring the human gumshoe Elijah Baley and his android partner R. Daneel Olivaw—is a puzzle (Tranter 2002; Leslie-McCarthy 2007). The protagonists are faced with a robot seemingly not complying with the Laws of Robotics, and the protagonists must reason and experiment to determine how situational context or human orders have led to the malfunction (Roberts 2005; Palumbo 2000). For example, the story ‘Liar!’ tells of an accidentally created mind-reading robot, Herbie (Asimov 2008, pp. 91–111). Herbie is caught in a paradox of the First Law. His mind-reading skills tell him how to answer a human so as not to cause harm. However, by misrepresenting to his human interrogators what he has read in the minds of others, he also harms. When Herbie is forced to comprehend the hurt from the social consequences of the misrepresentations and untruths: ‘you must tell them, but if you do you hurt, so you mustn’t; but if you don’t you hurt, so you must, but—’, he self-destructs (Asimov 2008, p. 110).

In both Čapek and Asimov, robots emerge as juridical beings. The robots manufactured by Rossum Corporation seem to obey a set of natural laws, a sort of ‘natural selection’ in their vanquishing of humanity and an evolutionary law of return in their becoming human. The industrial tools and domestic help constructed by US Robots and Mechanical Men can be seen as deontological beings possessing three internalised, hierarchal universal laws to make them safe. Indeed, Asimov’s robots rendered safe (or at least safer than Čapek’s genocidal variety) by the ‘Three Laws of Robotics’ are seen as an inspiration, if not a blueprint, for contemporary roboticists (Murphy and Woods 2009; Anderson 2008; McCauley 2007).

However, Asimov’s imagining shares more with Čapek than an implication of legality and the use of a derivative of the Czech word for ‘strenuous work’ (Segel 1995, p. 297) to denote the manufactured artificial beings of their titles. First, Asimov writes of a cornucopian human future (Tranter 2002); his stories begin in the then future of the 1990s and extend well into the twenty-first century.Footnote 6 Through robots, humans achieve exploration and exploitation of the solar system (Asimov 2008), peace and bountifulness on Earth (Asimov 2008), and faster-than-light interstellar travel (Asimov 1969, 2008). Second, Asimov’s robots are also distinctly embodied entities, whether the metal bodies of early prototypesFootnote 7 or the more human-esque, synthetic android bodies of later varieties (Asimov 2008).Footnote 8 Finally, there is a strict separation between humans and robots encoded by the very terms of the Laws of Robotics, which keeps distinct ‘humans’ and ‘robots.’ In Asimov’s universe, the human masters remain distinct from their robotic servants.Footnote 9 Like in Čapek, the only direction for change seems to be robots metamorphosing into humans, as told by Asimov in the later novella and played by Robin Williams in the film adaption of The Bicentennial Man (Asimov 1976; Columbus 1999; Coleman and Hanely 2009).

Human Futures, Bodies, and Separation in Science Fiction’s Robots

These tropes of human futures, bodies, and separation in Čapek and Asimov became the foundation for how mainstream science fictions portrayed artificial beings. Often these tropes are combined to reiterate the overarching narrative structure of Čapek or Asimov. A more sophisticated retelling of R.U.R. was the reimagined television series Battlestar Galactica (2004–2009) (Rymer 2003). The foundational premise of the series involves the annihilation of humanity by its rebelled robotic servants (the Cylons), which had evolved from metallic ‘toasters’ into fleshy humanoid models. The series further parallels R.U.R. in that by the end of the series, humans and Cylons form a mixed society, yet with a decidedly ‘human’ culture (Rymer 2009). Human future remains the focus along with embodiment, which is repeatedly emphasised intra-text through narratives and tropes about bodily reproduction (Tranter 2018b, 90 − 3). The one development from the R.U.R. template is that the new human culture is a hybrid of human and Cylon (Rose 2015). While this challenges the separation trope from R.U.R., it can be seen as a manifestation of Bicentennial Man, where the telos for robots is the human. Indeed, this is also the case for another well-known television robot, Data (Brent Spiner) from Star Trek: The Next Generation. Data’s narrative through the series and the subsequent films is to become ‘more human’ as he repeatedly saves the future human civilisation (Short 2003; Barrett and Barrett 2017).

These tropes ara also in mainstream science fictions that are not primarily concerned with robot narratives. The enduring characters of C-3PO and R2-D2 from George Lucas’ Star Wars franchise highlight the embodiment and separation of science fiction’s imagining of robots. The two ‘droids’ are presented as separate and different from the plethora of human and alien characters. Regardless of projected personality and appearance of sentience, Star Wars droids are presented as property to be traded and disassembled at will (Bryson 2010). Further, they are distinctly located within material bodies. C-3PO’s shiny gold humanoid body and R2D2 as a dustbin-on-wheels. R2-D2 needs to physically interface with computer systems; C-3PO can only communicate with humans, aliens, and other droids through speech.

This sense of material bodies and separation is not uniquely associated with Western robot narratives. Japanese robot stories often reiterate these tropes. The foundation of the Japanese robot and robot narratives is Osamu Tezuka’s Tetsuwan Atomu (Astro Boy) manga and anime (Sone 2017). Tezuka presents a humanoid boy robot, who interacts with and saves human society, but who also knows and identifies with being a robot (Tranter 2018a). Like George Lucas’ droids, Atomu is distinctly located within a body—his nuclear-powered, capable of flight, and weapon-equipped 4-feet-7-inches frame—but unlike Star Trek’s Data, who has a Pinocchio desire to become more human (Scheerer 1989), Atomu and his robot kin, remain separate and distinct from the human scientists, school children, and megalomaniacs of their shared world (Gibson 2010, 2012).

Even popular science fictions that attempt to problematise the founding tropes regarding robots often reiterate and reinforce others. The original Westworld film is a retelling of R.U.R. with the exception that it is not robot manifest destiny to kill and replace humans but rather a programming glitch (Crichton 1973; Telotte 1995). The recently reimagined HBO series Westworld brings to the forefront the sense of injustice regarding the treatment of the droids in Star Wars. It shifts identification from the ‘human’ amusement park visitors to the robotic ‘hosts’ and the horror, not of being hunted by killer robots, but that personhood, identity, and memories are an illusion (Le 2017; Winckler 2017). There is empathy for the robots, but only in so much as the robots are proxies for humans. There remains a particular focus on a distinction between humans and robots with the robots as entites located within a facsimile of a human body. The horror of the human/robot slippage for personhood and memory in Westworld has strong antecedents. It is a general theme across Philp K. Dick’s novels (Nelson 2009) and particularly emphasised in Ridley Scott’s adaption of Do Androids Dream of Electric Sheep? (Dick 1972) Blade Runner (Scott 1982), where the concluding suggestion is that gumshoe ‘replicant’ bounty hunter/investigator Deckard (Harrison Ford) is possibly a replicant. The body–memory/human–robot complex in Blade Runner (Landsberg 1995) becomes emphasised in Blade Runner 2049 (Villeneuve 2017), where, as in Battlestar Galactica, the horror of the indistinguishability between humans and replicants transforms into a quest for replicant sexual reproduction. The human body, in its biological mammalian specificity, becomes the paradigm for what it means to an entity in the material world.

There is a trajectory within mainstream science fiction’s robot oeuvre where this emphasis on the robot as clearly within a distinct, discrete body in time and space is inverted. The origins of the inversion are within Asimov’s original robot stories, where the robots merged with industrial and electronic systems to become an all-knowing, all-present entity (Asimov 2008, pp. 198–224). Rather than the becoming human of R.U.R., the robots become god-like; disembodied, all knowing, all seeing network that is everywhere. This can be seen in the chilling computer HAL in 2001: A Space Odyssey, who is the ship, a disembodied voice, and a blinking red light rather than a specific material entity like Star War’s R2D2 or Star Trek’s Commander Data (Nelson 2009). This inversion of the discrete material body trope was dramatised in the film I, Robot (Proyas 2004), loosely based on Asimov’s robot stories, where a single artificial intelligence (AI), VIKI, had merged with digital infrastructure and become manifested within the positronic robots of that world. Unlike the separate, embodied and distinct robots imagined by R.U.R and reiterated in mainstream science fiction, VIKI, and other filmic networked horrors such as the Terminator franchise’s Skynet or Marvel Avengers’ Ultron, show a tendency to immateriality in the imagining of artificial beings, of bodies being instrumental, temporary, and transient manifestations of an omnipresent techno-totality (Nelson 2009).

However, the anxiety and horror of HAL, Skynet, and Ultron comes from them as orchestrators of human ends. Each is R.U.R. but with just one big multi-bodied robot. While VIKI, following an Asimovian theme, is not inherently misanthropic, but rather motivated by the desire to care for and manage human futures by limiting free will, the emergence of the ‘disembodied’ immaterial robot still challenges human futures.

Nevertheless, even in imagining a robot as an immaterial omnipresent entity, the key tropes from twentieth-century science fiction are often reasserted. First, the narrative structure is always about human futures and that this future involves a cataclysmic distinction and separation between humans and the one big robots. Second, the presentation of an entirely bodiless entity is often redacted in the climax of these stories. The human protagonists thwart the robotic over-mind by physically destroying its ‘core.’ It turns out that the immaterial giant robot has a material center, a vestibule of a body – that can be materially affected. In this the previous suggestion of disembodiment turns out to be a projection and a ultimately a material body in space and time is reemphasised.

In conclusion, mainstream science fiction has been a storehouse for the cultural imaging of artificial beings. Notwithstanding, the diversity of narratives, subgenres and tones this cultural imaginary involves three related tropes: human futures, bodies, and separation. The focus might be on created beings, but the referent is the human and its futures. The basic human experience as an entity located within a material body in space and time remains the touchstone. Mainstream science fiction has an extended history of imagining robots but, in doing so, keeps seeing the human.

The Human Paradigm and Robot Law Scholarship

This part argues that legal thinking about robots tends to reiterate the human paradigm from science fiction. Robot law is a rapidly emergent field but with a pedigree. Just before and during World War II, there was mention of ‘robot bombs’ (Angell 1944) and ‘robot landings’ (Smith 1938). Nevertheless, the idea of a discrete scholarship on law and robots remained an extremely ‘minor jurisprudence’Footnote 10 during the twentieth century, Raymond August’s fabulous 1987 ‘Corpus Juris Roboticum’ a notable exception. However, since 2015 scholarship on corpus juris roboticum has entered the mainstream. There has emerged an expanding body of legal scholarship with a particular focus on robots, AI, and automated digital technologies. What is revealed by a review of legal writing about robots, AI, and automated digital technologies (robot law scholarship) is how it generally is grounded and emanates from the human paradigm established in science fiction. In parallel with science fiction’s robot narratives—while there are differences, challenges, and variations on the themes—there is a common core of human futures, bodies, and separation.

Human Futures

The anxiety about human futures trope is deployed in robot law scholarship in several ways. The primary location is in the introductory or general discussions of the current and future progress of robots, highlighting human anxieties about this future. It is common to see introductory remarks, either to an article in its entirety or to individual sections concerning the current state of technology, where anxiety about robots and AI and the implication for human futures is directly denoted. For instance, Mark Lemley and Bryan Casey (2019) write:

What happens when artificially intelligent robots don’t do what we expect…as robotics and AI systems increasingly integrate into our society, they will do bad things. (p. 1313)Footnote 11

These opening observations will commonly refer in general terms to the era of robotisation (Danaher 2016), the risks of the increasing integration of robots into human society (Lemley and Casey 2019), and the ‘technological revolution’ heralded by robots and various forms of automated technologies (Mazur 2018, p. 279), before delving into the specifics of the article. Opening remarks like ‘Robots and Artificial Intelligence are conquering our world’ (Simmler and Markwalder 2019, p. 1) or dramatic quotes used as chapter epigraphsFootnote 12 firmly frame the piece within a popular science fiction narrative. Tropes popularised in science fiction, such as robotic infiltration or encroachment in the human domain, arouse feelings of anxiety and urgency: ‘With each passing week, robots infiltrate deeper into our public spaces, places of work, and even bedrooms’ (Lemley and Casey 2019, p. 1318, our emphasis); they ‘slip into more and more domains of human life’ (Cardiell 2021, p. 247). The trust we have collectively placed in rapidly diffused new technologies without availing ourselves of ‘opportunities for reflection and debate’ has left us on a ‘razor’s edge’ (Guihot and Moses 2020, p. 354). Some avoid hyperbolic language in preference for describing a fundamental transformation of the human world and a ‘major paradigm shift’ (Mazur 2018, pp. 279–280), flagging long-standing ‘concerns about [AI’s] potential impact on humanity’ (Chesterman 2021, p. 1), or simply citing, in cautionary tones, the increasing independence of robots from ‘human supervisory control’ (Barfield 2018, p. 193). Others, though less evocative, unequivocally anchor the discussion within the narrative by, for example, making early reference to the acceptance and adoption of various robots and automated digital technologies and immediately highlighting a myriad of resulting questions to be answered and problems to be overcome: ‘What are the security and legal ramifications of the use of these new technologies? … What is the safety of these systems? … What are the privacy issues…? Could these systems fail?’ (Subramanian 2017, p. 82) There is a conception of robotic future in terms of binary outcomes: robots might ‘on the one hand, enhance and improve the quality of people’s lives but, on the other hand, can pose a threat to them’ (Katarzyna 2020, pp. 172–173).

Often robot law scholarship is predicated on anxieties about the legal and regulatory implications for future robotics. Sometimes the science fiction itself is explicitly deployed as the starting point, asking how we might apply Asimov’s laws to robots (Rigotti 2020) or treating Asimov as the genesis of robot regulation (Zödi 2020).Footnote 13 More prosaically, there is a worry about whether existing legal frameworks are sufficient and what changes can or should be made. Ronald Leenes and colleagues (2017) ask: ‘Are our existing normative frameworks adequate to deal with developments in robotics? Can new robotic technologies … be regulated within existing legal and ethical frameworks?’ (p. 2). Some scholars expressly engage with the anxiety trope, describing a ‘pervading sense of unease’ about the unknown impact on human life (Guihot et al. 2017, p. 385). Others endorse anxious ruminating on the potential adverse consequences for human society: ‘as artificial intelligence technologies continue to mature, humans continue to fret. And so we should.’ (Henderson 2019, p. 3, our emphasis). There is not uniformity on whether existing law and regulation is disruptive by robots. Some argue that are capable of dealing with novel and complex problems raised by developing robot and AI technology (see e.g., Bryan 2020). However, even where scholars expressly reject some of the more colorful manifestations of human anxieties, such as the ‘existential threat to humanity’ (Guihot et al. 2017, p. 389), other science fiction tropes—such as some miscalculations in design or robots’ realisation of their own superiority (and human obsoleteness)—are still acknowledged as being ‘so ubiquitous that [this fear] informs every level of discussion’ (Guihot et al. 2017, p. 398). The orientating point is still humans and the implications for human future.

While this aspect of the human paradigm is most evident in introductory or general remarks, anxiety about human futures also features in analysis that focusing on specific legal issues. An example is the future of the workplace and labor law. Legal literature acknowledges the transformative potential of robot and automated technology but tends to be framed by anxiety about the as-yet-unknown implications for human autonomy in the workplace and ‘a range of concerns’ which may be disruptive to the world of (human) work as it is currently known (Hendrickx 2018, pp. 366–67). A common fear is the R.U.R. trope of the potential for the ‘large scale replacement of human labour’ in the workplace by robots (Schofield-Georgeson 2020, pp. 264–65). In the context of personal care, Eduard Fosch-Villaronga (2020) has suggested that given the ‘malleability of the brain,’ the use of robots ‘will not leave human nature unchanged’ (p. 4). A further subject-specific example is the intersection of robots and human sexuality. The manufacturing and selling of sex robots is predicted to ‘create concerns’ across a variety of domains (Gersen 2019, pp. 1795–1796). However, what is interesting to note is that the anxiety in this context extends not only outward to broader society but also inward as unease around what robots might reveal about misogyny and the normalisation of violence towards women and children. For example, Gersen (2019) asks, ‘Will sex robots intensify human pleasures or magnify human horrors of exploitation, abuse, and rape?’ (pp. 1795–1796; see also Rigotti 2020). Similarly, literature addressing the adoption of robots in the aged and medical care industries highlights the apprehensions around implications for privacy, confidentiality, and autonomy (Blake 2019; Poulson et al. 2020, pp. 1–2) note the potential for ‘disastrous consequences’ due to ‘cyber-attack to these technologies’ or arising from the close physical interaction between the elderly and malfunctioning robots. Blake (2019) links anxieties about privacy to another theme comprising one version of the embodiment trope from science fiction, namely the potential for a version of robot ‘omnipresence’ in care robots (p. 567).

Robot law scholars also acknowledge and discuss the ways in which robots and emerging automated technologies can benefit society (Mazur 2018) and be a boon in a variety of industries (Subramanian 2017). Even where dystopian narratives are expressly recognised as a popular vision of the human future, at the other end of the spectrum usually sit the cornucopian dreams of far-reaching benefits from the development and application of this new technology (Guihot et al. 2017). The prevalence of the tropes around anxiety and human futures is not to suggest that robot law scholarship not highly sophisticated and nuanced. However, considering the language deployed, particularly in introductory comments and general observations on the progress of robots and automated systems, together with the challenges posed by how these technologies are framed, reveals how legal thinking and writing tend to reiterate and reinforce this aspect of the human paradigm from science fiction.

Bodies

Robot law scholarship also tends to conceive of robots and digital autonomy as ‘embodied’ in the narrow sense, that is located within a specific, identifiable system, vehicle, or machine. This downplays an essential feature of the digital: its diffusion and immateriality; its lack of a body. Given the tendency in legal scholarship to drills down into highly specific issues within broader legal subjects, this orientations towards robots as clearly identified material things, tends to be revealed in definitional choices, language bias, and the use of examples. Within robot law scholarship this is preference towards definitions of robots that favor robots as distinct material machines. For example, writing in the context of social robots, Ramesh Subramanian adopts a definition of a ‘physically embodied, autonomous agent’, certainly embodied (in contrast to ‘inanimate computers’), though not necessary humanoid (using ‘animaloid’ forms as the popular example) (see e.g., Subramanian 2017, pp. 85–86), while Ishii (2019) cites definitions which incorporate ‘physical agents that perform tasks by manipulating the physical world’ (p. 510). Along with autonomy, a ‘physical nature’ (that is, the ability to interact with material world) and ‘human-likeness’ are commonly assumed as essential robot features (Leenes et al. 2017, p. 4). Even where the limiting view of anthropomorphic robots is acknowledged as inadequate to cover the variety of possible shapes and forms, the definitions ultimately adopted tend to feature the elements of physicality and mechanics (see e.g., Hendrickx 2018). Leading scholars in robot law jurisprudence, such as Ryan Calo, define robots in the ‘strongest, fullest sense’ as ‘corporeal objects’, where embodiment is an essential quality (Calo 2015, pp. 531–532).Footnote 14

Even without expressly adopting definitions favoring material embodiment, the language used and examples employed in legal literature often reinforces the trope. For example, in exploring the possibilities of robot-caused harm, the scholars will acknowledge that many robot manufacturers might not be incentivised to create ‘humanoid robots’, yet in providing examples of alternatives forms, they will rely upon distinctly discrete machines such as cars and drones (see e.g., Danaher 2016). Scholars utilise examples of industrial machines and vehicles, complicated by their merging with AI (see e.g., Barfield 2018), or refer to humans being ‘hurt’ by robots and, as a consequence, robots being taken ‘control of’ and destroyed (see e.g., Mulligan 2018). The language implies the existence of some discrete material form capable of doing bodily harm and of being harmed. Imagined scenarios where robots are not as material tend to be less popular, for example, the potentially significant adverse consequences of robot lawyers or financial advisors providing negligent advice or even committing fraud.Footnote 15

In scholarship with a focus on personal injuries and related subjects, it could be expected that embodied robots would feature prominently. However, this conception of robots can also be seen in scholarship where bodily materiality does not seem as significant. For example, in considering hypothetically extending constitutional rights to advanced robots, R. George Wright expressly assumes robots as being ‘corporeal, embodied’ and rejects robots that do not physically manifest (Wright 2018, p. 615). Monika Simmler and Nora Markwalder (2019) assume that humanoid robots and other tangible machines, such as intelligent cars, will dominate the discussion on criminal culpability in the context of artificial intelligence. Similar tendences can be observed in the scholarship considering the future of advanced sex robots (‘life-size machine entities with human-like appearance, movement, and behavior’; Gersen 2019, pp. 1795–1798) and human-robot marriage (Walter 2020). While there might be a preferencing of humanoid entities; however, does this limited focus not deny the future possibility for (and likely present reality of) relationships—even kinds of sexual intimacy—between humans and non-embodied robots, such as entirely digitalised technologies?Footnote 16

In a large, growing and dynamic literature as robot law scholarship there are contributions that do not manifest science fiction’s human paradigm of robots. In investigating the ethical implications of domestic robots—a topic where embodiment tropes might be expected to dominate—Lachlan Urquhart, Dominic Reedman-Flint, and Natalie Leesakul expressly reject the limitation of materially framed robots and include disembodied forms of AI (Urquhart et al. 2019). Others highlight the potentially spurious basis for distinguishing between robots and AI based on embodiment or tangibility (Księżak and Wojtczak 2020; Guihot and Moses 2020). Leenes and colleagues (2017, p. 4) note the insufficiency of ‘common assumptions’ around the definitional dimensions of robots, highlighting the possibility of ‘non-physical…or non-human-like’ robots, while Morse (2019, p. 217) rejects the definition of robots as ‘objects’ in favor of a broader approach which admits ‘intelligent automatic systems’ such as TurboTax (see also the definitional choices of Lemley and Casey (2019) and Rachum-Twaig (2020). In discussing robots, ethics, and the rule of law, Pagallo (2017) avoids the accompanying images of humanoid robots or even machines, using examples such as ‘Vital,’ the first AI corporate board member, and drawing comparisons to the notion of autonomy already extended to corporations in some countries in the context of corporate criminal responsibility. Other lawyers delineate a boundary between robots, seen as essentially material bodied and even humanoid, versus AI and machine learning, though they acknowledge a blurring of the line between the concepts and accept the difficulty in predicting future development (Bennett and Daly 2020). Others still resist restrictive definitions altogether and argue that we should ‘embrace the ineffable nature of robots…we will know them when we see them’ (Casey and Lemley 2020, p. 294); however, in much robot law scholarship material ‘thingness,’ identifiable and discrete in space and time, something like the consumer goods made by United States Robot and Mechanic Men Inc in Asimov’s stories, is what is in focus.

Separation and Distinction

Within robot law scholarship there is a quarantining between humans as users, consumers, or creators of robots and robotic technologies. The robot and the human, as in science fiction, remain distinct. There is an acknowledgement that existing regulatory frameworks are predicated on assumptions, such as generic distinctions between ‘things and humans’ (Leenes et al. 2017, p. 2). There is worry at the prospect of increased robot autonomy detached from human control and robots ‘breaking norms’ understood by humans (Simmler and Markwalder 2019, pp. 8–9). There is a emphasis on the deeply humanness of the legal profession, such as the complexity of legal language, as a safeguard to social order—of ensuring ‘a rule of persons, not machines’ (Pasquale 2019, p. 1). A number of narratives familiar from science fiction are used to reinforce this trope, including robots as tools/property, robots as proxies for humans/human-robot empathy, and anxiety about blurring the lines of distinction.

As with R2D2 and C-3PO, lawyers write about robots in terms that emphasise their status as property owned and used by, and subservient to, humans. Choices of language are significant. Describing robots as ‘fighting our wars…manufacturing our products…caring for our most vulnerable…[and] driving our cars’ (Danaher 2016, p. 299, our emphasis) reinforces the separation of humans and robots and the notion of robots as tools for human ends. In the robot scholarship, the question as to the law’s adequacy to respond to advancing robot technology is often framed by a tension between robot autonomy and human control: ‘robots are becoming more and more independent from human supervisory control’ (Barfield 2018, p. 193). When grappling with the question of liability for robot-related harms or loss, lawyers understandably look to the designers and legal owners of robots and AI systems, but the language deployed is that of ‘human or corporate masters’ (Lemley and Casey 2019, p. 1314–1315). This view of robots is reflected in robot law scholar’s definitional preferences and conception of the future purpose of robotic and AI technologies. Not only is there a preference for definitions that require (mostly) physical embodiment with environmental sensory capabilities, but there is also an expectation that a robot’s purpose is to ‘enhance human possibilities for action’ (Simmler and Markwalder 2019, p. 5). The vision of the future is ‘increased convenience, comfort, companionship and greater security’ for human users of robots (Urquhart et al. 2019, p. 246). Bart Verheij summarises this neatly: ‘In short, AI should be good for us’ (2020, p. 181). In the context of criminal law, the idea of fully autonomous robots capable of the necessary mens rea for a criminal offense is dismissed, but robots are conceived as extensions or tools of human criminality and negligence (Pagallo 2017). This scenario is inverted in one imagining by Pagallo (2017), where robots eventually replicate themselves (another popular science fiction plot) and ‘use’ humans to commit crimes in the real world. The roles may be reversed, but the distinction remains.

The theme of human treatment of robots and its implications (familiar from science fiction such as Westworld) appears in the debates around sex robots. One of the anxieties regarding the use of sex robots is the unease around treating a robot only as an ‘object’ and ‘nonhuman’ (Gersen 2019, p. 1802; see also Rigotti 2020). The tension between the separation of humans and robots in this context is captured by Gersen (2019, p. 1809):

The more we attribute an involuntarily servile, nonhuman ‘nature’ to robots, the more unease we may feel about using them for sex, given our norms and culture of sexual consent. But also, the more that we come to think of robots as having human-like consciousness, will, sentience, and desire, the more wrong it may seem to use them simply to satisfy human sexual desire.

The debate on sex robots is important in the way it shows understanding and thoughts around robots are anchored in preexisting human constructs. For example, sex robots might be thought of as ‘bad’ if they replicate unequal human power relationships (Gersen 2019). Similarly, in considering whether to extend ‘rights to be protected from mistreatment’ to social robots, one of the reasons in support is the potential to ‘reinforce social norms against mistreatment that are valued in society’ (Bennett and Daly 2020, p. 78)Footnote 17 In this sense, robots are tools and the ultimate recipient of the benefits they offer are humans. The concern is the implications for existing senses of humanity.

The significance of human distinction and separation is also emphasised where scholars raise concerns with the slippage of human separation and distinction. There is speculation of how robotic prostheses ‘problematize the distinction between therapy and enhancement’ (Leenes et al. 2017, pp. 10–11), blurring the lines between human and machine. Wright, in considering the case of extending constitutional rights to robots, raises the question of whether, by doing so (or rather the basis on which such a move might be justified), ‘the constitutional rights of humans would be no less imperiled’ (Wright 2018, pp. 633–334). Might giving legal personhood to robots even lead to a loss of ‘human control,’ precipitating a ‘robot uprising’ (Genderen, 2018, p. 15)? As in science fiction, lawyers envisage a future (albeit a distant future) where a ‘master algorithm,’ by being able to ‘code its own progeny,’ might achieve that defining capacity of humanness: reproduction (Pasquale 2019, p. 5). Michael Froomkin and colleagues (2019) touch on the distinction differently by warning of a future where increased reliance on robot doctors reduces not only the number of human doctors in practice but, in turn, the number of human doctors available to train/program robot doctors.

Robot law scholarship also firmly secured humans and humanness as the referent against which robots and AI are viewed and measured. The ‘human’ was deployed in ways that problematise the perceived inadequacy of robots. For example, a difficulty for the law is thought of as being that robots ‘may not learn or reason like humans do’ (Barfield 2018, p. 193). Or put another way, the reason why scholars do not consider criminal responsibility attributable to robots is because of a fundamental orientation to not recognise robots as ‘equal to us’ on the basis that they lack human capacities (Simmler and Markwalder 2019, p. 20). However, humanness is also used as the bar or standard by which robots ultimately should be measured, reminiscent of Data’s progression towards increasing humanness in the Star Trek universe. In the context of extending human rights to robots, the threshold test is whether the ‘gap’ between humans and robots can be closed; it is those robots that ‘possess human-like qualities’ or AI that (though it may not be like a person) can ‘act’ like a person, to whom/which legal rights can and should be extended (Katarzyna 2020; Abbott 2021).Footnote 18 In considering the possibility of AI in litigation and judgment, Volokh (2018, p. 1137) imagines robots conversing or arguing in a way ‘indistinguishable’ from humans. Humanness is the end goal. Conversely, the literature reminds us that robots are less than, apart from, and which seek to ‘emulate’ human intelligence (Lemley and Casey 2019, p. 1327). Indeed, Evtimov and Calo \ describe AI as being understood ‘as a set of techniques aimed at approximating some aspect of human or animal cognition’ (Evtimov et al. 2019, pp. 895–896).Footnote 19

Of course, this distinction and separation are at times challenged, even where the trope is also reinforced. Pasquale (2019), for instance, advocates for a ‘complementary vision of human-machine cooperation’ (pp. 54–55). A simple example of such cooperation has been imagined (and is already being implemented) in terms of using AI assistants to support human judges in their decision-making.Footnote 20 Casey and Lemley have made the point that the lines have blurred, such that entities we think of as being robot or human are often ‘better understood as hybrids falling somewhere in between’ (Casey and Lemley 2020, p. 292–293). Though reinforcing the separation of human and robot in another way, they argue this very blurring necessitates a bright line in regulation delineating the intended subject, human or robot (Casey and Lemley 2020; cf. Abbott 2021). Chesterman (2021), in a pushing back against human-centrism, has suggested that increasingly autonomous AI ‘calls into question long-standing assumptions that humans are the source, the means, and the purpose’ of attempts to regulate this developing technology (p. 60). This ‘question(s) the assumption of our own centrality’, the warning and reaffirmation of our separation are included in that ‘it is not yet time to relinquish it’ (Chesterman 2021, p. 246). While not a coherent vision of a posthuman world, these views at least conceive of something approaching human-robot hybridity, However, this tendency is an undercurrent, and remains inform by the human paradigm of human futures, material bodies and distinction between humans and robots.

Talking Robots but Seeing ‘The Man’

What this part has traced is that the legal literature considering robots has as its touchstone humans. To quote Hendrickx (2018): ‘The ultimate central factor in the debate about robots and AI is the human factor … ultimately human freedom is at stake’ (p. 387). There is nothing challenging nor surprising in the fact that human lawyers writing about human law in the context of human technologists developing robots for use by humans, are focused on what this means for humans, their world, and their future. However, identifying this human-centrism through mainstream science fiction’s cultural imagination of artificial beings reveals that this human focus has deeper implications. The male science fiction authors of the twentieth-century robot canon and their mainstream televisual popularisers seemingly were unable—or unwilling—to imagine robots as fundamentally different. This meant that the popular narrations and representations of robots were as juridical beings, beings requiring law and bans (the R.U.R. robot) or internal ethical/legal constraints (the Asimov-ian robot) to ensure conservative human futures; that is, futures where the human as embodied, separate, and distinct endures. Given this cultural orientation, it is not surprising that legal scholars have embraced robots as objects of the juridical imagination—as something that requires law and professionalised legal thought, so to legislate for this future.

However, the human supposed by twentieth-century science fiction’s imagination of artificial beings and reified in robot law scholarship is not without its contestations. The human, particularly as materially embodied and distinct, presents as a conception of humanity that is deeply gendered and contested (West 1988). There is not just an human-centrism but an androcentrism to the ways that robots have been imaged in fiction and in law. This is at the fore in the sex robot discourses. The idea that androids styled as women could be used for sex runs deep in the teenage male daydreams of science fiction. It is implicit in R.U.R, entirely explicit in Dick’s textual and screened replicants, and now in the emerging sextech industry (Oliveira 2020). It is therefore not surprising that there has developed a substantial theme within the robot law scholarship of sexbots and the legalities and ethics of allowing another vector for hyper, propertied, hetero-normative masculinity. What this means is the human that the robot is consistently imagined through in popular culture and legal discourse is itself highly particular, less human, and in its dreams and nightmares of bodily, distinct futures, much more the man. This does make for main/male stream culturally aligned legal scholarship, but it is also enclosing and constraining.

The Enclosing and Constraining of Robot Law Scholarship

This part identifies that the human-centrism manifest in the robot law scholarship encloses and limits the legal imagery. By not clearly appreciating the complex, fluxing hybridity of humans and the digital, robot law scholarship speaks past its subject. There are two dimensions to this. The first is the digital and its relationship with the material. The second is the misconception of the human within the digital.

The first is the emphasis on robots as having distinct material bodies—that digital automation is in a physical entity, in a specific time and place—seems highly contested. While a trait within Asimov was robots ascending to a god-like immaterial, all-knowing master system, actual robots and automated digital technologies tend to be less embodied than imagined in science fiction but fall short of the alternative imagining of transcendent omnipresence. There are fears and dreams for the techno-total overlord. In the West, there are concerns that, lurking behind the innocuous voices of Siri or Alexa, there is an ever-present master AI datafy-ing human life to be exploited and monetarised by surveillance capitalism (Zuboff 2019). In the East, behind a digital great wall, a similar totality is being imagined and built to ensure that the party endures (Wong & Dobson, 2019). These desires of corporate and authoritarian data-lords for a version of mainstreams science fiction’s master robot, keeps being heralded, but seemingly remains on the horizon. The total robot overlord’s arrival in world history is entangled with the ontological dilemma of digital data. While textbooks define data as ‘collections of facts’ (Shukla et al. 2022, p. 1), this misses a critical dimensions. To extract a fact from the real into the digital, to collect it, to engage in datafication, strips the fact from its context (Crawford 2021, p. 93). The process of collection transmutes fact to data according to the collection parameters. Data has no inherent referent beyond its structural relationship to other data within the digital and those structural relations can be reprogrammed, deleted and corrupted: ‘the digital shows us unavoidably that nothing is permanent. Data are immaterial. They exist but they are no-thing’ (Coeckelbergh 2023, p. 9). The alignment of ones and zeros on a digital device can be a credit card number, a saved game, or a junk app. For the total robot to be, data and the real must correspond. To be exploitable for control and profit digital twins must reflect the data-producer (Goriunova 2019). This is why the data lords in the West and East are so keen on data accuracy, of correspondence between user and human unfettered by privacy screens (Tranter 2019), because without correspondence the total robot is partial and incomplete. It goes from techno-totality, with all the visions of power and control, to something less—a game of chance with probabilities of being right. This is not the digital Big Brother, but the digital Weird Uncle, who sometimes is right and meaningful but more often gets it so very wrong.

The problem of data for the arrived of science fiction’s imagined techno-total AI overlord is further compounded by technics. The capacity to datafy keeps eclipsing both storage and processing. This does suggest for the data-lords more investment in technical solutions (Schwarz 2018)—smarter recognition algorithms, even bigger data, quantum computing, DNA storage (Naeem et al. 2022)—to capture and render in the digital the totality of life. Yet there is always an excess, a vitality and exceptionalness to life that confounds representation (Braidotti 2019). Technics can approach the limit but will consistently be butting against the law of diminishing marginal returns.

What this means is that the disembodied master robot that knows all, as anticipated by mainstream twentieth century science fiction is delayed. The digital is experienced as both highly material and immaterial (Coeckelbergh 2023, p. 12). The robot has a material presence, encased in a physical form, the automated car, the auto-vacuum, and the sexbot, but this is also connected and immaterial. Data flows from the machine in the material to data nodes to be feasted upon by hungry AIs, which, in turn, inform the algorithms that animate the machine. This is the everyday of contemporary human life with the digital; the supposed seamless interactions between personal digital devices and various clouds, of the anxieties of access, backups, and bandwidth. There is continuous flow experienced through the televisual of the graphical user interface and the ubiquitous swiping right (Coeckelbergh 2023, pp. 10 − 2). Present in the materiality of touchscreens and digits is the immaterial of the digital archive and its AI predators. Not so much a ‘ghost in a shell’ (Giddens 2015; Bolton et al. 2007) but, to mix metaphors (and continents of origin), the ‘king’s two bodies’ (Kantorowicz 2016; Fukushima 2015). An intense embodied experience that is profoundly disembodied. A being there and everywhere instantaneously. This is the place where mainstream science fiction’s cultural imaginary of robots, and the ways that it informs legal discourse, fails. The robots are in the material. But also elsewhere, an avatar for a partial digital totality. This requires a distinctly different imaginary and lexicon than bequeathed by twentieth-century science fiction. Concerns with distinct categories—human, machine, embodiment, separation—needs to give way to concepts and terms of becoming, such as flux, flow, emergent (Gennaro & Kellner, 2022).

This follows to the second limitation. Not only is the robot misconceived, but so too is the human. There is an intimacy to the digital. The boundaries of the self and self-actualisation are changed (Englezos 2022). The precariat delivery rider is less envisaged as an individual human in the way that the man is presented in the robot law scholarship and more an assemblage composed of bicycle-human-device-platform (Standing 2018). Automated driving technology, as sold and experienced, is a hybrid form where the human and robot cooperate and share the dynamic driving tasks over a journey (Van Wynsberghe & Guimarães Pereira, 2022). The field of neural implants promises transhumanist dreams of a hardwired digitality (Lima 2022). As Ana Oliveria (2020) articulates, there is a ‘dematerialization of human bodies to the counter-materialization of digital bodies’ (p. 7). The real of the human and the artifice of the digital are bleeding and blurring (Hildebrandt 2015).

Furthermore, while the digital was challenging the embodied experience of what it means to be human, the traditional conceptions of the human as understood as the legal subject has been critiqued as a historical and political construct. Rather than an adequate encapsulating of the human experiences, it has been identified as based on conception of humans as rights bearing autonomous individualists that aligns with privileged, white masculinity (Cooke 2017). The embodied experiences of women, children, the less abled, the aged, and those with colored bodies have rarely been the proper subjects of law as the bearer of rights and legal protections (Naffine 2003, pp 363-7). Rather, there has been a negotiation of legalities that shift and interpose between a rights-bearing subject of law and an owned legal object (Kang 2011, pp. 111-2). Hybridity and transience between subjectivity and objectivity, of having to struggle to be a self-narrator within a loud semiotic field of discourses that define, constrain, and own, has been inimical to human experiences for all the others that did not correspond to the white man that has masqueraded as the human in legal discourses (Adebisi 2022; Käll 2022a). The human, even in modernity with its surface hard distinctions, has always been hybrid and transitory (Latour 1993), features that are becoming more obvious with the blurring of boundaries in the digital (Braidotti 2020, p. 1182). Indeed, robot law scholarship’s anxieties for the future of separate humans distinct from the robots who serve is dangerous mythmaking that tries to stabilise a subject and his law whose time has passed. This is the effect when mainstream science fiction’s robot tropes form the imagery for robot law scholars, it denies addressing the actual forms of life and power that are making the world. The robot is not to be feared or subjugated but is kin. Another assemblage of natureculture—to use Donna Haraway’s (2003) phrase that embraces the hybridity of the immaterial-material ‘all the way down’ (p. 12)—to interface with a posthuman present that is potentially new and different.

What this means for law, or more precisely for the normative orders of law, regulation and technological management (Brownsword 2019, 2020) that are being deployed in the present to secure preferred technological futures, is a need for critical re-examination of what is being ordered, for who and to what ends. To address this need there is a twofold task. The first is understanding of the ordering of human life with and through the digital. How are hybrid lives with and the through the digital being ordered, what values can be discerned, what processes identified, how is power constituted and circulating? Legal scholars are beginning to progress this agenda (See generally Hildebrandt 2015; Hildebrandt and O’Hara 2021; Noll 2019; Käll 2018, 2022b; Arvidsson 2018). The second task is imagining better futures for the hybrid lives of entities that are after the human. In this task the cultural imaginary of science fiction is significant. Science fiction has contained divergent imaginings that reacted and critiqued the male-stream projections of mainstream populist science fiction (Larbalestier 2006; Cashbaugh 2016; Cox-Palmer-White 2021). Science fiction also as a global genre is creating lived in worlds beyond the Western cultural mindset (Bisschoff 2020; Banerjee 2020; Langer 2011). There is also an emerging scholarship beginning this process of connecting these alternative imaginings to issues of justice, judgement and normative ordering (Sharma 2022; Travis 2022). This critical re-examination presents opportunities for ordering hybrids of digital and the real into the future, rather than looking back to tropes of separate humans and discrete robots that has been the predominate perspective within robot law scholarship.