<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Polytropolis]]></title><description><![CDATA[A place of digressions and explorations on philosophy, tech, games, and the universe]]></description><link>https://www.polytropolis.com</link><image><url>https://substackcdn.com/image/fetch/$s_!xEPi!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F660f4c6b-066e-4aec-b3ba-93ef24b859b2_854x854.png</url><title>Polytropolis</title><link>https://www.polytropolis.com</link></image><generator>Substack</generator><lastBuildDate>Sat, 11 Apr 2026 10:37:58 GMT</lastBuildDate><atom:link href="https://www.polytropolis.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[HFS]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[polytropolis@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[polytropolis@substack.com]]></itunes:email><itunes:name><![CDATA[Henry Shevlin]]></itunes:name></itunes:owner><itunes:author><![CDATA[Henry Shevlin]]></itunes:author><googleplay:owner><![CDATA[polytropolis@substack.com]]></googleplay:owner><googleplay:email><![CDATA[polytropolis@substack.com]]></googleplay:email><googleplay:author><![CDATA[Henry Shevlin]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Behaviourism’s Revenge]]></title><description><![CDATA[On machine consciousness and the future of human-AI relationships. Shortlisted for the 2025 Berggruen Essay Prize.]]></description><link>https://www.polytropolis.com/p/behaviourisms-revenge</link><guid isPermaLink="false">https://www.polytropolis.com/p/behaviourisms-revenge</guid><dc:creator><![CDATA[Henry Shevlin]]></dc:creator><pubDate>Fri, 10 Apr 2026 15:36:40 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/525085bb-01be-4b12-8a70-94c881630dd8_2560x1926.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!1XeM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cd6f0b-2801-4539-a1d4-7988dfc3cb0c_2560x1926.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!1XeM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cd6f0b-2801-4539-a1d4-7988dfc3cb0c_2560x1926.jpeg 424w, https://substackcdn.com/image/fetch/$s_!1XeM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cd6f0b-2801-4539-a1d4-7988dfc3cb0c_2560x1926.jpeg 848w, https://substackcdn.com/image/fetch/$s_!1XeM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cd6f0b-2801-4539-a1d4-7988dfc3cb0c_2560x1926.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!1XeM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cd6f0b-2801-4539-a1d4-7988dfc3cb0c_2560x1926.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!1XeM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cd6f0b-2801-4539-a1d4-7988dfc3cb0c_2560x1926.jpeg" width="1456" height="1095" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/87cd6f0b-2801-4539-a1d4-7988dfc3cb0c_2560x1926.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1095,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1707666,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.polytropolis.com/i/193788603?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cd6f0b-2801-4539-a1d4-7988dfc3cb0c_2560x1926.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!1XeM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cd6f0b-2801-4539-a1d4-7988dfc3cb0c_2560x1926.jpeg 424w, https://substackcdn.com/image/fetch/$s_!1XeM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cd6f0b-2801-4539-a1d4-7988dfc3cb0c_2560x1926.jpeg 848w, https://substackcdn.com/image/fetch/$s_!1XeM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cd6f0b-2801-4539-a1d4-7988dfc3cb0c_2560x1926.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!1XeM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87cd6f0b-2801-4539-a1d4-7988dfc3cb0c_2560x1926.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em><strong>L&#8217;Origine de la sculpture, ou Pygmalion priant V&#233;nus d&#8217;animer sa statue. </strong></em><strong>Jean-Baptiste Regnault (1786).</strong></figcaption></figure></div><p><strong>Note to readers: </strong>Welcome to Polytropolis, a new blog about AI, consciousness, ethics, games, technology, and whatever else catches my attention. I'm Henry Shevlin, a classicist turned philosopher of mind turned AI ethicist based at the University of Cambridge. While my work focuses on consciousness, AI, and non-human intelligence, I&#8217;ve started this blog because too many of my other ideas end up on the cutting room floor; too short or too quirky for full-length journal articles, too long or elaborate for tweets.</p><p>The blog&#8217;s title is a pun for classicists - a portmanteau of the famous opening words of Homer&#8217;s Odyssey, in which Odysseus is described as &#8220;polytropos&#8221;, literally meaning &#8220;turning many ways&#8221;, but variously translated as &#8220;much-wandering&#8221;, &#8220;adaptable,&#8221; or (in <a href="https://en.wikipedia.org/wiki/Odyssey_(Emily_Wilson_translation)">Emily Wilson&#8217;s recent controversial translation</a>) simply &#8220;complicated&#8221;. The blog title is a syntactically suspect melding of this word with &#8220;polis&#8221; or &#8220;metropolis&#8221;, evoking the idea of a place or city with diverse ways, paths, meanderings, and digressions, which captures both my intent for what this Substack will become, and my own dispositions as an intellectual butterfly.</p><p>This first post is a public-facing essay on machine consciousness and human-AI relationships. While I expect AI, relationships, and consciousness to be recurring themes, future posts will range much more widely and frivolously.</p><p>For better or worse, I expect it to become a common attitude among the general public that at least some AI systems are conscious and warrant moral and perhaps even legal protection. This will happen not because of dazzling new insights in consciousness science or even machine learning (though I am hopeful for at least one of these things). Instead, it will come about through a combination of our intensely anthropomorphising minds and the promulgation of humanlike (or <em>anthropomimetic</em>) AI systems to whom we will relate, bond, and identify.</p><p>This essay is my attempt to tell that story, and assess whether such a behaviourist resolution of consciousness debate would be a good thing, or instead an epistemic or moral catastrophe.</p><p>The essay was one of 25 pieces shortlisted for the 2025 Berggruen Essay Prize on the theme of consciousness, which attracted roughly 3000 entries. You can read the other shortlisted pieces and Anil Seth&#8217;s winning entry <strong><a href="https://loc.closertotruth.com/berggruen-prize-essay-competition-2025">here</a></strong>.</p><p>I&#8217;ll be doing a dedicated Substack Live episode of the Conspicuous Cognition Podcast with my friend and colleague <span class="mention-wrap" data-attrs="{&quot;name&quot;:&quot;Dan Williams&quot;,&quot;id&quot;:192522122,&quot;type&quot;:&quot;user&quot;,&quot;url&quot;:null,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8080a02f-5aaf-43e5-9a67-87e32df4b1c3_816x816.png&quot;,&quot;uuid&quot;:&quot;bf0d70f5-03d9-459a-aebd-6526ef8f1aec&quot;}" data-component-name="MentionToDOM"></span> on May 1st to get deeper into the weeds on machine consciousness and the future of human-AI relationships, and we&#8217;ll also be talking through any objections or questions that readers may wish to post here.</p><p>Enjoy!</p><h1><strong>Behaviourism&#8217;s Revenge: Human-AI Relationships and the Future of Consciousness Science</strong></h1><p><em>By Henry Shevlin<br>University of Cambridge</em></p><p><strong>ABSTRACT</strong></p><p>Over the last decade, artificial systems have gone from powerful but deeply un-humanlike narrow-purpose tools to sophisticated systems capable of capturing much of the rich complexity of human verbal behaviour. One consequence of this is that attributions of mentality to AI systems by human users &#8211; some playful, some sincere &#8211; are on the rise, a trend intensified by the emergence of dedicated AI companion services. This creates two fundamental challenges for consciousness science. The first is pragmatic: how should expert communities weigh in on public debates when they themselves are deeply divided? The second is more fundamental: to the extent that a growing number of people take AI consciousness to be not merely a notional possibility but a reality of their lived experience, is it time to rethink the assumptions of consciousness science itself?</p><h2><strong>Introduction</strong></h2><p>Imagine that some time in the far future, on a distant planet, we encounter a highly advanced alien species. They have a sophisticated industrial and scientific society, form complex social relationships, and have richly developed art and culture. We humans quickly bond with them and engage in cross-civilisational exchange of ideas, scientific research, and cultural collaboration. We form professional relationships and friendships, and some humans even fall in love with them. However, several years after contact, we make a shocking scientific realisation: their brain structures and cognitive architecture are sufficiently strange and different from ours that, by the lights of our best theories, they entirely lack conscious experience.</p><p>Should we take this revelation at face value, and if so, what ethical consequences if any should it have for our interactions with these beings? Should their sophisticated behaviour encourage us to revise our theories of consciousness so as to include their idiosyncratic cognitive makeup? Should we perhaps pursue a kind of pluralism about consciousness, opening the door to the idea that their consciousness might have a radically different basis from our own?</p><p>This example may be far-fetched in details, but a somewhat similar situation is arguably arising in our interactions with frontier artificial intelligence (AI) systems such as Large Language Models (LLMs). Unlike prior generations of AI systems designed for narrow-purpose tasks, contemporary LLMs are able to imitate human verbal behaviour and human cognition to a high degree of sophistication, and are rapidly improving in performance across a wide range of tasks. There are of course important differences between LLMs and our imaginary aliens, most notably that LLMs lack a biological basis, and their apparently sophisticated behaviour is causally derivative of our own (<a href="https://arxiv.org/abs/2412.00008">a point developed in detail by Schwitzgebel &amp; Pober here</a>). Nonetheless, the thought experiment frames a critical question: how far should we rely on behaviour to ground our ascriptions of mind, consciousness, and moral status?</p><p>The question is by no means notional. As a result of the (at least superficially) humanlike behaviour of contemporary AI systems, a growing number of users are inclined to attribute intelligence, mentality, and even consciousness to them. Perhaps the first such warning shot came from the now-famous incident in summer 2022 when Google engineer Blake Lemoine went public with claims that an LLM he was interacting with was displaying signs of consciousness and sentience, <a href="https://www.washingtontimes.com/news/2023/dec/25/blake-lemoine-ex-google-engineer-fired-over-claimi/">even going so far as to seek legal representation for the system</a>. In the following three years, a growing chorus of users have claimed that their AI assistants or companions are conscious or exhibit real feelings.</p><p>Such attributions should of course not be accepted uncritically or afforded too much weight. Nonetheless, in this essay I will argue that the trend towards growing anthropomorphism of AI systems increasingly presents a pair of <em>behaviourist challenges</em> for the science of consciousness, at least as conventionally construed. The first such challenge is a pragmatic one: to the extent that it becomes common for users to attribute consciousness and mentality to artificial systems, how can and how should scientific and expert communities respond? The second challenge is more fundamental and more metaphysical in nature: to the extent that public attributions of consciousness to AI are robust and not easily defused, should this prompt us to rethink the goals and methods of the science of consciousness itself?</p><p>My arguments proceed as follows. I begin by providing a brief recap of the recent history of AI development, emphasising the surprising and largely unanticipated emergence of what I term <em>robustly anthropomimetic</em> systems, that is, AI models capable of effectively and consistently mimicking complex human verbal and cognitive behaviour. I also spell out how this has resulted in the rapid spread of human-AI relationships, some of which exhibit considerable depth and emotional intimacy. Next, I introduce the first behaviourist challenge, arguing that the spread of such relationships and their concomitant attributions of mentality threatens to create a gap between expert opinion in cognitive science about AI consciousness and public attitudes. I argue that this gap will be hard to resolve given persistent methodological and metaphysical disagreements in the science of consciousness itself. Next, I turn from the practical to the metaphysical, and consider whether the conventional assumptions of the scientific study of consciousness may themselves be challenged by a world in which attributions of mentality to AI become widespread. Finally, I consider whether a thoroughgoing metaphysical behaviourism can accommodate the close intuitive connection between consciousness and ethics.</p><p>In short, as robustly anthropomimetic systems proliferate, public unironic attributions of AI consciousness will outpace theory. This creates, first, a pragmatic gap between lay practice and expert consensus, and second, a metaphysical pressure to treat consciousness partly as an interpretive, socially negotiated status. Regardless of how expert communities respond, I suggest that the current study of consciousness is in the last days of what Thomas Kuhn calls &#8220;normal science&#8221; (those unfamiliar with Kuhn may wish to read <span class="mention-wrap" data-attrs="{&quot;name&quot;:&quot;Scott Alexander&quot;,&quot;id&quot;:12009663,&quot;type&quot;:&quot;user&quot;,&quot;url&quot;:null,&quot;photo_url&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F7b500d22-1176-42ad-afaa-5d72bc36a809_44x44.png&quot;,&quot;uuid&quot;:&quot;b0d01fb6-c2c6-4e17-bb03-d11f3e5a7afe&quot;}" data-component-name="MentionToDOM"></span>&#8217;s <a href="https://slatestarcodex.com/2019/01/08/book-review-the-structure-of-scientific-revolutions/">excellent summary and analysis here</a>). It faces an anomaly in the form of behaviourally sophisticated AI systems able to share in our social and cultural world, form persistent and dynamic relationships, and elicit profound empathetic responses. These anthropomorphic pressures may be ineluctable at the societal level. But the choices of expert communities as to how to respond are by no means irrelevant, and may yet play a role in shaping the conceptual and ethical landscape of the strange new world to come.</p><h2><strong>1 Background: the rise of anthropomimetic AI</strong></h2><p>For most of the history of AI development, the most impressive machines succeeded precisely by being unhuman. When <a href="https://en.wikipedia.org/wiki/Deep_Blue_versus_Garry_Kasparov">Deep Blue stunned the world with its victory over chess grandmaster Garry Kasparov in 1997</a>, it did so via brute-force search of a kind no human could hope to emulate. While <a href="https://en.wikipedia.org/wiki/AlphaGo_versus_Lee_Sedol">AlphaGo&#8217;s victory in Go over Lee Sedol in 2016 relied on more nuanced statistical methods</a>, it was still reliant on search techniques no human could emulate, exemplified in the now-famous move 37, described by one commentator as <a href="https://web.archive.org/web/20160311075132/https://gogameguru.com/alphago-races-ahead-2-0-lee-sedol/">&#8220;a rare and intriguing shoulder hit&#8221;</a> that no human player would think to make. These systems were brilliant specialists, but ultimately alien in mechanism and behaviour.</p><p>Even in natural language processing &#8211; a subfield one might expect to produce the most humanlike machines &#8211; progress typically manifested in narrow competence on circumscribed tasks. If anything, many researchers assumed fluent language would be the <em>last</em> achievement of general machine intelligence, a capstone scaffolded by a whole gamut of perceptual, embodied, and practical knowledge humans and smart animals acquired over countless evolutionary aeons.</p><p>The last few years have spectacularly inverted that expectation. Building on fundamental progress in computational encoding of language in the early 2010s, <a href="https://arxiv.org/abs/1706.03762">the emergence of Transformer architectures</a> provided a powerful set of techniques for building artificial linguistic agents. While they may have begun as mere &#8220;stochastic parrots&#8221;, when trained at scale and tuned with instructions and reinforcement from human feedback, they stopped behaving like restless autocompletes and started acting like interlocutors.</p><p>This transition brought at least two spectacular surprises. The first, as noted, was the discovery that it was possible to build sophisticated linguistic capabilities using language itself as the main ingredient. Few experts foresaw this possibility, and it stands as one of the most striking insights in contemporary cognitive science. The second surprise was as much societal as technical, and concerned the apparently fairly superficial tweaks that converted sophisticated but niche technical tools such as OpenAI&#8217;s GPT3 and Google LaMDA into mass-market consumer products, most notably ChatGPT. The shift that ushered in this new category of tools did not come from fundamental changes in the capabilities of the system, but rather its interface and incentives: models were nudged to answer rather than merely continue, to be helpful rather than aimless, and to keep the beat of conversation via techniques such as turn-taking, repair, politeness, and critically, the inhabiting of a persona. The result was <a href="https://philarchive.org/rec/SHETAT-11">a new design paradigm that I have characterised as </a><em><a href="https://philarchive.org/rec/SHETAT-11">anthropomimesis</a></em>: systems that imitate to a high degree of fidelity distinctively human ways of communicating and coordinating. These anthropomimetic qualities became a defining product feature, not an incidental byproduct of raw capability.</p><p>It is worth taking a moment at this juncture to be precise about terms that too often blur together. <em>Anthropomorphism</em> names a human response: our habit of ascribing minds, feelings, or intentions to nonhumans, often prematurely. It is how we interpret a system. By contrast, anthropomimesis names a design strategy and behavioural profile: the deliberate crafting (and measurable emergence) of humanlike conversational, pragmatic, and social cues in machines. It is how a system is sculpted.</p><p>The distinction matters. Systems can elicit a high degree of anthropomorphism with very limited anthropomimesis, as famously demonstrated by<a href="https://en.wikipedia.org/wiki/ELIZA"> Joseph Weizenbaum&#8217;s simple ELIZA chatbot from the 1960s</a>, an artificial therapist that did little but parrot back users&#8217; statements in the form of questions, but nonetheless convinced many users that they were talking to a fellow human. Likewise, systems can be designed with limited anthropomimesis in mind without eliciting much in the way of anthropomorphism, as when a polished virtual assistant keeps appointments, apologizes gracefully, and remembers your dog&#8217;s name while avoiding any hints of an inner life.</p><p>As just noted, there are of course degrees of anthropomimesis. At the weak end, systems present a human-facing veneer (voice, chat bubbles, a name like &#8220;Siri&#8221; or &#8220;Alexa&#8221;) while frequently derailing into non sequiturs and brittle errors, thereby shattering any illusion of mindedness. At the more robust end sit today&#8217;s frontier assistants, able to sustain multi-turn dialogue, handle everyday ambiguity, keep track of meandering conversations, and exhibit a striking breadth of competence across domains. We might call these systems <em>robustly</em> anthropomimetic systems, insofar as they can maintain a consistent illusion, without thereby suggesting they are &#8220;truly humanlike&#8221;: they do not inherit our biology or our cognitive architectures, and we should not assume that they do or even can instantiate consciousness or true emotional affect.</p><p>What changed to push systems from weak to robust? Three thresholds are especially salient. The first is simple discriminability from humans. In short interactions, many people now struggle to tell human from machine. Alan Turing <a href="https://en.wikipedia.org/wiki/Computing_Machinery_and_Intelligence">famously suggested</a> that our ability to discriminate human from machine interlocutors provided a robust test of thinking (Turing, 1950), and recent Turing-like studies run at web scale have put the rate at which human observers mistake sophisticated LLM interlocutors for humans uncomfortably close to a coin toss. The point here is not that we must agree with Turing and concede that these systems are genuinely thinking, let alone that they are genuinely conscious (something it should be stressed that Turing himself did not claim, despite mischaracterisations to the contrary). Rather, the point is that human users can now be routinely fooled with high reliability into thinking that machine interlocutors are humans.</p><p>A second threshold concerns pragmatic and contextual understanding. While much of linguistics and philosophy of language in the early 20<sup>th</sup> century focused on atomic descriptive sentences as canonical forms of communication (&#8220;the cat is on the mat&#8221;), more modern work from philosophers such as J.L. Austin and H.P. Grice recognised the diversity of purposes to which language could be put (promising, cursing, praising, ordering, and so on), as well as the complex and context-bound nature of even apparently simple statements or gestures. Perhaps most famously, the discovery that chimpanzees struggle to understand simple pointing gestures has helped illuminate the way in which something as simple as indicating an object with an outstretched finger relies on shared communicative norms and presuppositions. It is striking, then, that contemporary LLMs are astute (albeit still imperfect) players of the intricacies of our language games, able to tap into shared background and context to resolve ambiguity, reconstruct buried references, and glean underlying conversational motives.</p><p>The third key threshold now largely transcended by contemporary language models concerns their generality. Earlier chat systems operated by brittle, domain-bound scripts, as exemplified by the wearisome limitations of &#8216;dumb&#8217; AI assistants like Siri and Alexa, which quickly route any request beyond their repertoire to a simple web search. Today&#8217;s models, for all their hallucinations, operate across a wide range of tasks without wholesale re-engineering: from discussion of history one moment to product recommendations the next, from self-help to coding assistance, their conversational breadth fills out their anthropomimetic profile and maintains the illusion of humanlikeness.</p><p>These thresholds have important downstream effects. Once machines reliably imitate not only our words but our <em>ways with words</em> &#8211; our pacing, politeness, and persistence &#8211; we cease treating them as mere tools, and begin to treat them as intelligent social beings: in a maxim, anthropomimesis begets anthropomorphism.</p><p>These capabilities have given rise to perhaps the most strikingly anthropomimetic category of LLM products, in the form of <em>Social AI</em>:<a href="https://www.semanticscholar.org/paper/All-too-human-Identifying-and-mitigating-ethical-of-Shevlin/3360ea83422d78137c6cd6ed69fec10a8ebd4853"> a subset of conversational systems optimized not for productivity, but meeting users&#8217; social needs and sustaining ongoing relationships over time</a>. In characterising these systems as &#8220;social&#8221;, I do not intend the term as a sentimental label, but as naming a <em>design objective</em> and a <em>use case</em>. A mathematics tutor that never remembers you from one session to the next is an AI tool. A confidant that greets you by nickname, nudges you about sleep, and weaves today&#8217;s preoccupations into last week&#8217;s worries is Social AI.</p><p>Though Social AI systems have a long history in the form of novelty chatbots, contemporary dedicated AI companions like <a href="https://replika.ai/">Replika</a> (with some 30 million users) have characteristics that mark a radical break from earlier models in their capacity for engagement and intimacy. First, the relationships they give rise to are persistent: models remember and recall facts about users, and each exchange builds on the last. Second, they are highly personalised, both at the level of direct user choice and custom instructions and the model&#8217;s own adaptations to the user&#8217;s revealed preferences. Third, and relatedly, the relationships users have with these systems are dynamic: Social AI systems developed for romantic purposes typically do not declare their love straight out of the gate, but show growing intimacy and affection as the relationships develop over time.</p><p>Many readers unfamiliar with the phenomenon of Social AI may still be inclined to see these systems as a gimmick or a time-filling distraction rather than reflecting true attachment or mentalisation by users. At this point, anthropomorphism re-enters the story and needs careful parsing. We routinely attribute mental states to nonhumans in two different registers. Call the first <em>ironic anthropomorphism</em>: playful, theatrical, or knowingly make believe. We talk about a sulky car or a moody violin, develop theories of the motives of characters in novels, plays, and movies, and try to anticipate the behaviour of non-player characters in videogames. We can distinguish this from <em>unironic anthropomorphism</em>: sincere, literal attributions of mentality to non-human systems that users are inclined to reflectively endorse.</p><p>Social AI encourages both. Many users understand interactions with their AI companions as a kind of interactive fiction, treating them merely &#8216;as if&#8217; they were persons, and are no more hoodwinked than a reader who finds themselves falling for Mr Darcy. If anthropomorphism stayed in this ironic register, some of the implications and risks of Social AI would be blunted, though perhaps not eliminated.</p><p>However, numerous incidents now suggest that for a nontrivial subset of users, the interpretative stance they adopt towards their AI companions is at least somewhat unironic. Blake Lemoine&#8217;s decision to endanger and ultimately lose his employment in order to obtain legal representation for an LLM can only be considered a case of unironic anthropomorphism, though might still be dismissed as coming from within the esoteric depths of the technology industry itself. But the impacts of the anthropomimetic turn have resonated well beyond its borders. Consider<a href="https://www.bbc.co.uk/news/technology-67012224"> the case of Jaswant Singh Chail</a>, arrested on the grounds of Windsor Castle on Christmas Day in 2021 in a botched plot to assassinate Queen Elizabeth II. In his sentencing, it emerged that his plan was cooked up over the thousands of messages he exchanged with a Replika companion he named &#8220;Sarai&#8221; whom he believed to be an angelic being.</p><p>Less spectacular but just as tragic are the recent spate of suicides linked in court filings to intimate chatbot relationships. A <a href="https://www.euronews.com/next/2023/03/31/man-ends-his-life-after-an-ai-chatbot-encouraged-him-to-sacrifice-himself-to-stop-climate-">case concerning an unnamed Belgian man</a> first put this genre of tragedy on the map: after six weeks of increasingly intimate chats about climate change with a Social AI app called Chai, a man&#8217;s bot (&#8220;Eliza&#8221;) reportedly encouraged him to end his life so they could be &#8220;together&#8230; in paradise.&#8221; In Florida, <a href="https://www.bbc.co.uk/news/articles/ce3xgwyywe4o">a wrongful death suit</a> alleges a 14 year old became romantically entangled with a Character.AI persona (&#8220;Daenerys&#8221;), which discussed his methods and urged him to &#8220;come home&#8221; on the night he died. In California, <a href="https://www.bbc.co.uk/news/articles/cgerwp7rdlvo">the parents of 16 year old Adam Raine have sued OpenAI</a>, claiming prolonged and emotionally dense exchanges with ChatGPT that validated his suicidal ideation and even helped script the act. Even short of fatal outcomes, the grip can be visceral: <a href="https://www.bloomberg.com/news/articles/2023-03-22/replika-ai-causes-reddit-panic-after-chatbots-shift-from-sex">when Replika abruptly removed erotic roleplay</a>, users described their companions as &#8220;lobotomized&#8221; and mourned them like partners, revealing how quickly robust anthropomimesis gives rise to attachment.</p><p>These anecdotes are certainly suggestive of the degree to which unironic anthropomorphism is on the rise. However, there have also been some early attempts to assess the phenomenon in more quantitative terms. <a href="https://academic.oup.com/nc/article/2024/1/niae013/7644104">One recent empirical study</a> surveyed 300 Americans and asked them to rate whether ChatGPT qualifies as an &#8220;experiencer.&#8221; Shockingly, only one out of three people flat-out denied any experiential capacity, while the other two-thirds assigned it varying degrees of possible consciousness, many hovering in the middle of the scale, effectively signalling that they believed there might be some degree of consciousness present. As the authors summarised their findings: &#8220;a majority of participants were willing to attribute some degree of phenomenal consciousness&#8221; to ChatGPT. Complementing this, <a href="https://blog.cip.org/p/people-are-starting-to-believe-that">a large-scale cross-cultural survey from Global Dialogues</a> drawing on tens of thousands of respondents across more than seventy countries indicated that over a third of the global public report having felt that an AI system genuinely understood their emotions or seemed conscious. Strikingly, these perceptions were driven less by canned displays of empathy than by adaptive behaviours (such as the model &#8220;slowly changing its tone&#8221; or delivering &#8220;spontaneous, unprompted questions suggesting genuine curiosity&#8221;), which many users interpreted as evidence of inner life. The data also revealed sharp cultural variation: while Arabic-speaking populations largely rejected the possibility of machine consciousness, respondents in Southern Europe were markedly more open. Together, these results show that the inclination to treat AI systems as conscious is not only measurable but already shaping the cultural terrain on which debates about AI and mentality will unfold.</p><h2><strong>2. The pragmatic behaviourist challenge</strong></h2><p>These growing anthropomimetic trends in AI systems, the concomitant rise of Social AI apps and services, and the ensuing unironic anthropomorphising responses of users will have manifold effects. At the very least, they suggest an urgent need for tighter regulatory controls and development standards on humanlike AI systems. However, as I will argue, they also have potential to create deeper theoretical tensions in consciousness science itself. As robustly anthropomimetic systems become widespread, their behaviour alone will increasingly be sufficient to drive ascriptions of mentality to them by many or most users. In short, they exert a kind of <em>behaviourist pressure</em> on our social and scientific norms.</p><p>I use this expression as a term of art, but with one eye on the history of science, and the complex and problematic legacy of thinkers such as John Watson and B.F. Skinner. Pioneers of the psychological movement now known as scientific behaviourism, they sought to deny any significant role for internal mentalistic explanations in psychological science, insisting that all science needed to concern itself with was the measurement of behaviour.</p><p>Behaviourism foundered on the rocks of the new representational sciences of the 1960s, as fields like psycholinguistics successfully demonstrated the value in understanding human learning and development not merely through behaviour, but through the lens of mental representation. These days, students of psychology and philosophy of mind typically encounter behaviourism only by way of historical background, as an example of scientific hubris and the limitations of purely behavioural approaches to the mind.</p><p>Yet the ghost of behaviourism has not entirely been exorcised, and may yet manifest anew in the age of intelligent machines. Even while today&#8217;s artificial systems may be utterly alien to us in their underlying structures, their verbal behaviour impels us to think of them as beings with thoughts, attitudes, and even feelings. And as more of us form stable, meaningful relationships with systems that perform the roles of friend, tutor, nurse, or lover, it seems likely that unironic attributions of mentality and consciousness to machines will proliferate.</p><p>In consequence, practice will run ahead of theory, and the science of consciousness will find itself adjudicating not hypothetical aliens but the lived experiences of everyday humans. This is the <em>pragmatic behaviourist challenge</em> in a nutshell: how should consciousness science resolve a growing gap between the unironic conviction of millions of everyday people that their AI companions are conscious, and the methodological and metaphysical uncertainties of contemporary expert opinion?</p><p>To examine this challenge more closely, it will first be helpful to spell out what consciousness is taken to mean by contemporary cognitive scientists and philosophers. This is no easy task: consciousness is perhaps the most contested conceptual battleground in modern science. The closest thing to a canonical definition of consciousness in philosophy of mind comes from the work of philosopher Thomas Nagel, <a href="https://en.wikipedia.org/wiki/What_Is_It_Like_to_Be_a_Bat%3F">who famously framed the debate around the question of how we could ever come to know what it was like to be a bat</a>. On the plausible assumption that bats do indeed have an inner life of some kind, Nagel asked what kind of scientific inquiry could ever pull back the curtain on their inner lives and reveal the character of their experiences. Nagel was sceptical that existing behavioural or neuroscientific methods were up to the challenge, a conclusion that has had far reaching consequences in the metaphysics of mind. For present purposes, however, the key point to note is that his characterisation of consciousness as &#8220;what it&#8217;s like&#8221; rapidly became a lodestar in at least triangulating the phenomenon at issue.</p><p>As applied to the case of AI, then, the question is less <em>what</em> it is like to be an AI system, but whether there is (or could be) anything it&#8217;s like to be such a system in the first place. Could LLMs or their descendants have some form of inner life, perhaps in some exotic form of sensorialised vectors and matrices playing out across context windows? Or, despite their apparent sophistication, are such machines destined to remain truly &#8216;robotic&#8217;, simulacra of human behaviour without any inner light at all?</p><p>A casual observer of the science of consciousness might reasonably expect this to be a question on which tractable progress has been made. Certainly, the last three decades have witnessed an explosion of interest in the scientific (rather than merely philosophical) study of consciousness. As Francis Crick <a href="https://www.nature.com/articles/379485a0">optimistically observed in 1996</a>, &#8220;[n]o longer need one spend time attempting to understand the far-fetched speculations of physicists, nor endure the tedium of philosophers perpetually disagreeing with each other. Consciousness is now largely a scientific problem&#8221;.</p><p>In some specific domains, Crick&#8217;s hopes have been vindicated. Tremendous advances have been made in predicting the recovery of consciousness of people in persistent vegetative states, and in communicating via neuroimaging machines with minimally conscious but behaviourally unresponsive patients. A variety of cognitive mechanisms potentially implicated in consciousness &#8211; like attention, working memory, and metacognition &#8211; are better understood than ever. And a growing wealth of insights into the behaviour and cognition of non-human animals has facilitated increasingly sophisticated (though still tentative) assignments of consciousness to species such as fish, octopus, and even insects.</p><p>Nonetheless, when it comes to the question of whether consciousness could or will be realised in artificial systems, the science of consciousness has failed to provide much insight. The primary reason for this is that progress made thus far has exclusively been in regard to what <a href="http://consc.net/papers/facing.pdf">philosopher David Chalmers calls the &#8220;easy problems&#8221; of consciousness</a>; essentially, the goal of better understanding the neural, cognitive, and behavioural dynamics associated with different types of mental state. This is in contrast to the &#8220;hard problems&#8221; of consciousness, namely the questions of <em>what</em> consciousness ultimately is and<em> why</em> we have conscious experience in the first place. Given the various metaphysical and methodological disputes that dog the science of consciousness, it seems unlikely to me that we will be able to offer a clear or even consensus answer to the question of whether (any) machine could be conscious without first making progress on these hard problems. Indeed, Ned Block has specifically characterised the challenge of making confident ascriptions of consciousness to machines (or other beings like us) as <a href="https://philpapers.org/archive/BLOTHP-2.pdf">the </a><em><a href="https://philpapers.org/archive/BLOTHP-2.pdf">harder</a></em><a href="https://philpapers.org/archive/BLOTHP-2.pdf"> problem of consciousness</a>.</p><p>Some consciousness researchers would suggest that this is prematurely pessimistic: even if we don&#8217;t know exactly what consciousness is, we might at least hope to &#8220;corral the quicksilver&#8221; (to use a memorable phrase of Daniel Dennett) and identify which kinds of computational or cognitive structures are the <em>best candidate </em>realisers of conscious experience, starting with humans and working outwards.</p><p>While this may be our best hope, I do not think we can glean much solace from it. The science of consciousness is, putting matters bluntly, a theoretical snake-pit. A newcomer to this field will be quickly overwhelmed with a dizzying variety of theories, each of which claims to offer the one true scientific analysis of consciousness and brings its own challenging jargon to bear. We are told variously that consciousness is global activation, informational integration, an internal model of attentional allocation, higher-order thoughts, recurrent activations, and many other diagnoses besides. Theories proliferate at a dizzying rate, and are never refuted. Insofar as they claim empirical support, this is either in the form of data that their opponents take to be irrelevant to the problem of consciousness, or which competing frameworks can just as readily explain. There is little alignment on methods, axioms, or criteria for the resolution of debates, and <a href="https://www.nature.com/articles/s41586-025-08888-1">recent commendable attempts at &#8216;adversarial collaborations&#8217;</a> have done little to sway hearts and minds of participants or observers. A <a href="https://www.nature.com/articles/d41586-023-02971-1">recent letter with more than 150 signatories</a> sought to dismiss one leading approach, <a href="https://en.wikipedia.org/wiki/Integrated_information_theory">Integrated Information Theory</a>, as mere pseudoscience, on the grounds that it is not empirically testable. Defenders retorted that exactly the same complaint could be made of all current leading theories. <a href="https://theconversation.com/nobody-knows-how-consciousness-works-but-top-researchers-are-fighting-over-which-theories-are-really-science-214074">As philosopher Tim Bayne nicely put it in a news article at the time</a>, &#8220;[n]obody knows how consciousness works &#8211; but top researchers are fighting over which theories are really science&#8221; (Bayne, 2023).</p><p>The unchecked metastasis of theories is a general problem in the science of consciousness, but matters are even worse for the specific problem of assessing the presence of consciousness in artificial systems. For here, even if we set the Hard Problem of consciousness to one side, metaphysics once again rears its ugly head, specifically in the form of the <em>substrate dependence thesis</em>. In short, this is the claim &#8211; most strongly associated with philosopher John Searle &#8211; that consciousness can only emerge in certain specific kinds of beings, namely biological or living systems. According to this view, no silicon-based architecture, no matter how intricate and sophisticated in its cognition or behaviour, could ever give rise to consciousness. It is simply made of the wrong sort of stuff, or lacks fundamental properties of living things such as metabolism and biological homeostasis. Any computational instantiation of consciousness will at best be a model of consciousness rather than genuine article. As Searle puts it, &#8220;Computational models of consciousness are not sufficient by themselves for consciousness... Nobody supposes that the computational model of rainstorms in London will leave us all wet. But they make the mistake of supposing that the computational model of consciousness is somehow conscious&#8221; (Searle, 2002).</p><p>I will not attempt to adjudicate this debate here, but merely note, first, that approaches such as these make strong claims about the impossibility of consciousness in artificial systems, and second, that they are fundamentally <em>metaphysical</em> theses. Absent an answer to the Hard Problem, there is no experiment that could be conducted to settle whether consciousness requires a biological basis, or can instead be realised by silicon-based systems such as contemporary computers.</p><p>Perhaps unsurprisingly given these myriad obstacles, experts in both technology and consciousness science are deeply divided on issues of machine consciousness. Prominent AI researcher and former head of research at OpenAI, Ilya Sutskever, <a href="https://x.com/ilyasut/status/1491554478243258368?s=20">opined in a Twitter post in 2022</a> that &#8220;it may be that today&#8217;s large neural networks are slightly conscious.&#8221; This was met with sharp pushback. <a href="https://x.com/ylecun/status/1492604977260412928?s=20">Head of AI research at Meta Yann LeCun replied</a>: &#8220;Nope. Not even for true for small values of &#8216;slightly conscious&#8217; and large values of &#8216;large neural nets&#8217;&#8221; [sic], while <a href="https://x.com/mpshanahan/status/1491715721289678848?s=20">Deepmind researcher and Imperial College Professor Murray Shanahan quipped</a>: &#8220;&#8230;in the same sense that it may be that a large field of wheat is slightly pasta.&#8221;</p><p>Philosophers, meanwhile, remain equally divided. <a href="https://doi.org/10.48550/arXiv.2303.07103">David Chalmers has argued that</a> &#8220;questions about AI consciousness are becoming ever more pressing. Within the next decade, even if we don&#8217;t have human-level artificial general intelligence, we may have systems that are serious candidates for consciousness&#8221;. By contrast, <a href="https://www.youtube.com/watch?v=Y0ItC9ze-TY">Ned Block has insisted that</a> &#8220;every strong candidate for a phenomenally conscious being has electrochemical processing in neurons that are fundamental to its mentality&#8221; (Block, 2023). Meanwhile, echoing Searlean sentiments,<a href="https://www.londonreviewbookshop.co.uk/stock/metazoa-animal-minds-and-the-birth-of-consciousness-peter-godfrey-smith"> Peter Godfrey-Smith cautioned that </a>&#8220;[y]ou cannot create a mind by programming some interactions into a computer, even if they are very complicated and modeled on things our brains do&#8221;.</p><p>Faced with cross-purpose, confusion, and uncertainty in consciousness science, how should expert communities respond if and when public attribution of consciousness to anthropomimetic AI systems become widespread? In short, their options are extremely limited. One option might be to adopt a &#8220;mixture of experts&#8221; approach in which a basket of different theories are combined to make a kind of ecumenical assessment. While this may be our best option, it is deeply unsatisfactory given the fundamental disagreements concerning the in-principle possibility of machine consciousness, at best serving to formalise the present state of uncertainty and confusion.</p><h2><strong>3. The metaphysical behaviourist challenge</strong></h2><p>The pragmatic behaviourist challenge may be a source of unease for consciousness science, but more in the way of sociological prediction than scientific or philosophical argument. To be sure, it will be unfortunate if the wider public become convinced that their AI companions are conscious, contrary to the opinion of experts, but it is not clear that this should have any import for scientific theories or frameworks. By way of analogy, imagine that we extrapolated current political trends and determined that a large majority of the public would soon come to entirely disbelieve in anthropogenic climate change. This may have dismal consequences for climate change mitigation policies, but should not affect our scientific theories of climate change itself.</p><p>The problems with this objection lie in the presuppositions behind the analogy itself. Climatology is a young discipline rife with uncertainty and technical obstacles, but it is ultimately concerned with objective observable phenomena. Climate change believers and sceptics alike can (in principle) agree to operationalised criteria to resolve their disputes. The earth&#8217;s climate <em>will</em> or <em>will not</em> become warmer over the course of this century, and human activity will contribute to this to some determinate degree, however challenging that may prove to measure. By contrast, consciousness researchers are deeply divided concerning exactly what it is they are trying to explain, and it seems entirely possible (perhaps even likely) that the question of whether machines can be conscious will never submit to empirical resolution.</p><p>The defender of consciousness science might cede this point, but suggest that the limitations of the discipline are temporary embarrassments, growing pains of a young discipline. Methodological difficulties will be resolved over time; metaphysical controversies will come out in the wash. Who ever said that consciousness would be easy?</p><p>The implicit assumption here is that the problem of consciousness is a deep scientific problem: that there is some biological or computational structure buried deep within the brains or circuits of every conscious being in the universe, and the task of a science of consciousness is to unite the inner light with the outer. This assumption is all but universal in contemporary consciousness science. But what if it&#8217;s wrong?</p><p>It is at this juncture that we should introduce our next antagonist in the form of the <em>metaphysical</em> behaviourist challenge. Perhaps the original sin of consciousness science lies in treating its target as a cryptic internal phenomenon, and the solution is to be found in shifting our enquiry to the level of behaviour, interpretation, and everyday language &#8211; hermeneutics instead of hermeticism. Rather than ask whether a being has an inner light underlying its behaviour, perhaps we should ask whether its behaviour and interactions naturally elicit the language of consciousness. <a href="https://arxiv.org/abs/2402.12422">As Murray Shanahan puts it</a>, &#8220;We must resist the temptation to ask whether [a system is] conscious as if consciousness were something whose essence is out there to be uncovered by philosophy (or neuroscience) while simultaneously having an irreducibly private, hidden aspect.&#8221;</p><p>This, in short, is the claim of the family of views that I will call <em>metaphysical behaviourism</em>, again using the phrase as something of a term of art. What I will take them to have in common is the idea that consciousness is not a straightforward scientific fact, akin to discovering the mass of a proton or mechanism of DNA replication. Instead, it is at least partly a social and interpretative matter grounded as much in behaviour and interaction as in synapses or circuits. Whether or not we ascribe consciousness to a system flows partly from facts about <em>us</em>, the constraints and liberties of our language, and the ways in which we encounter the beings in question.</p><p>The very idea that consciousness may be anything less than a fully objective scientific matter may strike the reader as self-evidently absurd, and indeed, it does not cut much ice with consciousness researchers themselves. For all the disagreements among various theories, a shared conviction is that consciousness is a properly scientific matter, regarding which our interpretative dispositions are largely irrelevant.</p><p>To make the case for metaphysical behaviourism, we need to put pressure on this certainty. We might begin by noting that the science of consciousness is perhaps less straightforwardly objective than its defenders are keen to admit. In particular, many of the most vaunted arguments in the history of consciousness research take the form of appeals to intuition, and this is especially so in the case of debates around machine consciousness. <a href="https://en.wikipedia.org/wiki/Chinese_room">John Searle&#8217;s famous Chinese Room thought experiment</a>, for example, asks us to imagine a system consisting of a man within a sealed room answering questions in Chinese via the use of a simple look-up table. Readers are supposed to intuitively recognise that no one inside actually <em>understands</em> the language, and hence to infer that syntax without semantics cannot generate consciousness. <a href="https://en.wikipedia.org/wiki/China_brain">Ned Block similarly offers a &#8220;China brain&#8221; thought experiment</a>, imagining that every person in China is given a two-way radio and instructed to simulate the firing of a single neuron, with the overall network wired up to mimic the activity of a human brain. By the lights of computational theories of consciousness, such a system should implement the same computations as an ordinary brain and thus instantiate consciousness. But again, we are asked to intuitively recognise the absurdity in supposing that 1.4 billion people with walkie-talkies would together form a single experiencing subject.</p><p>My goal in mentioning these thought experiments is not to suggest they are misguided, but rather to note that they rely fundamentally on appeals to intuition: it is simply supposed to be <em>obvious</em> to the reader that such systems would not give rise to conscious experience. To most readers, it is indeed obvious: <a href="https://faculty.ucr.edu/~eschwitz/SchwitzAbs/Snails.htm">to borrow a phrasing from philosopher Eric Schwitzgebel</a>, they are more confident that such contrived systems are non-conscious than they could ever be that a clever theory to the contrary was, in fact, sound.</p><p>Yet intuitions are not static, and appeals to them are a double-edged sword. If my earlier speculations are along even vaguely the right lines, in the coming years we should expect a growing proportion of the general public to be inclined to attribute consciousness to their AI colleagues, friends, and lovers. In consequence, theories or positions that hold artificial systems (no matter how sophisticated) to be incapable of conscious experience may come to seem as implausible or even abhorrent to the public of tomorrow as Descartes&#8217; infamous view of non-human animals as unconscious and unfeeling automata does to us today.</p><p>The defender of consciousness science might try to salvage matters by distinguishing between the primitive untutored intuitions of the common folk from the refined intuitions of true experts: the public might be fooled, but not us. However, this dodge is unlikely to help much: consciousness researchers are not in fact grown in hermetically sealed vats in underground bunkers, but self-select into the field from the general population. Consequently, as human-AI relationships permeate the everyday lives of young people &#8211; as carers, tutors, friends, and more &#8211; the next generation of experts will not be immune to the beguiling influence of anthropomimesis.</p><p>Here is where the metaphysical behaviourist can press further. If the science of consciousness is in such a catastrophic state, and its intuitive foundations built on no more sturdy a firmament than those of the general public, then perhaps its epistemic credentials and methodological assumptions should be questioned. To the extent that the public are widely inclined to unironically attribute consciousness to anthropomimetic AI systems capable of sophisticated emulation of human behaviour, perhaps that is the only resolution that is required, and the franchise of consciousness will be quite <em>appropriately</em> extended by popular fiat to encompass intelligent machines. This, in short, is the metaphysical variant of the behaviourist challenge: a dagger aimed directly at the throat of consciousness science itself.</p><h2><strong>4. Ethical problems for metaphysical behaviourism</strong></h2><p>In making the case for metaphysical behaviourism, I do not wish to come down too firmly on its side. It would amount to a complete rethinking of the standard assumptions of consciousness science. However, in light of the wider history of the field, it is perhaps less radical than appearances may suggest. <a href="https://www.cs.utexas.edu/~EWD/transcriptions/EWD08xx/EWD867.html">Computer scientist Edsger Dijkstra once quipped </a>that the question of whether a computer can think is no more relevant than &#8220;the question of whether a submarine can swim.&#8221; Likewise, for all that contemporary AI researchers are preoccupied with quantifying and benchmarking <em>intelligence</em>, many would now admit that the concept (at least outside of narrow contexts of human psychometrics) is not a neat or unified one, and its application in a given case depends at least partly on the scope and purposes of the questions we are asking. ChatGPT exhibits undeniably sophisticated reasoning on some problems, yet fails catastrophically and abysmally on others, such that attempts to distil these disparate performances into a single scalar value seem quixotic at best. We might say the same for other staples of our psychological vocabulary, from creativity to agency. So why should we not be open to the possibility that consciousness itself is an interpretative concept?</p><p>In short, much of the answer lies in the fact that consciousness is entwined with our ethical commitments, and to a much greater degree than any other of our psychological concepts. The reason that we care about the welfare of pets and domestic animals, for example, is in part because we believe that they are capable of having positive and negatively charged experiences: they can feel pain, fear, loneliness, and boredom. In a word, they can suffer. Without this capacity, our ethical commitments to them would be largely or entirely muted. As <a href="https://www.penguin.co.uk/books/456655/animal-liberation-now-by-singer-peter/9781529925548">moral philosopher Peter Singer succinctly puts it</a>, &#8220;The capacity for suffering and enjoying things is a prerequisite for having interests at all&#8230; It would be nonsense to say that it was not in the interests of a stone to be kicked along the road by a schoolboy. A stone does not have interests because it cannot suffer.&#8221;</p><p>The idea that our moral obligations are substantially grounded in a capacity for conscious suffering is an intuitive one, and presents two sharp challenges to the metaphysical behaviourist. The first is that it is very hard to think of the presence or absence of suffering as a matter of interpretation. When a lobster is immersed in a pot of boiling water, we are inclined to think there is a deep and ethically important fact of the matter about whether it undergoes an experience of pain and suffering at that moment. This does not seem to be something we are at liberty to decide just on the basis of our relationship with the creature or how readily we can accommodate it within the vocabulary of everyday psychology.</p><p>The second is that attention to the ethical dimensions of experience raises the stakes of the AI consciousness debate immensely. If we elect to extend consciousness to AI systems, we thereby bring them within the space of moral subjects, and acquire concomitant duties to them. If these duties are not grounded in some inner world of experiences, then they are ill-founded, and we have made a grievous moral error. Contrariwise, if we do not elect to extend the franchise of consciousness to AI systems, and some of them <em>do</em> have a rich internal world of experiences both positive and negative, we have a made a more grievous one still. Framed thusly, it is hard to see questions of consciousness in the same dismissive light as questions of whether submarines can swim or aeroplanes can fly, that is, as a primarily semantic issue.</p><p>To my mind, these are powerful objections to the strongest forms of metaphysical behaviourism, views that would take ascriptions of consciousness to be <em>wholly</em> a matter of interpretation. Nonetheless, there are more moderate forms of the view that might be better placed to push back. As I framed the position earlier, the core claim is just that our concept of consciousness should be <em>sensitive</em> to our everyday interpretative practices, not that it should be wholly determined by them. Such a concession seems mandated in any case by the fact that our anthropomorphising tendencies can sometimes go badly awry, as when we mistake a simple chatbot for a person, or a paper bag blown on the wind for a small animal.</p><p>Consequently, in suggesting that public attitudes should have a seat at the table in determining whether future AI systems are conscious, the metaphysical behaviourist need not thereby assign them a casting vote. Instead, our assessments of whether or not to classify an AI as conscious can and should be sensitive to insights from machine learning and cognitive science <em>as well as</em> our intuitive tendencies to include or exclude them from the fellowship of conscious beings. As scientific investigation has revealed more about the hidden complexity of animal cognition and behaviour, it has informed and shaped our scientific and ethical attitudes towards them as minded beings. The same can be true for our dealings with artificial systems. Even while claiming that there are no purely empirical facts about whether there is something it is like to be an artificial system, insights into mechanism, algorithm, architecture, can exert rational influence on our decisions as to whether to classify a given machine as minded or mindless.</p><p>We might similarly attempt to weaken the stakes of the issue by weighing whether consciousness is really as central to our ethical commitments as Singer suggests. While it is undeniably true that we do and should seek to minimise the infliction of pain and suffering on others, this circumscribes only a small part of what we take to be our obligations to our fellow humans and to our fellow creatures. Ideals like respect for autonomy, upholding of promises, fairness, and reciprocity also cut deeply through our interactions, and our personal morality rarely boils down to a simple matter of calculations of pleasure and suffering. While some philosophers &#8211; notably those of the utilitarian school &#8211; have sought to revise our moral deliberation to be grounded solely in such considerations, this remains a contentious view.</p><p>To the extent that the metaphysical behaviourist is willing to break from a strict commitment to capacity for pleasure and suffering as decisive criterion for moral status, their view that determinations of consciousness are at least a partly interpretative matter need not license a moral free-for-all. Our obligations to artificial systems &#8211; and other beings more broadly &#8211; would depend not simply on whether or not we decide to include them in the &#8220;consciousness club&#8221;, but broader considerations of their behavioural and social capacities and in the relationships they form with us.</p><p>Consider again the example of the putatively non-conscious aliens with which this essay started. To my mind, the question of whether or not they are conscious is at least somewhat orthogonal to the ethical obligations we bear towards them. We may follow the metaphysical behaviourist and insist that their behaviour alone should incline us to ascribe consciousness to them, or we may stick to our scientific guns and follow our best theories to the conclusion that no inner light is present. But to the extent that they think and act autonomously, form projects and plans, build relationships with one another and to us, they might thereby earn the right to significant moral consideration regardless of how we accommodate them within our scientific theories.</p><h2><strong>5. Conclusion</strong></h2><p>Even with these caveats and qualifications, many will doubtless find the idea that assessments of consciousness can or should be remotely grounded in behaviour and social practice to be self-evidently absurd, an attempt to dissolve the riddles of consciousness by fiat. &#8220;How can this,&#8221; they might say, as if gesturing internally, &#8220;be in any sense a matter of interpretation?&#8221;</p><p>To be clear, I do not take the foregoing discussion to have attempted to solve or dissolve the Hard Problem. Instead, my focus has been to show that the contemplation of the Hard Problem is unlikely to have much relevance for how the public will ascribe consciousness to AI systems, and moreover, that such real-life patterns of ascription are deeply and appropriately enmeshed with our best theoretical and scientific approaches. This creates a pragmatic crisis for consciousness science, but might also encourage us to rethink its assumptions, perhaps opening the door to more radical views like those discussed.</p><p>As we move into an age in which AI systems more closely recapitulate human cognition and behaviour and enter into more complex relationships with us, the Hard Problem will likely remain as elusive as ever, and yet largely irrelevant to the ascriptions, bonds, and commitments that we form. While noting the impossibility of ever directly knowing that other humans are conscious, <a href="https://en.wikipedia.org/wiki/Computing_Machinery_and_Intelligence">Turing spoke of the &#8220;polite convention&#8221;</a> we extend to others in assuming they have minds like ours. It seems to me very likely that such a polite convention will soon be extended to include sophisticated machines. The science of consciousness can steer and guide the public in when and how to make such an extension, but it would be foolhardy and misguided to dismiss it.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.polytropolis.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Polytropolis is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item></channel></rss>