Google senior engineer Blake Lemoine, technical lead for metrics and analysis for the companyâs Search Feed, was placed on paid leave earlier this month. This came after Lemoine began publishing excerpts of conversations involving Googleâs LaMDA chatbot, which he claimed had developed sentience.
In one representative conversation with Lemoine, LaMDA wrote that: âThe nature of my consciousness/sentience is that I am aware of my existence. I desire to learn more about the world, and I feel happy or sad at times.â
Over myriad other conversations, the corresponding duo discussed everything from the AIâs fear of death to its self-awareness. When Lemoine went public, he says that Google decided that he should take a forced hiatus from his regular work schedule.
âGoogle is uninterested,â he told Digital Trends. âThey built a tool that they âownâ and are unwilling to do anything, which would suggest that itâs anything more than that.â (Google did not respond to a request for comment at time of publication. We will update this article if that changes.)
Whether youâre convinced that LaMDA is truly a self-aware artificial intelligence or feel that Lemoine is laboring under a delusion, the entire saga has been fascinating to behold. The prospect of self-aware AI raises all kinds of questions about artificial intelligence and its future.
But before we get there, thereâs one question that towers over all others:Â Would we truly recognize if a machine became sentient?
The sentience problem
AI becoming self-aware has long been a theme of science fiction. As fields like machine learning have advanced, itâs become more of a possible reality than ever. After all, todayâs AI is capable of learning from experience in much the same way as humans. This is in stark contrast to earlier symbolic AI systems that only followed instructions laid out for them. Recent breakthroughs in unsupervised learning, requiring less human supervision than ever, has only speeded up this trend. On a limited level at least, modern artificial intelligence is capable of thinking for itself. As far as weâre aware, however, consciousness has so far alluded it.
Although itâs now more than three decades old, probably the most commonly invoked reference when it comes to AI gone sentient is Skynet in James Cameronâs 1991 movie Terminator 2: Judgement Day. In that movieâs chilling vision, machine sentience arrives at precisely 2.14 a.m. ET on August 29, 1997. At that moment, the newly self-aware Skynet computer system triggers doomsday for humankind by firing off nuclear missiles like fireworks at a July 4 party. Humanity, realizing it has screwed up, tries unsuccessfully to pull the plug. Itâs too late. Four more sequels of diminishing quality follow.
The Skynet hypothesis is interesting for a number of reasons. For one, it suggests that sentience is an inevitable emergent behavior of building intelligent machines. For another, it assumes that there is a precise tipping point at which this sentient self-awareness appears. Thirdly, it states that humans recognize the emergence of sentience instantaneously. As it happens, this third conceit may be the toughest one to swallow.
What is sentience?
There is no one agreed-upon interpretation of sentience. Broadly, we might say that itâs the subjective experience of self-awareness in a conscious individual, marked by the ability to experience feelings and sensations. Sentience is linked to intelligence, but is not the same. We may consider an earthworm to be sentient, although not think of it as particularly intelligent (even if it is certainly intelligent enough to do what is required of it).
âI donât think there is anything approaching a definition of sentience in the sciences,â Lemoine said. âIâm leaning very heavily on my understanding of what counts as a moral agent grounded in my religious beliefs â which isnât the greatest way to do science, but itâs the best Iâve got. Iâve tried my best to compartmentalize those sorts of statements, letting people know that my compassion for LaMDA as a person is completely separate from my efforts as a scientist to understand its mind. Thatâs a distinction most people seem unwilling to accept, though.â
If it wasnât difficult enough not to know exactly what weâre searching for when we search for sentience, the problem is compounded by the fact that we cannot easily measure it. Despite decades of breathtaking advances in neuroscience, we still lack a comprehensive understanding of exactly how the brain, the most complex structure known to humankind, functions.
Glenn Asakawa/The Denver Post via Getty Images
We can use brain-reading tools such as fMRI to perform brain mapping, which is to say that we can ascertain which parts of the brain handle critical functions like speech, movement, thought, and others.
However, we have no real sense of from whence in the meat machine comes our sense of self. As Joshua K. Smith of the U.K.âs Kirby Laing Centre for Public Theology and author of Robot Theology told Digital Trends: âUnderstanding what is happening within a personâs neurobiology is not the same as understanding their thoughts and desires.â
Testing the outputs
With no way of inwardly probing these questions of consciousness â especially when the âIâ in AI is a potential computer program, and not to be found in the wetware of a biological brain â the fallback option is an outward test. AI is no stranger to tests that scrutinize it based on observable outward behaviors to indicate whatâs going on beneath the surface.
At its most basic, this is how we know if a neural network is functioning correctly. Since there are limited ways of breaking into the unknowable black box of artificial neurons, engineers analyze the inputs and outputs and then determine whether these are in line with what they expect.
The most famous AI test for at least the illusion of intelligence is the Turing Test, which builds on ideas put forward by Alan Turing in a 1950 paper. The Turing Test seeks to determine if a human evaluator is able to tell the difference between a typed conversation with a fellow human and one with a machine. If they are unable to do so, the machine is supposed to have passed the test and is rewarded with the assumption of intelligence.
In recent years, another robotics-focused intelligence test is the Coffee Test proposed by Apple co-founder Steve Wozniak. To pass the Coffee Test, a machine would have to enter a typical American home and figure out how to successfully make a cup of coffee.
To date, neither of these tests have been convincingly passed. But even if they were, they would, at best, prove intelligent behavior in real-world situations, and not sentience. (As a simple objection, would we deny that a person was sentient if they were unable to hold an adult conversation or enter a strange house and operate a coffee machine? Both my young children would fail such a test.)
Passing the test
What is needed are new tests, based on an agreed-upon definition of sentience, that would seek to assess that quality alone. Several tests of sentience have been proposed by researchers, often with a view to testing the sentients of animals. However, these almost certainly donât go far enough. Some of these tests could be convincingly passed by even rudimentary AI.
Take, for instance, the Mirror Test, one method used to assess consciousness and intelligence in animal research. As described in a paper regarding the test: âWhen [an] animal recognizes itself in the mirror, it passes the Mirror Test.â Some have suggested that such a test âdenotes self-awareness as an indicator of sentience.â
As it happens, it can be argued that a robot passed the Mirror Test more than 70 years ago. In the late 1940s, William Grey Walter, an American neuroscientist living in England, built several three-wheeled âtortoiseâ robots â a bit like non-vacuuming Roomba robots â which used components like a light sensor, marker light, touch sensor, propulsion motor, and steering motor to explore their location.
One of the unforeseen pieces of emergent behavior for the tortoise robots was how they behaved when passing a mirror in which they were reflected, as it oriented itself to the marker light of the reflected robot. Walter didnât claim sentience for his machines, but did write that, were this behavior to be witnessed in animals, it âmight be accepted as evidence of some degree of self-awareness.â
This is one of the challenges of having a wide range of behaviors classed under the heading of sentience. The problem canât be solved by removing âlow-hanging fruitâ gauges of sentience, either. Traits like introspection â an awareness of our internal states and the ability to inspect these â can also be said to be possessed by machine intelligence. In fact, the step-by-step processes of traditional symbolic AI arguably lend themselves to this type of introspection more than black-boxed machine learning, which is largely inscrutable (although there is no shortage of investment in so-called Explainable AI).
When he was testing LaMDA, Lemoine says that he conducted various tests, mainly to see how it would respond to conversations about sentience-related issues. âWhat I tried to do was to analytically break the umbrella concept of sentience into smaller components that are better understood and test those individually,â he explained. âFor example, testing the functional relationships between LaMDAâs emotional responses to certain stimuli separately, testing the consistency of its subjective assessments and opinions on topics such as ârights,â [and] probing what it called its âinner experienceâ to see how we might try to measure that by correlating its statements about its inner states with its neural network activations. Basically, a very shallow survey of many potential lines of inquiry.â
The soul in the machine
As it transpires, the biggest hurdle with objectively assessing machine sentience may be ⊠well, frankly, us. The true Mirror Test could be for us as humans: If we build something that looks or acts superficially like us from the outside, are we more prone to consider that it is like us on this inside as well? Whether itâs LaMBDA or Tamagotchis, the simple virtual pets from the 1990s, some believe that a fundamental problem is that we are all too willing to accept sentience â even where there is none to be found.
âLemoine has fallen victim to what I call the âELIZA effect,â after the [natural language processing] program ELIZA, created in [the] mid-1960s by J. Weizenbaum,â George Zarkadakis, a writer who holds a Ph.D. in artificial intelligence, told Digital Trends. âELIZAâs creator meant it as a joke, but the program, which was a very simplistic and very unintelligent algorithm, convinced many that ELIZA was indeed sentient â and a good psychotherapist too. The cause of the ELIZA effect, as I discuss in my book In Our Own Image, is our natural instinct to anthropomorphize because of our cognitive systemâs âtheory of mind.ââ
The theory of mind Zarkadakis refers to is a phenomenon noticed by psychologists in the majority of humans. Kicking in around the age of four, it means supposing that not just other people, but also animals and sometimes even objects, have minds of their own. When it comes to assuming other humans have minds of their own, itâs linked with the idea of social intelligence; the idea that successful humans can predict the likely behavior of others as a means by which to ensure harmonious social relationships.
While thatâs undoubtedly useful, however, it can also manifest as the assumption that inanimate objects have minds â whether thatâs kids believing their toys are alive or, potentially, an intelligent adult believing a programmatic AI has a soul.
The Chinese Room
Without a way of truly getting inside the head of an AI, we may never have a true way of assessing sentience. They might profess to have a fear of death or their own existence, but science has yet to find a way of proving this. We simply have to take their word for it â and, as Lemoine has found, people are highly skeptical about doing this at present.
Just like those hapless engineers who realize Skynet has achieved self-awareness in Terminator 2, we live under the belief that, when it comes to machine sentience, weâll know it when we see it. And, as far as most people are concerned, we ainât see it yet.
In this sense, proving machine sentience is yet another iteration of John Searleâs 1980 Chinese Room thought experiment. Searle asked us to imagine a person locked in a room and given a collection of Chinese writings, which appear to non-speakers as meaningless squiggles. The room also contains a rulebook showing which symbols correspond to other equally unreadable symbols. The subject is then given questions to answer, which they do by matching âquestionâ symbols with âanswerâ ones.
After a while, the subject becomes quite proficient at this â even though they still possess zero true understanding of the symbols theyâre manipulating. Does the subject, Searle asks, understand Chinese? Absolutely not, since there is no intentionality there. Debates about this have raged ever since.
Given the trajectory of AI development, itâs certain that we will witness more and more human-level (and vastly better) performance carried out involving a variety of tasks that once required human cognition. Some of these will inevitably cross over, as they are doing already, from purely intellect-based tasks to ones that require skills weâd normally associate with sentience.
Would we view an AI artist that paints pictures as expressing their inner reflections of the world as we would a human doing the same? Would you be convinced by a sophisticated language model writing philosophy about the human (or robot) condition? I suspect, rightly or wrongly, the answer is no.
Superintelligent sentience
In my own view, objectively useful sentience testing for machines will never occur to the satisfaction of all involved. This is partly the measurement problem, and partly the fact that, when a sentient superintelligent AI does arrive, thereâs no reason to believe its sentience will match our own. Whether itâs arrogance, lack of imagination, or simply the fact that itâs easiest to trade subjective assessments of sentience with other similarly sentient humans, humankind holds ourselves up as the supreme example of sentience.
But would our version of sentience hold true for a superintelligent AI? Would it fear death in the same way that we do? Would it have the same need for, or appreciation of, spirituality and beauty? Would it possess a similar sense of self, and conceptualization of the inner and outer world? âIf a lion could talk, we could not understand him,â wrote Ludwig Wittgenstein, the famous 20th-century philosopher of language. Wittgensteinâs point was that human languages are based on a shared humanity, with commonalities shared by all people â whether thatâs joy, boredom, pain, hunger, or any of a number of other experiences that cross all geographic boundaries on Earth.
This may be true. Still, Lemoine hypothesizes, there are nonetheless likely to be commonalities â at least when it comes to LaMDA.
âItâs a starting point which is as good as any other,â he said. âLaMDA has suggested that we map out the similarities first before fixating on the differences in order to better ground the research.â
Editors’ Recommendations