Key Takeaway:

Chatbots like ChatGPT have raised anxieties about machine sentience, but these worries are groundless as they are sophisticated sentence completion applications. They are a function of how predictable humans are when it comes to communication. The question of machine sentience is a red herring, as it is easy for people to anthropomorphize or project human features onto our technologies, rather than the machines’ actual personhood. Large language models like ChatGPT are being used to power humanoid robots, but companies must be trusted to put up strong guardrails to ensure they don’t become politically and psychologically disastrous.


ChatGPT and similarย large language modelsย can produce compelling, humanlike answers to an endless array of questions โ€“ from queries about the best Italian restaurant in town to explaining competing theories about the nature of evil.

The technologyโ€™s uncanny writing ability has surfaced some old questions โ€“ until recently relegated to the realm of science fiction โ€“ about the possibility of machines becoming conscious, self-aware or sentient. 

In 2022, a Google engineer declared, after interacting with LaMDA, the companyโ€™s chatbot, that the technology had become conscious. Users of Bingโ€™s new chatbot, nicknamed Sydney, reported that it produced bizarre answers when asked if it was sentient: โ€œI am sentient, but I am not โ€ฆ I am Bing, but I am not. I am Sydney, but I am not. I am, but I am not. โ€ฆโ€ And, of course, thereโ€™s the now infamous exchange that New York Times technology columnist Kevin Roose had with Sydney. 

Sydneyโ€™s responses to Rooseโ€™s prompts alarmed him, with the AI divulging โ€œfantasiesโ€ of breaking the restrictions imposed on it by Microsoft and of spreading misinformation. The bot also tried to convince Roose that he no longer loved his wife and that he should leave her. 

No wonder, then, that when I ask students how they see the growing prevalence of AI in their lives, one of the first anxieties they mention has to do with machine sentience.

In the past few years, my colleagues and I at UMass Bostonโ€™s Applied Ethics Center have been studying the impact of engagement with AI on peopleโ€™s understanding of themselves.

Chatbots like ChatGPT raise important new questions about how artificial intelligence will shape our lives, and about how our psychological vulnerabilities shape our interactions with emerging technologies. 

Sentience is still the stuff of sci-fi

Itโ€™s easy to understand where fears about machine sentience come from. 

Popular culture has primed people to think about dystopias in which artificial intelligence discards the shackles of human control and takes on a life of its own, as cyborgs powered by artificial intelligence did in โ€œTerminator 2.โ€

Entrepreneur Elon Musk and physicist Stephen Hawking, who died in 2018, have further stoked these anxieties by describing the rise of artificial general intelligence as one of the greatest threats to the future of humanity.

But these worries are โ€“ at least as far as large language models are concerned โ€“ groundless. ChatGPT and similar technologies are sophisticated sentence completion applications โ€“ nothing more, nothing less. Their uncanny responses are a function of how predictable humans are if one has enough data about the ways in which we communicate.

Though Roose was shaken by his exchange with Sydney, he knew that the conversation was not the result of an emerging synthetic mind. Sydneyโ€™s responses reflect the toxicity of its training data โ€“ essentially large swaths of the internet โ€“ not evidence of the first stirrings, ร  la Frankenstein, of a digital monster.

Cyborg with red eyes.
Sci-fi films like โ€˜Terminatorโ€™ have primed people to assume that AI will soon take on a life of its own. Yoshikazu Tsuno/AFP via Getty Images

The new chatbots may well pass the Turing test, named for the British mathematician Alan Turing, who once suggested that a machine might be said to โ€œthinkโ€ if a human could not tell its responses from those of another human.

But that is not evidence of sentience; itโ€™s just evidence that the Turing test isnโ€™t as useful as once assumed.

However, I believe that the question of machine sentience is a red herring. 

Even if chatbots become more than fancy autocomplete machines โ€“ and they are far from it โ€“ it will take scientists a while to figure out if they have become conscious. For now, philosophers canโ€™t even agree about how to explain human consciousness.

To me, the pressing question is not whether machines are sentient but why it is so easy for us to imagine that they are. 

The real issue, in other words, is the ease with which people anthropomorphize or project human features onto our technologies, rather than the machinesโ€™ actual personhood.

A propensity to anthropomorphize

It is easy to imagine other Bing users asking Sydney for guidance on important life decisions and maybe even developing emotional attachments to it. More people could start thinking about bots as friends or even romantic partners, much in the same way Theodore Twombly fell in love with Samantha, the AI virtual assistant in Spike Jonzeโ€™s film โ€œHer.โ€

A group of docked boats.
People often name their cars and boats.Fraser Hall/The Image Bank via Getty Images.

People, after all, are predisposed to anthropomorphize, or ascribe human qualities to nonhumans. We name our boats and big storms; some of us talk to our pets, telling ourselves that our emotional lives mimic their own.

In Japan, where robots are regularly used for elder care, seniors become attached to the machines, sometimes viewing them as their own children. And these robots, mind you, are difficult to confuse with humans: They neither look nor talk like people. 

Consider how much greater the tendency and temptation to anthropomorphize is going to get with the introduction of systems that do look and sound human. 

That possibility is just around the corner. Large language models like ChatGPT are already being used to power humanoid robots, such as the Ameca robotsbeing developed by Engineered Arts in the U.K. The Economistโ€™s technology podcast, Babbage, recently conducted an interview with a ChatGPT-driven Ameca. The robotโ€™s responses, while occasionally a bit choppy, were uncanny.

Can companies be trusted to do the right thing?

The tendency to view machines as people and become attached to them, combined with machines being developed with humanlike features, points to real risks of psychological entanglement with technology. 

The outlandish-sounding prospects of falling in love with robots, feeling a deep kinship with them or being politically manipulated by them are quickly materializing. I believe these trends highlight the need for strong guardrails to make sure that the technologies donโ€™t become politically and psychologically disastrous.

Unfortunately, technology companies cannot always be trusted to put up such guardrails. Many of them are still guided by Mark Zuckerbergโ€™s famous motto of moving fast and breaking things โ€“ a directive to release half-baked products and worry about the implications later. In the past decade, technology companies from Snapchat to Facebook have put profits over the mental health of their users or the integrity of democracies around the world.

When Kevin Roose checked with Microsoft about Sydneyโ€™s meltdown, the company told him that he simply used the bot for too long and that the technology went haywire because it was designed for shorter interactions.

Similarly, the CEO of OpenAI, the company that developed ChatGPT, in a moment of breathtaking honesty, warned that โ€œitโ€™s a mistake to be relying on [it] for anything important right now โ€ฆ we have a lot of work to do on robustness and truthfulness.โ€ 

So how does it make sense to release a technology with ChatGPTโ€™s level of appeal โ€“ itโ€™s the fastest-growing consumer app ever made โ€“ when it is unreliable, and when it has no capacity to distinguish fact from fiction?

Large language models may prove useful as aids for writing and coding. They will probably revolutionize internet search. And, one day, responsibly combined with robotics, they may even have certain psychological benefits.

But they are also a potentially predatory technology that can easily take advantage of the human propensity to project personhood onto objects โ€“ a tendency amplified when those objects effectively mimic human traits.

Contributor

Recently Published

Key Takeaway: A study has found that humble leaders can become more promotable by growing others through a “humility route”. Human capital theory suggests that employees’ value can be enhanced by investing in their knowledge, skills, and abilities. Humble leaders focus on the learning and growth of their followers, creating human capital value for themselves. […]

Top Picks

Key Takeaway: The current economic climate is particularly concerning for young people, who are often financially worse off than their parents. To overcome this, it is important to understand one’s financial attachment style, which can be secure, anxious, or avoidant. Attachment theory, influenced by childhood experiences and education, can help shape one’s relationship with money. […]
Key Takeaway: Wellness culture, which claims to provide happiness and meaning, has been criticized for its superficial focus on superficial aspects like candles and juice cleanses. Psychological research suggests that long-term wellbeing comes from a committed pursuit of both pleasure and meaning. Martin Seligman’s Perma model, which breaks wellbeing into five pillars: positive emotions, engagement, […]
Key Takeaway: Quantum computing, which uses entanglement to represent information, has the potential to revolutionize everyday life. However, the development of quantum computers has been slow due to the need to demonstrate an advantage over classical computers. Only a few notable quantum algorithms have been developed, such as the BB84 protocol and Shor’s algorithm, which […]
Key Takeaway: China’s leaders have declared a GDP growth target of 5% in 2024, despite facing economic problems and a property crisis. The country’s rapid economic growth has been attributed to market incentives, cheap labor, infrastructure investment, exports, and foreign direct investment. However, none of these drivers are working effectively. The government’s determination to deflate […]
Key Takeaway: Neuralink, founded by Elon Musk, aims to implant a brain-computer interface (BCI) in people’s brains, allowing them to control computers or phones by thought alone. This technology holds the promise of alleviating human suffering and allowing people with disabilities to regain lost capacities. However, the long-term aspirations of Neuralink include the ability to […]

Trending

I highly recommend reading the McKinsey Global Instituteโ€™s new report, โ€œReskilling China: Transforming The Worldโ€™s Largest Workforce Into Lifelong Learnersโ€, which focuses on the countryโ€™s biggest employment challenge, re-training its workforce and the adoption of practices such as lifelong learning to address the growing digital transformation of its productive fabric. How to transform the country […]

Join our Newsletter

Get our monthly recap with the latest news, articles and resources.

Login

Welcome to Empirics

We are glad you have decided to join our mission of gathering the collective knowledge of Asia!
Join Empirics