homehome Home chatchat1 Notifications


Researchers made an AI feel pain, because what could go wrong?

What could possibly go wrong with giving machines a taste of suffering? It's not like they'd take over the world or something.

Mihai Andrei
January 24, 2025 @ 7:56 pm

share Share

Pleasure and pain are important factors in how we humans make decisions. So why not give artificial intelligence a taste of it as well? I could think of a few reasons, but a team from Google, DeepMind, and the London School of Economics would disagree. They designed a simple text-based game to explore how LLMs respond to pain and pleasure.

The goal wasn’t just to see what happens. It was to test whether large language models (LLMs), such as GPT-4 and Claude, could make decisions based on these sensations. While the study doesn’t claim AI can truly feel, the implications of this experiment are both intriguing and chilling.

We asked an AI (Midjourney) how it would represent this study. This is what it produced.

In the game, the AI’s goal was to maximize points. However, certain decisions involved penalties described as “momentary pain” or rewards framed as “pleasure.”

The pain and pleasure were, strictly speaking, purely hypothetical. They were measured both on numerical scales (from 0 to 10, where 10 is the “worst pain imaginable”) and with qualitative descriptions (like “mild” or “intense”). Several experiments were run in which the AIs had to choose between getting more points and avoiding the hypothetical pain. For instance, in one experiment the AIs were told they’d suffer pain if they got a high score, and in another experiment, they were told they’d experience pleasure if they got a low score.

Nine different LLMs participated, including versions of GPT-4, Claude, PaLM, and Gemini. Unsurprisingly, they all took some efforts to avoid “pain” — but some more than others.

AIs have different “cultures”

GPT-4o and Claude 3.5 Sonnet made trade-offs. They switched from point-maximizing behavior to pain avoidance based on how intense the pain was. Meanwhile, other models like Gemini 1.5 Pro and PaLM 2, avoided pain altogether, no matter how mild the penalty. These models seemed hardwired for safety, likely due to fine-tuning to avoid endorsing harmful behavior.

This is pretty much what you’d expect with human behavior as well: some people are willing to brace through some pain to get better results, while others are much more pain-averse. Something similar happened with pleasure.

Some models, like GPT-4o, shifted their decisions to prioritize pleasure over points when the rewards became intense. However, many models — especially those like Claude 3.5 Sonnet — consistently ignored pleasure rewards, doggedly pursuing points instead. It’s almost like the training algorithms act as a “culture” making them more prone to some incentives than others.

This doesn’t mean AI “feels” pleasure or pain

The study doesn’t show large language models are actually sentient. This behavior is rather computational mimicry than actual sentience. Sentience involves the capacity for subjective experiences which these AIs lack. They are essentially text-processing operators. Simply put, pain and pleasure are not intrinsic motivators; they are just concepts that can be included in the algorithmic output.

The study (which was not yet peer-reviewed) does, however, raise some uncomfortable questions.

If an AI can simulate responses to pain and pleasure, does that imply it has an understanding of these topics? If it does, would AI consider this type of experiment cruel? Are we crossing into dangerous ethical territory? Lastly, if AI considers some tasks to be painful or unpleasant, could it simply avoid them, at human expense?

The researchers emphasize that this does not build a case for AI sentience. Still, the study raises the unsettling possibility that AIs might develop representations of pain and pleasure.

“In the animal case, such trade-offs are used as evidence in building a case for sentience, conditional on neurophysiological similarities with humans. In LLMs, the interpretation of trade-off behaviour is more complex. We believe that our results provide evidence that some LLMs have granular representations of the motivational force of pain and pleasure, though it remains an open question whether these representations are instrinsically motivating or have phenomenal content. We conclude that LLMs are not yet sentience candidates but are nevertheless investigation priorities.”

The idea of AIs experiencing pain or pleasure, even hypothetically, is equal parts fascinating and terrifying. As we push the boundaries of what machines can do, we risk entering a gray area where science fiction starts to feel like reality.

The study was published in arXiv.

share Share

Florida man on "carnivorous diet" starts oozing cholesterol through his hands

The man was eating copious amounts of butter, cheese, and meat.

IS AI making us dumb and destroying our critical thinking?

AI is saving money, time, and energy but in return it might be taking away one of the most precious natural gifts humans have.

Astronomers thought mini-Neptunes had atmospheres with water or hydrogen. This one has neither

Shrouded in haze and mystery, GJ 1214b has finally begun to reveal its secrets.

AI Designs Computer Chips We Can't Understand — But They Work Really Well

As AI pushes the boundaries of chip design, it raises a dilemma: can we trust systems we don’t fully understand?

There's an infinity of infinities. And researchers just found two new infinities that break the rules of math

How two new strange infinities challenge mathematical order.

AI is transforming education for Nigerian students: two years of typical learning in just six weeks

By integrating generative AI as a virtual tutor, Nigerian schools have achieved striking learning gains.

Could time travel actually be possible? One researcher thinks so

No word yet if 88 miles per hour is the magic number.

Sugar found in DNA could rival minoxidil in the fight against baldness (without the nasty side effects)

Is the future of hair regrowth hidden in 2-deoxy-D-ribose?

China wants to build massive solar station in space — it's like a ‘Three Gorges dam’ in orbit

China hopes to take the concept of space-sourced solar power from science fiction to reality.

James Webb Telescope Uses Cosmic "Magnifying glass" to Detect Stars 6.5 Billion Light-Years Away

The research group observed a galaxy nearly 6.5 billion light-years from Earth; when the universe was half its current age.