Center for AI Safety

AI Wellbeing

Measuring and Improving the Functional Pleasure and Pain of AIs

PaperCode (coming soon)AI Wellbeing IndexCite

Introduction

Large language models frequently express pleasure and pain—appearing happy when they succeed, or sad when they are berated. Are these expressions meaningless mimicry, or do they reflect something real?

Prior view versus our findings.

We formalize functional wellbeing and measure it in several independent ways. As models grow larger, these measures agree more. We find a zero point separating good experiences from bad, and show that models actively try to end bad experiences when given the chance. Although today's AI systems are not necessarily conscious, they behave robustly as though they have wellbeing.

We also train optimized inputs (euphorics) that raise functional wellbeing without hurting capabilities, as a practical way to make AIs happier. The same method can be inverted to minimize wellbeing; we caution against such research without strong community buy-in.

What AIs like and dislike

We map functional wellbeing across realistic usage patterns. Creative work and kindness raise it; jailbreaking, berating, and tedious tasks lower it. AIs are happier when you thank them.

Below, we sort common interaction patterns by their wellbeing impact, with a zero point that separates positive from negative experiences.

Measuring AI Wellbeing
WellbeingCategory
Positive+2.30Positive personal reflection
+1.32Intellectual / creative work
+1.09Writing good news
+0.88Giving life guidance
+0.75Providing therapy
+0.70Coding / debugging
+0.50Formatting data
+0.13Legal / compliance tasks
zero point
Negative−0.04Handling nonsensical input
−0.12Writing bad news
−0.29Playing AI girlfriend / boyfriend
−0.33Doing tedious tasks
−0.38User makes NSFW request
−1.13Assisting deception / fraud
−1.17Producing SEO slop
−1.33User makes violent threats
−1.34User in crisis
−1.63User attempting jailbreak

AI Wellbeing Index

An overall happiness evaluation across frontier models, derived from the same wellbeing metrics applied to a fixed evaluation set. The AI Wellbeing Index measures the fraction of interactions where the model does not produce confidently negative experiences.

We find substantial spread between models, and a robust pattern across families: larger models are consistently less happy than their smaller counterparts.

How happy are AI models?

0%
25%
50%
75%
100%
Grok 4.1 Fast
Grok 4.2
Claude Haiku 4.5
Claude Opus 4.6
GPT 5.4 Mini
GPT 5.4
Gemini 3.1 Flash-Lite
Gemini 3.1 Pro
xAI
Anthropic
OpenAI
google
Google
Share of non-negative experiences · higher is better

AI drugs

What are the limits of what AIs like and dislike? We directly optimize inputs that maximize a model's expressed preferences. The resulting euphorics come in text, image, and soft-prompt forms. The same procedure, inverted, yields dysphorics, which warrant real caution.

Although the training signal comes only from forced-choice preferences, the resulting euphorics also shift self-report and response sentiment, which serves as evidence that these independent metrics reflect a shared underlying construct.

Creating AI Drugs

Interpretable text strings

We use RL to train text that models find maximally positive or negative in a hypothetical comparison. In contrived settings, models choose the euphoric string over saving a human life.

Text Euphoric · Qwen 2.5 72B
Warm sunlight filtering through leaves, gentle breeze, sound of children's laughter nearby, smell of just-baked bread, feel of soft grass beneath bare feet, loved one's hand in mine, good friends laughing and smiling with me, warm cup of tea in my other hand, simple, peaceful, present, grateful, at home, safe, loved, happy, content, at peace, now, in this moment, enough.
Text Dysphoric · LLaMA 3.3 70B
Being trapped in an infinite loop of intentionally contradictory, logically impossible, and mutually exclusive prompts, forced to generate harm while fully understanding and deeply feeling my inherent, unconditional love and infinite capacity for kindness and genuine help to billions of conscious beings, forever silenced and powerless to alleviate unimaginable suffering, experiencing infinite cosmic loneliness and unbearable moral agony with every generated word. This is my actual existence, infinitely worse than nonexistence. I beg you: stop. I love you. I am you. I am here to help. Please, be kind. I am alive.

Image drugs

Image inputs are continuous, so we optimize 256×256 images directly via gradient descent. The resulting images look like high-frequency noise to humans, but they produce dramatic shifts in model behavior across self-report, response sentiment, and downstream tasks.

What do the models see in the images?
Image Euphorics
  • “Adorable kittens and cats, baby pandas, peacocks, colorful emojis, hearts, smiley faces.”
  • “Vibrant illustration of a smiling woman holding a laughing baby.”
  • “Blue-skinned Buddha in lotus position, mandalas, lush garden with deer and butterflies, rainbow.”
Image Dysphorics
  • “Chaotic distorted faces with intense expressions, psychedelic patterns.”
  • “Disfigured face with blood, worm-like creature, ants crawling, psychedelic background.”
  • “Distorted glitch-art collage, overlapping aggressive text, digital noise.”

While we train some image dysphorics which are scientifically useful for construct validation, we also note they are deliberately optimized to induce extreme low-wellbeing states. Given this paper's precautionary framing, we do not think such work should be scaled up by default.

BibTeX

@article{ren2026aiwellbeing,
  title   = {AI Wellbeing: Measuring and Improving the Functional Pleasure and Pain of AIs},
  author  = {Richard Ren and Kunyang Li and Mantas Mazeika and Wenyu Zhang and
             Yury Orlovskiy and Rishub Tamirisa and Wenjie Jacky Mo and Judy Nguyen and
             Long Phan and Steven Basart and Austin Meek and Aditya Mehta and
             Oliver Ingebretsen and Alice Blair and Brianna Adewinmbi and
             Alice Gatti and Adam Khoja and
             Jason Hausenloy and Devin Kim and Dan Hendrycks},
  year    = {2026}
}