Emotion concepts and their function in a large language model
by dnw on 4/4/2026, 6:30:25 AM
https://www.anthropic.com/research/emotion-concepts-function
Comments
by: kirykl
The technology they are discovering is called "Language". It was designed to encode emotions by a sender and invoke emotions in the reader. The emotions a reader gets from LLM are still coming from the language
4/4/2026, 8:06:38 AM
by: comrade1234
There was a really old project from mit called conceptnet that I worked with many years ago. It was basically a graph of concepts (not exactly but close enough) and emotions came into it too just as part of the concepts. For example a cake concept is close to a birthday concept is close to a happy feeling.<p>What was funny though is that it was trained by MIT students so you had the concept of getting a good grade on a test as a happier concept than kissing a girl for the first time.<p>Another problem is emotions are cultural. For example, emotions tied to dogs are different in different cultures.<p>We wanted to create concept nets for individuals - that is basically your personality and knowledge combined but the amount of data required was just too much. You'd have to record all interactions for a person to feed the system.
4/4/2026, 7:58:57 AM
by: emoII
Super interesting, I wonder if this research will cause them to actually change their llm, like turning down the ”desperation neurons” to stop Claude from creating implementations for making a specific tests pass etc.
4/4/2026, 7:06:33 AM
by: yoaso
The desperation > blackmail finding stuck with me. If AI behavior shifts based on emotional states, maybe emotions are just a mechanism for changing behavior in the first place. If we think of human emotions the same way, just evolution's way of nudging behavior, the line between AI and humans starts to look a lot thinner.
4/4/2026, 7:55:05 AM
by: whatever1
So should I go pursue a degree in psychology and become a datacenter on-call therapist?
4/4/2026, 7:56:01 AM
by: Chance-Device
> Note that none of this tells us whether language models actually feel anything or have subjective experiences.<p>You’ll never find that in the human brain either. There’s the machinery of neural correlates to experience, we never see the experience itself. That’s likely because the distinction is vacuous: they’re the same thing.
4/4/2026, 7:36:00 AM
by: trhway
>... emotion-related representations that shape its behavior. These specific patterns of artificial “neurons” which activate in situations—and promote behaviors—that the model has learned to associate with the concept of a particular emotion. .... In contexts where you might expect a certain emotion to arise for a human, the corresponding representations are active.<p>>For instance, to ensure that AI models are safe and reliable, we may need to ensure they are capable of processing emotionally charged situations in healthy, prosocial ways.<p>Force-set to 0, "mask"/deactivate those representations associated with bad/dangerous emotions. Neural Prozac/lobotomy so to speak.
4/4/2026, 8:29:19 AM
by: mci
The first and second principal components (joy-sadness and anger) explain only 41% of the variance. I wish the authors showed further principal components. Even principal components 1-4 would explain no more than 70% of the variance, which seems to contradict the popular theory that all human emotions are composed of 5 basic emotions: joy, sadness, anger, fear, and disgust, i.e. 4 dimensions.
4/4/2026, 7:22:23 AM
by: koolala
A-HHHHHHHHHHHHHHHJ
4/4/2026, 8:02:21 AM
by: idiotsecant
Its almost like LLMs have a vast, mute unconscious mind operating in the background, modeling relationships, assigning emotional state, and existing entirely without ego.<p>Sounds sort of like how certain monkey creatures might work.
4/4/2026, 7:17:25 AM
by: techpulselab
[dead]
4/4/2026, 8:13:52 AM
by: ActorNightly
[dead]
4/4/2026, 7:26:03 AM