Hacker News Viewer

Emotion concepts and their function in a large language model

by dnw on 4/4/2026, 6:30:25 AM

https://www.anthropic.com/research/emotion-concepts-function

Comments

by: kirykl

The technology they are discovering is called "Language". It was designed to encode emotions by a sender and invoke emotions in the reader. The emotions a reader gets from LLM are still coming from the language

4/4/2026, 8:06:38 AM


by: comrade1234

There was a really old project from mit called conceptnet that I worked with many years ago. It was basically a graph of concepts (not exactly but close enough) and emotions came into it too just as part of the concepts. For example a cake concept is close to a birthday concept is close to a happy feeling.<p>What was funny though is that it was trained by MIT students so you had the concept of getting a good grade on a test as a happier concept than kissing a girl for the first time.<p>Another problem is emotions are cultural. For example, emotions tied to dogs are different in different cultures.<p>We wanted to create concept nets for individuals - that is basically your personality and knowledge combined but the amount of data required was just too much. You&#x27;d have to record all interactions for a person to feed the system.

4/4/2026, 7:58:57 AM


by: emoII

Super interesting, I wonder if this research will cause them to actually change their llm, like turning down the ”desperation neurons” to stop Claude from creating implementations for making a specific tests pass etc.

4/4/2026, 7:06:33 AM


by: yoaso

The desperation &gt; blackmail finding stuck with me. If AI behavior shifts based on emotional states, maybe emotions are just a mechanism for changing behavior in the first place. If we think of human emotions the same way, just evolution&#x27;s way of nudging behavior, the line between AI and humans starts to look a lot thinner.

4/4/2026, 7:55:05 AM


by: whatever1

So should I go pursue a degree in psychology and become a datacenter on-call therapist?

4/4/2026, 7:56:01 AM


by: Chance-Device

&gt; Note that none of this tells us whether language models actually feel anything or have subjective experiences.<p>You’ll never find that in the human brain either. There’s the machinery of neural correlates to experience, we never see the experience itself. That’s likely because the distinction is vacuous: they’re the same thing.

4/4/2026, 7:36:00 AM


by: trhway

&gt;... emotion-related representations that shape its behavior. These specific patterns of artificial “neurons” which activate in situations—and promote behaviors—that the model has learned to associate with the concept of a particular emotion. .... In contexts where you might expect a certain emotion to arise for a human, the corresponding representations are active.<p>&gt;For instance, to ensure that AI models are safe and reliable, we may need to ensure they are capable of processing emotionally charged situations in healthy, prosocial ways.<p>Force-set to 0, &quot;mask&quot;&#x2F;deactivate those representations associated with bad&#x2F;dangerous emotions. Neural Prozac&#x2F;lobotomy so to speak.

4/4/2026, 8:29:19 AM


by: mci

The first and second principal components (joy-sadness and anger) explain only 41% of the variance. I wish the authors showed further principal components. Even principal components 1-4 would explain no more than 70% of the variance, which seems to contradict the popular theory that all human emotions are composed of 5 basic emotions: joy, sadness, anger, fear, and disgust, i.e. 4 dimensions.

4/4/2026, 7:22:23 AM


by: koolala

A-HHHHHHHHHHHHHHHJ

4/4/2026, 8:02:21 AM


by: idiotsecant

Its almost like LLMs have a vast, mute unconscious mind operating in the background, modeling relationships, assigning emotional state, and existing entirely without ego.<p>Sounds sort of like how certain monkey creatures might work.

4/4/2026, 7:17:25 AM


by: techpulselab

[dead]

4/4/2026, 8:13:52 AM


by: ActorNightly

[dead]

4/4/2026, 7:26:03 AM