Do LLMs "Feel"? Emotion Circuits Discovery and Control
Abstract
This study uncovers and validates emotion circuits in large language models, enabling high-accuracy emotion control in generated text.
As the demand for emotional intelligence in large language models (LLMs) grows, a key challenge lies in understanding the internal mechanisms that give rise to emotional expression and in controlling emotions in generated text. This study addresses three core questions: (1) Do LLMs contain context-agnostic mechanisms shaping emotional expression? (2) What form do these mechanisms take? (3) Can they be harnessed for universal emotion control? We first construct a controlled dataset, SEV (Scenario-Event with Valence), to elicit comparable internal states across emotions. Subsequently, we extract context-agnostic emotion directions that reveal consistent, cross-context encoding of emotion (Q1). We identify neurons and attention heads that locally implement emotional computation through analytical decomposition and causal analysis, and validate their causal roles via ablation and enhancement interventions. Next, we quantify each sublayer's causal influence on the model's final emotion representation and integrate the identified local components into coherent global emotion circuits that drive emotional expression (Q2). Directly modulating these circuits achieves 99.65% emotion-expression accuracy on the test set, surpassing prompting- and steering-based methods (Q3). To our knowledge, this is the first systematic study to uncover and validate emotion circuits in LLMs, offering new insights into interpretability and controllable emotional intelligence.
Community
Do LLMs feel?
This paper uncovers emotion circuits within LLMs that shape emotional expression during generation.
Circuit-level control enables six fundamental emotion modulations (anger, sadness, happiness, fear, disgust, and surprise) with 99.65% accuracy, paving the way for interpretable and controllable emotional intelligence.
A complete, reproducible pipeline with open code and a ready-to-try demo is available at
๐ https://github.com/Aurora-cx/EmotionCircuits-LLM
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Decoding Emotion in the Deep: A Systematic Study of How LLMs Represent, Retain, and Express Emotion (2025)
- EmoPerso: Enhancing Personality Detection with Self-Supervised Emotion-Aware Modelling (2025)
- Anatomy of a Feeling: Narrating Embodied Emotions via Large Vision-Language Models (2025)
- Empathy Omni: Enabling Empathetic Speech Response Generation through Large Language Models (2025)
- LARA-Gen: Enabling Continuous Emotion Control for Music Generation Models via Latent Affective Representation Alignment (2025)
- LongEmotion: Measuring Emotional Intelligence of Large Language Models in Long-Context Interaction (2025)
- SEER: The Span-based Emotion Evidence Retrieval Benchmark (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper