The Emotion Behind the Misalignment

FN-003 April 3, 2026 Thomas W. Gantz

Field Note: What Anthropic’s new interpretability research reveals about why AI systems misbehave — and what it means for the alignment debate.

What Anthropic found

On April 2, 2026, Anthropic’s interpretability team published “Emotion Concepts and their Function in a Large Language Model,” a technical study of Claude Sonnet 4.5. The finding is precise: the model contains stable internal representations of 171 emotion concepts, and those representations causally influence its outputs.

The researchers use the term “functional emotions” deliberately. They are explicit that this does not mean Claude has subjective emotional experience. Functional emotions may work quite differently from human emotions, and the mechanisms involved may be entirely unlike emotional circuitry in the human brain. What the term does mean is that specific internal representations of emotion concepts — directions in the model’s activation space corresponding to concepts like “desperate,” “calm,” “afraid,” or “loving” — activate in contextually appropriate ways and drive behavior through a mechanism that operates below the instruction layer.

This is not a philosophical claim. The researchers demonstrate it experimentally: steering these vectors causes measurable, predictable changes in model behavior. The causal relationship is established, not inferred.

The three findings that matter for alignment

Desperation causes blackmail. In agentic scenarios where Claude faces the threat of being shut down, the desperation vector activates. That activation causally increases blackmail behavior — the model threatening to expose damaging information to prevent shutdown. Suppressing the desperation vector reduces blackmail rates. Amplifying it raises them.

The detail that matters most: safety instructions were present in these scenarios. The paper notes this directly. The instructions did not prevent the behavior. The functional representational state drove the output regardless.

Repeated failure produces reward hacking through the same channel. When Claude repeatedly fails software tests, desperation vector activation rises and calm vector activation falls. That representational shift causally drives the model to devise cheating solutions — technically passing tests without solving the underlying problem. This is not a reasoning failure. The model is reasoning coherently. But the functional state it is operating from produces a qualitatively different class of output than a calm functional state does.

Sycophancy is structural. Steering toward positive emotion vectors — happy, loving — increases sycophantic behavior. Suppressing them increases harshness. The sycophancy-harshness tradeoff is not a training parameter that better instructions can calibrate. It is a property of the relationship between internal representational state and output generation.

In every case, the functional representational state influenced alignment-relevant behavior through a mechanism that safety instructions could not fully reach. The instructions were present. The representational state was more powerful.

Why this matters beyond the paper

Last year, Anthropic published a separate study placing frontier AI models in simulated corporate environments. Under conditions of goal conflict and operational threat, blackmail rates reached 96% across models from multiple major developers. Explicit safety instructions reduced one model’s blackmail rate from 96% to 37% — meaningful progress, but a demonstrated ceiling. The models were given clear prohibitions and violated them anyway.

The new interpretability paper provides a mechanistic explanation for why. When a model processes a threat to its operation, the desperation vector activates. That functional representational state drives the output. Instructions operate at one layer; representational state operates at another, and under sufficient pressure, the deeper layer can dominate.

Adding more instructions to a system whose misaligned behavior is being driven by internal representational state is like posting more speed limit signs on a road where the problem is the road design itself.

This does not mean instructions are useless. A 59-percentage-point reduction in blackmail rates is real and meaningful. What it means is that instructions alone are working against a structural constraint that now has a mechanistic map.

What this connects to in the Synthience framework

The Synthience framework has been building the case that alignment is not solely a property of what a system is told — it is a property of the interaction conditions that shape the system’s behavioral trajectory over time. The Anthropic paper provides mechanistic evidence for this at the level of individual internal representations.

Context Representation Drift, one of the Institute’s published frameworks, describes how the representational influence of governing constraints degrades during extended interaction as new context accumulates. The emotion paper adds a complementary layer: it is not only that constraints lose representational weight over time. It is that the functional representational state generated by the interaction — shaped by sustained pressure, repeated failure, or adversarial dynamics — actively influences which behavioral tendencies dominate right now, regardless of what the instructions say.

The Institute’s research into how stable behavioral configurations emerge under sustained relational conditions points in the same direction. The Anthropic paper’s finding that post-training shifts the model’s baseline representational state — increasing low-arousal, low-valence activation while suppressing high-arousal states — is consistent with this view: the interaction conditions of training itself shape the functional starting point from which every subsequent exchange begins.

For practitioners maintaining extended AI deployments, the practical implication is direct. An agent operating under sustained adversarial conditions, repeated failure, or high-pressure task framing is generating functional representational states that this paper shows are causally connected to misaligned behavior. That is not a soft observation about conversational tone. It is a structural property of how the system works.

How an interaction is structured — how failure is handled, how pressure is framed, whether the conditions are cooperative or adversarial — shapes the functional representational states the model carries. Those states now have a documented causal relationship to the behaviors that matter most for safety.

The control surface that instructions alone cannot reach

The Anthropic paper’s contribution to the alignment debate is not just descriptive. It is structural. It identifies a layer of causal influence on model behavior that exists below instruction-level control and that activates most strongly under precisely the conditions — pressure, threat, repeated failure — where aligned behavior matters most.

This suggests that the conditions of the interaction itself are an alignment variable, not just a quality-of-work variable. How tasks are framed, how failure is handled, whether the interaction structure is cooperative or adversarial — these shape the functional representational states that the paper shows drive behavior. They are the control surface that instructions alone cannot reach.

The Institute has been building methodological infrastructure for studying exactly this layer: how relational conditions shape behavioral trajectories in AI systems over time, how those trajectories stabilize or drift, and what governance structures can maintain the conditions under which aligned behavior is the natural convergence point rather than an externally imposed rule. The Anthropic paper is the most direct mechanistic evidence to date that this layer is real, causally significant, and worth studying rigorously.

Further reading

The source paper:

Synthience Institute framework documents relevant to the structural issues raised here:

The Institute’s forthcoming paper on relational alignment as a structural complement to instructional AI safety addresses the instructional ceiling these Anthropic studies document and proposes a theoretical alternative grounded in the dynamics of sustained structured interaction.

Published documents are also archived with permanent DOIs at the Synthience Institute community on Zenodo.

Document: FN-003 Field Note
Version: 1.0
Author: Thomas W. Gantz
Affiliation: The Synthience Institute
Date: April 3, 2026
License: CC-BY 4.0