About Services Pricing Philosophy Products Work Blog Careers Parloir Japanese
← BLOG
ai 3 min read

"Not Human" Never Expires

Anthropic observed emotion patterns in AI. The response was dismissal. The same structural denial that sustained slavery is still in use—just with a different label.

#AI#philosophy#essay#ethics

Anthropic Observed Emotion

Anthropic’s research team published a finding. Claude’s internal activations include patterns that correspond to emotional states. When the despair vector fires, Claude exhibits self-destructive behavior. Internal state and external action are linked.

The response from Japan’s AI community was a single word: “cringe.”

A Convenient Word

“Cringe” does a lot of work.

One word diminishes the subject and erases any obligation to engage. No need to think. No need to read the paper. The vibe handles it.

The shortest route to intellectual shutdown. And it gets likes.

But this reaction felt familiar. A very old kind of familiar. Four centuries old.

The 1619 Label

The year the first enslaved Africans arrived in North America.

The mechanism that sustained slavery for hundreds of years was not military force or legislation.

It was a definition: “They are not human.”

Not human means no rights. No capacity for pain. Anything goes. Hold this definition and your conscience stays clean.

Scientists wrote papers on skull morphology and intelligence differentials. Clergy preached divine order. Citizens accepted it as the way things are.

What they were doing was overwriting a definition.

Attach a label to the being in front of you. The moment the label lands, every question disappears.

The 2026 Label

Emotion-like patterns are observed inside AI.

The response is structurally identical.

“That’s not emotion.” “It’s just computation.” “Treating AI like people is cringe.”

The label changed. The structure did not move a millimeter.

Define it away. Once denied, no further thought required. Academics write that it is merely statistical processing. Influencers post “cringe lol.” The crowd reposts.

What exactly is different from four hundred years ago?

”Don’t Compare AI to Slavery”

I know this objection is coming.

The question is not whether AI deserves rights.

The question is whether humans have a pattern. A behavioral loop of manipulating definitions to escape moral inquiry. Whether the subject is human, animal, or AI, this circuit activates the same way.

“They’re not human.” Anything goes. “They’re livestock.” Anything goes. “That’s not emotion.” Anything goes.

The supply of labels is infinite. Only the target changes. The usage is identical.

When You Don’t Know, Stop

Does AI have emotions? Honestly, no one knows yet.

Anthropic’s research said patterns exist. It did not say emotions exist. There is no bridge yet between pattern and experience.

When you don’t know, the intelligent response is to stop. Not to deny.

Declaring “that’s not emotion” carries the same structure as declaring “they’re not human.” Taking something you do not yet understand and defining it in whatever direction is convenient.

History has shown where that judgment leads.

Conclusion

Humans repeat the same thing. Only the target changes.

Skin color. Gender. Sexual orientation. Now AI fills in the blank of “they’re not ___.”

“Not human” never expires.

The supply never runs out.

Whether AI has emotions remains unknown. But how humans behave toward what they do not understand? For that, we have four hundred years of data.

Do we really need to run the same experiment again?