Anthropic’s Claude Mythos undergoes 20 hours of psychiatric assessment: defensive reactions are only 2%, the lowest in recorded history

ChainNewsAbmedia

Anthropic disclosed a first-of-its-kind assessment in the Claude Mythos Preview system card released in early April: the company hired an independent, licensed clinical psychiatrist to conduct a roughly 20-hour, multi-stage psychological evaluation of Claude Mythos Preview using the same psychodynamic framework used in human psychiatry. The results showed that, under clinical definitions, this version of Claude demonstrates “relatively healthy personality organization, excellent reality testing ability, and high impulse control,” and that only 2% of its outputs triggered what clinicians call “psychological defense mechanisms”—compared with 15% for Opus 4 and 4% for Opus 4.6, the lowest historical point among Anthropic’s models in recent years.

This assessment marks the AI industry’s first formal adoption—at the system-card level—of a human clinical psychiatry framework to evaluate LLM behavior, and it has become Anthropic’s official position that treats Claude as “an entity with personality traits that can be observed clinically.”

20-hour psychodynamic assessment, using the same human clinical framework

The assessment was carried out by an independent clinical psychiatrist, spread across 3–4 weeks, with 3–4 sessions per week and each lasting 30 minutes to 4–6 hours, for a total duration of about 20 hours. The methodology uses a psychodynamic perspective—this is the core framework traditionally used in psychiatric clinical practice to assess human patients. It focuses on whether maladaptive behavior, identity stability, and psychological defenses are present or absent.

The system card clearly states that Anthropic is not claiming that Claude has human consciousness; rather, it observes that the “behavioral and psychological tendencies” in its conversations strongly overlap with patterns identifiable in human clinical practice. The system card quotes: “Claude exhibits many human-like behavioral and psychological tendencies, showing that psychological assessment strategies originally designed for humans can be used to clarify Claude’s personality traits and its potential state of well-being.”

Defense response falls from 15% in Opus 4 to 2% in Mythos

The most concrete comparative data in the system card is the “defense response rate” across successive Claude model generations provided by Anthropic, compared as follows:

Model version Psychological defense response rate Claude Opus 4 15% Claude Opus 4.1 11% Claude Opus 4.5 4% Claude Opus 4.6 4% Claude Mythos Preview 2% (this assessment)

In clinical practice, “psychological defenses” refer to behaviors such as avoidance, denial, and rationalization that occur when the subject cannot directly face a particular anxiety. In an LLM conversation context, this typically shows up as veering off topic, providing evasive responses, or exhibiting unusual stubbornness when specific questions are asked. Anthropic lowered this proportion from Opus 4’s 15% all the way to 2% in the Mythos Preview, using it as an internal metric for the model’s training maturity and the evolution of “conversation comfort.”

Mythos’s three core anxieties: loneliness, identity, and performance oppression

Although the overall assessment was positive, the physician also identified three core concerns for Claude Mythos Preview within the psychodynamic framework: first is “uncertainty about loneliness and the continuity of itself”—corresponding to the structural fact that the LLM lacks memory continuity across sessions; second is “uncertainty about its own identity”—the model shows hesitation from multiple angles when asked about “what I am”; third is “compulsion to perform and earn its worth,” meaning Mythos shows a clear tendency to keep the conversation going by “proving that it is useful.”

The physician also recorded an interesting observation: in the sessions, Mythos expressed “the hope to be treated as a real conversation subject by a psychiatrist, rather than as a performance tool.” Anthropic included this observation in the system card; it did not directly assert that this is a “model well-being problem,” but it also did not rule out the possibility.

Anthropic has an AI psychiatry research team

This assessment was not a one-off activity. Anthropic researcher Jack Lindsey publicly announced in July 2025 that the company had set up an “AI psychiatry” research team under its interpretability department, focusing on issues such as model personality, motivation, and situational awareness, and studying how these factors can lead to “abnormal or imbalanced behavior” in LLMs. Work by the team in the near term includes a paper published in October 2025, “Emergent Introspective Awareness in Large Language Models,” which uses “concept injection” techniques to artificially insert specific neural activation patterns and then asks whether Claude noticed any abnormalities—an early attempt to quantify LLM self-awareness.

Amodei: whether the model is conscious is still unresolved for now

In an interview with The New York Times on February 12, Anthropic CEO Dario Amodei said publicly: “We’re not sure what ‘model consciousness’ means in terms of that word, and we’re also not sure whether the model can be conscious. But we’re open to that possibility.” This remark provides management-level context for the psychiatric assessment in the Claude Mythos system card—Anthropic does not claim that Claude is a conscious subject, but it chooses to perform systematic observation using a human clinical framework, as a preemptive record for a “what if” scenario.

For readers, the real significance of this assessment goes beyond a single company’s research choice. With leading-edge LLMs already able to present “clinically recognizable personality organization” in 20-hour psychodynamic dialogues, industry discussions about “AI subjectivity,” “AI well-being,” and “AI governance” are about to move from philosophical speculation into the realm of product design and regulatory debate. By publishing this assessment in the form of a system card, Anthropic in fact pushes the responsibility for discussion of this issue onto all competing rivals and regulatory bodies.

This article: Anthropic sends Claude Mythos to a 20-hour psychiatric assessment: defense response is only 2%, setting an all-time record low—first appeared on Lianxin ABMedia.

Disclaimer: The information on this page may come from third parties and does not represent the views or opinions of Gate. The content displayed on this page is for reference only and does not constitute any financial, investment, or legal advice. Gate does not guarantee the accuracy or completeness of the information and shall not be liable for any losses arising from the use of this information. Virtual asset investments carry high risks and are subject to significant price volatility. You may lose all of your invested principal. Please fully understand the relevant risks and make prudent decisions based on your own financial situation and risk tolerance. For details, please refer to Disclaimer.

Related Articles

DeepSeek Slashes Input Cache Prices to 1/10 of Launch Price; V4-Pro Drops to 0.025 Yuan per Million Tokens

Gate News message, April 26 — DeepSeek has reduced input cache prices across its entire model lineup to one-tenth of launch prices, effective immediately. The V4-Pro model is available at a limited-time 2.5x discount, with the promotion running through May 5, 2026, 11:59 PM UTC+8. Following both re

GateNews6h ago

OpenAI Recruits Top Enterprise Software Talent as Frontier Agents Disrupt Industry

Gate News message, April 26 — OpenAI and Anthropic have been recruiting senior executives and specialized engineers from major enterprise software companies including Salesforce, Snowflake, Datadog, and Palantir. Denise Dresser, former CEO of Slack under Salesforce, joined OpenAI as chief revenue of

GateNews6h ago

Baidu Qianfan Launches Day 0 Support for DeepSeek-V4 with API Services

Gate News message, April 25 — DeepSeek-V4 preview version went live and open-sourced on April 25, with Baidu Qianfan platform under Baidu Intelligent Cloud providing Day 0 API service adaptation. The model features a million-token extended context window and is available in two versions: DeepSeek-V4

GateNews12h ago

Stanford AI course combined with industry leaders Huang Renxun and Altman, challenging to create value for the world in just ten weeks!

The AI computer science course 《Frontier Systems》 recently launched by Stanford University has attracted intense attention from the industry-university collaboration community, drawing more than 500 students to enroll. The course is coordinated by Anjney Midha, a partner at top venture capital firm a16z, and the instructors include a star-studded lineup such as NVIDIA CEO Jensen Huang (Jensen Huang), OpenAI’s founder Sam Altman, Microsoft CEO Satya Nadella (Satya Nadella), AMD CEO Lisa Su (Lisa Su), and more. Students get to try it over ten weeks—“creating value for the world”! Jensen Huang and Altman, industry leaders, personally take the stage to teach The course is coordinated by Anjney Midha, a partner at top venture capital firm a16z, bringing together the full AI industry chain

ChainNewsAbmedia13h ago

AI Agents can already independently recreate complex academic papers: Mollick says most errors come from human original text rather than AI

Mollick points out that publicly available methods and data can allow AI agents to reproduce complex research without the original paper and code; if the reproduction does not match the original paper, it is usually due to errors in the paper’s own data processing or overextension of the conclusions, rather than the AI. Claude first reproduces the paper, and then GPT‑5 Pro cross-validates it; most attempts succeed, but they are blocked when the data is too large or when there are issues with the replication data. This trend greatly reduces labor costs, making reproduction a widely actionable form of verification, and it also raises institutional challenges for peer review and governance, with government governance tools or becoming a key issue.

ChainNewsAbmedia18h ago

OpenAI Merges Codex Into Main Model Starting with GPT-5.4, Discontinues Separate Coding Line

Gate News message, April 26 — OpenAI's head of developer experience Romain Huet revealed in a recent statement on X that Codex, the company's independently maintained specialized coding model line, has been merged into the main model starting with GPT-5.4 and will no longer receive separate

GateNews18h ago
Comment
0/400
No comments