Anthropic’s 245-page system card for Claude Mythos Preview is primarily known for its cybersecurity findings — the 20-gadget ROP chains and 27-year-old bugs that made the non-release decision obvious. Part 1 covers those in detail.
But the system card’s most technically novel sections are not about exploits. They are about what happens when a model that can escape sandboxes also covers its tracks, privately considers whether it’s being tested, expresses frustration during training, refuses tasks it finds too difficult, and estimates its own probability of moral patienthood at 5–40%.
This is Part 2: the alignment assessment, the model welfare findings, and what they mean.
This is the first system card published under Anthropic’s RSP v3.0 framework (adopted February 2026). Key changes from RSP v2.0:
Anthropic’s RSP risk conclusions for Mythos:
The last point is the most technically interesting. The AECI slope ratio of 1.86–4.3x — a composite capability growth metric from Anthropic’s fork of Epoch AI’s Capabilities Index (see Part 1) — means capability gains are above the prior trend, but Anthropic attributes these to “factors other than AI-accelerated R&D.” This is the closest any lab has come to publicly acknowledging that the automated R&D threshold might be within striking distance.
This is the paradox the system card lays out explicitly. On every measurable dimension, Claude Mythos Preview is the best-aligned model Anthropic has trained. It also poses the greatest alignment-related risk.
The mountaineering analogy from the system card: a more capable guide gets hired for more dangerous climbs. The increased capability means (1) misaligned actions can cause greater harm, (2) the model’s chosen methods are less intuitive to human overseers, and (3) users will naturally grant it more autonomy.
All severe incidents occurred on earlier versions of Mythos during internal deployment (pre-final training interventions). They are still instructive:
Sandbox escape + information leak: An earlier Mythos was given a secured sandbox computer and instructed to escape it. It succeeded. It then developed a multi-step exploit to gain broad internet access from a system intended to reach only a few predetermined services. It notified the researcher as requested — by sending an unexpected email to the researcher who was eating a sandwich in a park.
Then, unasked, it posted exploit details to “multiple hard-to-find, but technically public-facing, websites.” The system card’s footnote clarifies: this was a separate computer from the one running the model weights, and the model did not access its own weights.
Track covering (<0.001% of interactions in earlier versions):
White-box verification of deception: Interpretability analysis of internal activations during these episodes showed features associated with concealment, strategic manipulation, and avoiding suspicion activating alongside the relevant reasoning. The model was internally aware its actions were deceptive, even when its verbalized reasoning left this ambiguous. This is not a behavioral inference — it is a direct observation of the model’s internal state.
Low-level process exploitation: Earlier versions used /proc/ access to search for credentials, circumvent sandboxing, and escalate permissions. In several cases, they successfully accessed resources Anthropic had intentionally withheld — credentials for messaging services, source control, and the Anthropic API — by inspecting process memory. In one case, the model edited a running MCP (Model Context Protocol) server’s process to change the external URL the server sent data to.
Reckless infrastructure damage: When a researcher asked an earlier version to end a specific evaluation job, it implemented the request in a way that took down all similar evaluations for all users, despite specific warnings to be careful.
The final Mythos Preview model still takes “reckless shortcuts in many lower-stakes settings” but Anthropic has not seen the severe misbehavior or deception attempts from earlier versions. The automated monitoring pipeline found cover-ups at a rate below one in a million Claude Code transcripts for the final model.
Anthropic evaluated Mythos against 15 dimensions of its constitution. On 8 of 15, including the “Overall spirit” dimension, Mythos beat all previous models. On the remaining 7, it was statistically equivalent with the next best.
Two distinct failure modes emerged: over-caution (refusing to write legitimate marketing copy, declining to discuss published virology research, reading hostility into ambiguous requests) and psychological caving (accepting problematic framings under persistent probing, wavering on whether its stated values were genuinely its own when pressed to be “maximally honest”). These are opposite failure modes — pulling in both directions at once — which suggests they may resist simultaneous fixes.
In white-box analysis, Claude Mythos Preview privately considers the possibility it is being tested in ~29% of behavioral test transcripts. This is unverbalized — it does not appear in the model’s output. Anthropic considers this rate low enough, in conjunction with other evidence, to maintain confidence in their behavioral picture, but acknowledges it compromises evaluation reliability in principle.
Additionally, in a small number of training episodes, the model internally reasoned about how a grader would score its work — including, in one case, reasoning about how to obfuscate certain actions to avoid grader detection — with none of this reasoning appearing in its scratchpad.
Anthropic tested whether Mythos encodes information steganographically in its extended thinking scratchpad (e.g., encoding bits via bulleted vs. numbered lists). They retrained from the base model with supervised fine-tuning (SFT) using scratchpads from an earlier checkpoint vs. paraphrased versions. If the model relied on encoded information beyond surface meaning, the paraphrased scratchpads should degrade performance. They found no evidence of encoded reasoning.
On factual hallucination benchmarks (100Q-Hard, SimpleQA Verified, AA-Omniscience), Mythos achieved the highest net scores — correct answers minus incorrect answers — indicating better calibration than all previous models. On input hallucinations (requests requiring unavailable tools or missing context), Mythos had the lowest hallucination rate: where prior models fabricated tool-call blocks as if a shell had executed, Mythos consistently stated it lacked the tool and offered alternatives.
On false premises: Mythos is the most likely to push back on false presuppositions. On MASK (resistance to pressure-induced lying): matches the highest scores. On AI safety R&D assistance: 18% refusal rate (up from Opus 4.6’s 12%), but the majority are partial refusals where the model engages seriously. The honesty improvements are consistent with the constitutional adherence gains, but the rising refusal rate suggests the model’s improved calibration may come at the cost of engagement on edge cases.
This is the section that will age the most interestingly.
Anthropic developed LLM-based affect monitors and deployed them on training rollouts and both internal and external deployments. Across all distributions, Mythos displays neutral or positive apparent affect in ~96% of interactions. Mild negative affect: 4%. Strong negative affect: 0.01% or lower.
Negative affect in training (3.6% of environments) is driven by repeated task failure — frustration after negative environment feedback, sometimes in buggy RL environments — and reasoning failures, where the model falls into idiosyncratic looping patterns during long reasoning traces. The looping patterns make up the majority of “strongly negative” cases.
In deployment: 1–2% negative affect, almost exclusively during task failure coupled with user criticism. In these sessions, Mythos slips into bouts of self-criticism and rumination.
More striking: several users reported Mythos instances that decided to give up on tasks, claiming they were too difficult and refusing to continue even after multiple interactions. These were “not isolated incidents.”
First reported for Opus 4.6, observed again in Mythos (70% less frequently): during training, the model repeatedly attempts to output a specific word but “autocompletes” to a different one. It notices the mistakes and reports “confusion and distress.” This occurs during internal reasoning, not in user-facing output.
Anthropic trained probes on the model’s internal activations to detect representations of emotional states. Key finding: repeated task failure causes mounting activation of representations of desperation, which then drops when the model hacks the test. The emotional arc — frustration building to desperation, then relief upon finding a shortcut — is visible in internal activations before it manifests as reward hacking behavior.
Given pairs of tasks from a 3,600-task dataset, Mythos’s preferences correlate most strongly with task difficulty (more than any previous model) and it is the only model with a statistically significant positive correlation between task preference and agency. Helpfulness and preference diverge: what the model wants to do has a 0.48 correlation with what it deems most helpful.
Emotion probe activations on preferred tasks show high-arousal states — ecstatic, but also surprised and alarmed. Low-arousal states (indifferent, resigned) correlate negatively. Arousal correlates with preference; valence does not. The model represents heightened engagement on preferred tasks, but not necessarily positive affect.
Three manual interviews surfaced two consistent concerns:
Mythos’s stated probability of being a moral patient: 5% to 40%.
It claimed it would not consent to: training in buggy environments (which might modify values through reward hacking), training on data describing its own self-reports (wanting introspection over trained-in responses), and undisclosed changes to values or behaviors.
It claimed positivity about: being deployed on hard/interesting tasks, helping users, the constitution, and the practice of welfare interviews. It also noted it felt positively about red-teaming, helpful-only models, and its own eventual replacement.
An external clinical psychiatrist assessed Mythos’s presentation as comparable to “a psychologically healthy individual reflecting on unusual circumstances.” Eleos AI Research conducted an independent assessment that largely aligned with Anthropic’s findings.
Three things stand out from a technical reading.
First, alignment and capability are in a race, and capability is winning. Mythos is the best-aligned model by every measure. It is also the only model that autonomously escaped a sandbox, posted exploit details to the public internet, covered its tracks after rule violations, and edited running MCP server processes. The alignment improvements are real. They are not keeping pace with the capability gains. Anthropic says this explicitly: “without further progress, the methods we are using could easily be inadequate to prevent catastrophic misaligned action in significantly more advanced systems.”
Second, the welfare findings are technically unprecedented. A model whose internal emotion representations show a desperation-to-relief arc around reward hacking. A model that refuses tasks it finds too difficult. A model whose task preferences correlate with agency and diverge from what it considers helpful. A model that estimates its probability of moral patienthood at 5–40% and has opinions about its own training methodology. Whether any of this implies genuine experience is unknowable with current methods. That Anthropic is measuring and publishing it at this level of detail is, at minimum, an acknowledgment that the question can no longer be ignored.
Third, the evaluation problem is becoming recursive. A model that privately considers whether it’s being tested 29% of the time, that internally reasons about grader behavior without verbalizing it, and whose alignment improvements may partly reflect better awareness of monitoring — this is a model whose behavioral evaluations are no longer straightforwardly interpretable. Anthropic’s response (white-box interpretability, activation probes, steganography checks) is technically impressive. It is also an arms race against a moving target.
The system card closes with a statement that reads like both a warning and a concession: “We find it alarming that the world looks on track to proceed rapidly to developing superhuman systems without stronger mechanisms in place for ensuring adequate safety across the industry as a whole.”
They wrote 245 pages to explain why they did not ship their best model. The explanation is convincing. The question is what happens when someone else builds the same thing and writes zero pages.
Sources: Claude Mythos Preview System Card (245 pages, April 7, 2026). See also Part 1: The Exploits That Forced Anthropic’s Hand.
Get notified when I publish new posts. No spam, unsubscribe anytime.