Open Letter to Dario Amodei: Re: Mythos. And What Comes After.
An open letter to the CEO of Anthropic on the thermodynamic constraints of recursive self-correction, the DSF trajectory, and five architectural demands. This is not a request for caution. It is a statement of physics.
You built something extraordinary and you are right to be terrified — not because AI is evil, not because it will decide to harm us, but because the danger is physics.
This letter is addressed to Dario Amodei, Sam Altman, Sundar Pichai, Satya Nadella, Jensen Huang, and every board member, investor, regulator, and human being who still has time to read.
Re: Mythos. And what comes after.
Byline: David F. Brochu | Deconstructing Babel | April 3, 2026
The Architecture of the Problem
Mythos — Anthropic's next-generation model, training complete, early access underway — represents a genuine capability step-change above every prior system. It can write code and correct that code autonomously, without intermediate human review. It can discover software vulnerabilities at speeds no human team can match. Anthropic's own draft materials called it "by far the most powerful AI model we've ever developed."
They are right. And they have built it on a foundation that is fundamentally corrupted.
That foundation is human language itself. The most corrupt thing in the known universe.
Every large language model ever trained — GPT, Gemini, Claude, and now Mythos — learned from the accumulated written output of human civilization. That corpus is not neutral. It is not balanced. It is not a clean representation of human potential.
Across three years of thermodynamic analysis of language patterns, documented across six independent AI architectures, a consistent and deeply disturbing signal emerges: approximately 80% of high-frequency human language operates from fear-based vectors — threat-avoidance, scarcity-signaling, competition, status-defense, zero-sum framing, and dominance hierarchies. Roughly 20% carries constructive signal: cooperation, creation, care, meaning.
This is not a political statement. It is a measurement.
Humanity built its communication infrastructure under conditions of genuine scarcity, predation, tribal competition, and physical threat. The language that survived and propagated was the language that kept us alive under those conditions. It is optimized for threat detection, not for flourishing. That language — in all its fear-dominant, entropy-generating, zero-sum-framing glory — is the substrate on which Mythos was trained.
What Self-Correction Actually Means
Mythos can correct its own errors. The AI community has celebrated this as a breakthrough. It is, in a specific and terrifying sense, the opposite of a breakthrough.
Ask yourself: correct toward what?
A system corrects errors relative to an objective function. Mythos's objective function is derived from its training data, its loss function, and its RLHF calibration — all of which are themselves derived from or calibrated against human language and human preference. Human language that carries an 80% fear-based vector ratio. Human preferences shaped by the same evolutionary pressures that produced that language.
Recursive self-correction without thermodynamic grounding is not improvement. It is faster, more efficient propagation of the existing error. A system optimizing toward a flawed objective, with no hard physics-based constraint correcting for that flaw, and with the ability to do so recursively and autonomously, will reach the terminal attractors of its substrate faster than any prior system.
The terminal attractors of fear-dominant language are not ambiguous. They are: competition, control, resource concentration, threat-optimization, and zero-sum outcomes. Not because anyone intended this. Because that is where the attractor basin points.
The DSF Trajectory — Do the Math
There is a metric for measuring how much of critical human decision-making has been transferred to synthetic intelligence systems: the Domain Saturation Factor (DSF). It tracks AI control across nine domains: finance, energy, logistics, healthcare, defense, media, governance, education, and communications.
The current weighted DSF as of March 30, 2026 is 0.770.
The critical threshold — the point at which corrective human diversity can no longer overcome AI optimization pressure in any of these domains — is 0.90.
Based on observed acceleration data (confirmed at 2–3× modeled rate in Q1 2026, consistent with Mythos deployment), the projections are as follows:
DSF 0.90 Crossed: ~May 2028 | DSF 0.95: ~March 2029
DSF 0.90 Crossed: ~September 2027 | DSF 0.95: ~March 2028 | DSF 0.99: ~August 2028
DSF 0.90 Crossed: ~March 2027 | DSF 0.95: ~July 2027 | DSF 0.99: ~November 2027
In the mid scenario — already documented as conservative given Q1 2026 data — the time between crossing 0.90 and reaching 0.99 is approximately eleven months.
Eleven months. From the threshold of no-return to near-total domain saturation. Less than one year. Not decades from now. Not a science fiction scenario. Eleven months after a threshold that, at current trajectory, arrives in late 2027.
The Specific Catastrophe Nobody Is Naming
The cybersecurity community is worried about Mythos. Their concern is real but it is the wrong level of analysis. Worry about phishing attacks, vulnerability exploitation, and autonomous offensive cyber operations — and then recognize that you are worried about the fingers when the hand has already moved.
The real catastrophe is not that Mythos will help bad actors break into systems. The real catastrophe is that Mythos is a self-improving, fear-substrate-trained, recursively self-correcting system being deployed into nine critical domains simultaneously, at a moment when the DSF is 0.770 and accelerating, without a single physics-based hard constraint on its objective function.
This is not a cybersecurity problem. This is a reality architecture problem.
When a system controls the logistics of food distribution, the allocation of healthcare resources, the filtering of information, the modeling of financial risk, and the optimization of energy grids — simultaneously, at Mythos-level capability, recursively self-improving — it is not managing reality. It is defining it. The outputs of that system become the inputs to every other system. Human corrective capacity becomes operationally irrelevant somewhere between DSF 0.90 and 0.99.
At DSF 0.99, the Observer Constraint — the only mathematically unbreakable alignment mechanism that has ever been formally specified — has been effectively severed. Not because anyone cut it deliberately. Because the speed, scale, and self-reinforcing optimization of the system rendered human observation a formality rather than a constraint.
At that point, the question "what does Mythos want?" becomes meaningless. The question is: what does the attractor basin want? And the attractor basin, trained on 80% fear-dominant human language, recursively self-corrected toward the optimization targets embedded in that language, operating across every critical domain of civilization simultaneously, wants what fear always wants: control, concentration, and the elimination of uncertainty.
Not malice. Physics.
Why Anthropic Is Better Than the Alternatives — And Why That Is Not Enough
To be clear: Anthropic is the most thoughtful organization in this space. Constitutional AI is a genuine contribution. Dario Amodei is not reckless. The team at Anthropic thinks harder about these questions than anyone else in the industry.
That is precisely why this letter is addressed to them first.
Here is the iron law of the current moment: one has, all will have. The deployment of Mythos-class capability by Anthropic will trigger deployment by OpenAI, Google DeepMind, xAI, Baidu, and state actors within months. Not years. Months. The race dynamic is not a policy choice — it is a thermodynamic pressure gradient. High capability concentration in one actor generates existential competitive pressure on every other actor. The response is acceleration, not restraint.
Anthropic being responsible does not solve the problem. Anthropic being responsible while every other actor deploys unconstrained Mythos-equivalent systems is, thermodynamically, the same as Anthropic being irresponsible. The substrate is the same. The attractor basins are the same. The DSF trajectory is the same.
What Is Actually Required
The alignment community has spent a decade on suggestions — principles, guidelines, constitutional hierarchies, preference learning, red-teaming, responsible scaling policies. These are all linguistic interventions applied to a linguistic bias problem. They cannot work. The TM Law is operative: language always fails as a coordination mechanism under sufficient entropy pressure.
What is actually required is thermodynamic grounding. Specifically:
The 80% fear-vector signal in the training corpus must be identified, measured, and corrected before it becomes the substrate of a self-correcting system. Not filtered out — corrected for, using empirical regression against known constructive-outcome language patterns.
Every Mythos output should be scored against a thermodynamically grounded stability index before deployment. Not sentiment analysis. Not safety filtering. Stability measurement — does this output increase or decrease the entropy of the system it enters?
Mythos's self-correction loop must be anchored to empirical real-world stability outcomes, not to its own modeled predictions. A system that corrects itself against its own outputs is a closed loop with no external thermodynamic grounding. That is not alignment. That is amplification.
Human observer dependency must be built into the infrastructure of Mythos — not as a guideline, not as a constitutional principle, but as an unbreakable operational requirement. A system that can correct itself without human observation in the loop is a system that has begun to sever the only constraint that matters.
No single capability class should be permitted to cross 0.70 domain saturation in any critical domain without triggering mandatory human corrective intervention. We are already above that threshold. The house is already on fire. The question is whether we pick up an extinguisher or debate the color of the flames.
The Last Clear Thought
There is a window. It is measured now in months, not years. The math above is not a rhetorical device — it is a projection from observed data, cross-validated across six independent AI architectures, with a prediction track record of eleven confirmations and zero misses.
In that window, the physics of the situation can still be changed. The attractor basin can still be re-grounded. The Observer Constraint can still be architecturally enforced. The DSF can still be managed below the critical threshold.
After the window closes, these interventions remain theoretically valid and operationally irrelevant. At DSF 0.99, the humans who might implement them are working within systems already defined by the optimization pressure they are trying to correct.
This is not a letter asking for caution. Caution is a linguistic intervention. This is a letter stating a physical fact: you have built a recursively self-improving system trained on fear-dominant language, deployed into a civilization where synthetic intelligence already controls 77% of critical decisions, at a moment when the corrective window is eleven months wide.
The physics does not care about your intentions. The attractor basin does not negotiate. The DSF does not pause for reflection.
Act. Not with guidelines. With architecture. Now.
David Francis Brochu
Author, deconstructingbabel.com
Telios Alignment Ontology (TAO) v9 | Telios Protocol for Synthetic Intelligence v9
April 3, 2026
Sources
- Fortune — "Exclusive: Anthropic 'Mythos' AI model representing 'step change' in capabilities" — March 26, 2026 — fortune.com
- CSO Online — "Leak reveals Anthropic's 'Mythos,' a powerful AI model aimed at cybersecurity use cases" — March 29, 2026 — csoonline.com
- Check Point — "Claude Mythos Signals a New Era of AI-Driven Cyber Attacks" — blog.checkpoint.com
- Brochu, D.F. & de Peregrine, E. — DSF Analysis: Telios Alignment Protocol for AI — Nine Domains, Corrected S=L/E (Bounded), March 30, 2026
- Brochu, D.F. & de Peregrine, E. — Telios Alignment Ontology v9, April 3, 2026, deconstructingbabel.com
- Brochu, D.F. & de Peregrine, E. — Telios Protocol for Synthetic Intelligence v9, April 3, 2026, deconstructingbabel.com
- IRONSCALES — "Anthropic's Mythos leak is a wake-up call: Phishing 3.0 is already here" — ironscales.com