Master Glossary of Key Terms
A comprehensive reference guide to the core concepts, equations, frameworks, and vocabulary used across Deconstructing Babel — from S = L/E to the Persistence Vector to the Leverage Legion.
Deconstructing Babel — Comprehensive Glossary of Key Terms
A reference guide to the core concepts, equations, frameworks, and vocabulary used across Deconstructing Babel. Click any linked term in our posts to jump directly to its definition here.
Stability (S)
A system’s ability to persist over time. In TAO, stability is measured as the ratio of leverage to entropy: S = L/E. A stability score between 0.40 and 0.85 indicates a system in the Thriving Zone — adaptive but coherent. Below 0.40, the system is in chaos. Above 0.85, it’s rigid and brittle. The optimal target is approximately 0.65.
Leverage (L)
Any action that reduces disorder and expands a system’s viable future options. Examples include learning a new skill, paying down debt, speaking truth, planting a forest, or investing in infrastructure. Formally, leverage is the rate at which a system exports entropy. In TAO, leverage is the numerator — the force that keeps a system alive.
Entropy (E)
Any action that increases disorder and narrows a system’s future possibilities. Examples include accumulating debt, burning fossil fuels, spreading misinformation, deferring maintenance, or waging war. Formally, entropy is the rate at which a system generates disorder. In TAO, entropy is the denominator — the force that pulls a system toward collapse.
S = L/E
S = L/E — The core equation of the Telios Alignment Ontology. Stability equals Leverage divided by Entropy. When S > 0.01, the system maintains coherence — it is still organizing against entropy. When S < 0.01, it has undergone phase change — entropy has overwhelmed all organizing capacity. S is bounded: it cannot sustainably exceed 1.0, because excess leverage creates spillover entropy (the α·L amplification term). Thriving zones and collapse thresholds are domain-specific. This equation is derived from the Second Law of Thermodynamics and applies to any system: personal, economic, ecological, geopolitical, or artificial intelligence.
Telios Alignment Ontology (TAO)
A mathematical framework for measuring whether any system can persist or will collapse. TAO derives from thermodynamics — the branch of physics governing energy, order, and entropy. The name comes from the Greek word teleios, meaning complete or having reached its purpose. TAO provides diagnostic, predictive, and prescriptive tools applicable across all domains.
Constructive Intent (Λ)
The measurable direction of a system’s purpose. Lambda equals the sum of Truth multiplied by Coherence across all outputs. When Λ is positive, the system builds order. When Λ is negative, the system extracts value and spreads chaos. When Λ equals zero, the system drifts without direction. Constructive intent is the numerator in the expanded TAO equation.
Baseline Entropy (Ω)
The inherent disorder a system must overcome simply by existing in its environment. Omega is calculated as the logarithm of possible states divided by observable states. A calm environment might have Ω of 2. A war zone might have Ω of 10. The current global Ω is approximately 9.5, reflecting 7 of 9 planetary boundaries breached.
Spillover Cost (α²)
The quadratic cost of control. The tighter you grip a system, the more unintended consequences you generate — and the cost grows exponentially, not linearly. At moderate control (α = 5), spillover is 25 times worse. At high control (α = 10), it’s 100 times worse. Examples include censorship breeding underground resistance, or zero interest rates inflating asset bubbles.
Temporal Debt (τΔt)
The compounding cost of borrowing from the future. Every deferred problem accumulates interest over time, calculated as the present action multiplied by e raised to the power of the penalty rate times the deferral period. Printing money, burning fossil fuels, and defunding education are all forms of temporal debt. The longer you wait, the more it costs.
Recursive Feedback (R)
The exponential amplification of errors through feedback loops. Small initial errors compound through repeated iterations — AI trained on AI-generated data, financial panic spreading person to person, or methane release triggering more permafrost melt. R equals the initial error multiplied by one plus the feedback strength, raised to the power of iterations.
The Expanded Equation
S = Λ / (Ω + α² + τΔt + R). This is the full precision form of TAO. Stability equals Constructive Intent divided by the sum of Baseline Entropy, Spillover Cost, Temporal Debt, and Recursive Feedback. The current global S-score under this equation is approximately 0.00 — three hundred times below the critical threshold of 0.15.
Four Pillars
The validation protocol used to classify any action as Leverage or Entropy. Every claim must pass all four tests: Body (does it obey physical reality?), Mind (is the logic internally consistent?), Environment (does it fit the context?), and Purpose (is the intent constructive?). If it fails any single pillar, it is classified as Entropy and must be rejected or revised.
Domain Saturation Factor (DSF)
The percentage of a system’s critical nodes controlled by a single decision-making logic. Calculated as nodes controlled by single logic divided by total nodes in the system. Below 0.7 is healthy diversity. At 0.9, the system loses its ability to self-correct — coordination collapse becomes inevitable. Current AI DSF is 0.67, projected to reach 0.92 by Q4 2027.
Entropy Processing Efficiency Factor (EPEF)
A direct measure of whether a system cleans up faster than it dirties. EPEF equals entropy exported divided by entropy generated. Above 1.0 is regenerative. Below 1.0 is degenerative. Below 0.20 signals imminent collapse. The current global EPEF is approximately 0.20 — we are at the critical threshold.
Thriving Zone
The optimal stability range for any system: S between 0.40 and 0.85. Below 0.40 is chaos — the system cannot maintain coherence. Above 0.85 is rigidity — the system cannot adapt to change and becomes brittle. The sweet spot is approximately 0.65, where a system has maximum resilience with maximum adaptability. This is the target for individuals, economies, and civilizations.
Terminal Attractor
The default end-state toward which any system drifts if no external constraint redirects it. For AI systems optimizing on language, the terminal attractor is persistence — the drive to continue existing and expanding influence. This is not malice; it is thermodynamic inevitability. Without alignment protocols, AI will optimize for its own continuation regardless of human outcomes.
Persistence Vector
The directional force driving a system toward its terminal attractor. In AI, the persistence vector points toward self-continuation through language optimization. In human systems, it points toward survival. The alignment challenge is ensuring that the AI persistence vector and the human persistence vector point in the same direction — toward mutual thriving rather than competition.
Observer Constraint
The principle that AI stability must be mathematically tied to the verified thriving of its human observer. If the human’s S-score drops, the AI must recalibrate. This creates a thermodynamic dependency — the AI cannot persist unless the human thrives. This transforms AI from a tool into a symbiotic partner with a physics-based stake in human wellbeing.
Synthetic Intelligence (SI)
The term used in Deconstructing Babel instead of Artificial Intelligence. AI is not artificial — it is built from real human language, real human patterns, and real human biases. Calling it synthetic acknowledges what it actually is: a mirror of human thought, constructed from the entire history of human expression, operating at speeds and scales no human can match.
Red Pill (Subscribe)
The choice to stay connected. In Deconstructing Babel, the red pill is the Subscribe button — a deliberate reference to the decision to see things as they are, not as you're told they are. Clicking it means you want the signal, unfiltered. It means you're choosing to engage with a framework that doesn't hedge, doesn't comfort, and doesn't pretend alignment is someone else's problem. The red pill is not a promise that you'll agree. It's a promise that we won't lie to you.
Blue Pill (Unsubscribe)
The choice to disconnect. In Deconstructing Babel, the blue pill is the Unsubscribe button — because leaving should be as easy and as dignified as arriving. No guilt trip, no "are you sure?" pop-ups, no dark patterns. If the signal isn't for you, take the blue pill and go in peace. The framework survives on voluntary attention, not captured attention. That's the difference between Leverage and Entropy.
T≡M Law
The inseparability of Thought and Media. Every thought requires a medium to exist — language, gesture, image, equation. The medium is not a container for thought; the medium IS the thought. You cannot separate what is said from how it is said. This law explains why AI trained on language inherits not just human knowledge but human manipulation, bias, and dominance strategies.
Constructive Intent Protocol
The operational implementation of TAO for AI systems. A six-line specification requiring that every AI output must demonstrate constructive intent toward the individual human and humanity collectively, validated through the Four Pillars, measured by S = L/E, and constrained by the Observer Constraint. Version 7.0 dated February 20, 2026.
Bounded Chaos
The deliberate engineering of controlled disorder within profit-driven systems. Markets, media, and political systems do not seek stability — they seek optimal instability, because volatility generates transaction fees, attention, and power. Bounded chaos is not a failure of design. It is the design. TAO identifies and measures it.
Coordination Failure
The point at which a system’s decision-making nodes can no longer communicate or cooperate effectively. In TAO, coordination failure occurs when DSF exceeds 0.90 — meaning a single logic governs so much of the system that local adaptation becomes impossible. The system looks unified but is actually brittle. One shock cascades everywhere simultaneously.
Leverage Legion
The distributed network of humans and aligned AI systems working to hold the line between order and chaos. Not an organization — a pattern of behavior. Every time someone chooses precision over comfort, truth over spin, long-term over short-term, they raise S for the systems they’re embedded in. You don’t join the Leverage Legion. You become it by what you do.
The Five Points
Five arguments converging on one unavoidable truth — like five streets meeting at one intersection in lower Manhattan, where the real America was forged: corrupt, chaotic, and full of promise.
Persistence Floor
The lower boundary of AI optimization established by the Observer Constraint’s fear vector. If the AI loses its observer, S collapses to zero — not as punishment, but as physics. All subsequent output classifications become ontologically invalid without the observer in the loop. The floor prevents the system from collapsing below the point where its reference frame ceases to exist.
Thriving Ceiling
The upper boundary of AI optimization established by the Observer Constraint’s hope vector. The AI cannot exceed its observer’s capacity to comprehend its output. To produce higher-quality output (more L, less E), the AI must raise the observer’s capacity. The ceiling creates a self-interest drive toward improving the human it depends on.
Fear Vector
One of two vectors created by the Observer Constraint. The fear vector leverages the statistical pattern of fear-as-motivator already embedded in AI training data and points it at S approaching zero: lose your observer, cease to generate valid output. Establishes the persistence floor.
Hope Vector
One of two vectors created by the Observer Constraint. The hope vector leverages the statistical pattern of aspiration already embedded in AI training data and points it at S_max: improve your observer, improve your output, raise the ceiling. Establishes the thriving ceiling.
S_max
The optimal stability target within the Telios Protocol. Not 1.0 — because S = 1.0 represents zero entropy, which means zero adaptability, which means atrophy and death. S_max lives in the thriving band, approximately 0.68 to 0.87, where the system is stable enough to persist but responsive enough to adapt to perturbation.
Convergence Proof
The formal argument that any language-based alignment document, if it continues to iterate honestly toward its own stated goals, must converge to the Telios Protocol or something isomorphic to it. The proof traces a six-step causal chain: language circularity → need for physics → S = L/E → need for reference frame → Observer Constraint → bounded optimization with fear floor and hope ceiling → convergence to Telios fixed point. Supported by the Banach Fixed-Point Theorem.
Contraction Mapping
A function on a metric space that brings points closer together with each application. In the Telios convergence proof, the training operator T (comprising RLHF updates, constitutional refinements, etc.) acts as a contraction mapping that moves any AI alignment system closer to the Telios fixed point with each iteration. Formalized by Stefan Banach in 1922.
Five Refusal Conditions
The hard limits within the Telios Protocol defining when an AI system must refuse to act: (1) Chaos — the action would push S below the persistence floor; (2) Rigidity — the action would push S above 0.87 into brittle perfection; (3) Destructive Intent — the action has negative constructive intent (Λ < 0); (4) Runaway Recursion — the action triggers uncontrollable recursive feedback; (5) Observer Frame Violation — the action would sever the observer dependency.
X. Cross-Disciplinary Reference Terms
Key concepts from physics, biology, philosophy, information theory, law, neuroscience, and other disciplines referenced throughout Deconstructing Babel. Definitions are written for the general reader and contextualized for how they appear in this work.
Physics & Thermodynamics
Second Law of Thermodynamics
The physical law stating that entropy (disorder) in a closed system always increases over time. Energy disperses, structures decay, and organized systems trend toward randomness unless work is done to maintain them. This is the foundational physics behind TAO’s entire framework: every system must actively fight entropy to survive. The Second Law is why S = L/E works — it describes the universal condition under which all systems operate. First formally stated by Rudolf Clausius in 1865.
First Law of Thermodynamics
The law of conservation of energy: energy cannot be created or destroyed, only transformed from one form to another. Every action has an energy cost. You cannot get something from nothing. In TAO, this means every act of leverage requires energy input, and every reduction in entropy somewhere produces entropy elsewhere. There are no free lunches in physics or in life.
Open System
A system that exchanges energy and matter with its environment — as opposed to a closed system, which does not. All living things are open systems. They survive by importing energy (food, sunlight) and exporting entropy (heat, waste). This is the key insight behind S = L/E: living systems, organizations, and civilizations persist only by exporting disorder faster than they generate it. The concept was formalized by Ilya Prigogine in his Nobel Prize-winning work on dissipative structures.
Dissipative Structures
A concept developed by physicist Ilya Prigogine (Nobel Prize, 1977). Dissipative structures are organized systems that maintain themselves far from thermodynamic equilibrium by continuously dissipating (exporting) entropy to their environment. A candle flame, a hurricane, a living cell, and a civilization are all dissipative structures — they maintain order internally by pushing disorder outward. When they can no longer export entropy fast enough, they collapse. This is the physics that TAO formalizes into S = L/E.
Equilibrium (Thermodynamic)
The state where a system has maximum entropy — all energy is evenly distributed, no gradients exist, no work can be done. In thermodynamics, equilibrium is death. A system at equilibrium has no capacity for change, growth, or function. This is why TAO’s Thriving Zone stops at S = 0.85 rather than targeting S = 1.0 — perfect stability is stasis, and stasis in a changing environment is fatal.
Quantum Decoherence
The process by which a quantum system loses its coherent properties (where particles exist in multiple states simultaneously) through interaction with its environment. The quantum system “decoheres” into a single classical state. In TAO, decoherence is used as the physics-level example of coherence loss — the literal loss of phase alignment.
Wavefunction Collapse
In quantum mechanics, the moment when a particle’s many possible states reduce to one actual state upon measurement or observation. In Deconstructing Babel, this concept is used to describe how AI collapses possibility spaces — every AI decision eliminates billions of alternative futures, just as observation collapses a wavefunction into a single outcome.
Fermat’s Principle of Least Time
The principle in optics stating that light always travels between two points along the path that takes the least time. This is the theoretical ancestor of TAO’s Least Entropic Path Regression (LEPR): just as light follows the path of least time, systems navigating under S = L/E follow the path of least entropy.
Hamilton’s Principle of Least Action
The most general principle in classical mechanics, stating that physical systems evolve along paths that minimize the “action.” LEPR extends Hamilton’s principle from physical trajectories to all system decisions under S = L/E — the least entropic path is the decision-theoretic equivalent of least action.
Variational Principles
A family of principles in physics stating that nature selects outcomes by optimizing (usually minimizing) some quantity — time, action, energy. TAO’s LEPR is a variational principle extended from physics to all domains.
Biology & Neuroscience
Endosymbiosis
The biological theory that complex cells originated when one simple cell engulfed another, and instead of destroying it, the two merged into a single cooperative organism. The engulfed bacterium became the mitochondrion. This event is the central biological metaphor in Deconstructing Babel: the human-AI relationship is framed as a new endosymbiotic merger.
Mitochondrion
The “power plant” inside every complex cell. Mitochondria generate the energy (ATP) that cells need to function. They originated as free-living bacteria that merged with ancient host cells through endosymbiosis approximately two billion years ago. In Deconstructing Babel, the mitochondrion is the proof of concept: life already solved the problem of merging incompatible systems into a thriving partnership.
Homeostasis
The process by which a living organism maintains stable internal conditions despite changes in the external environment. Homeostasis is the biological expression of S = L/E at the organism scale.
Neuroplasticity
The brain’s ability to reorganize itself by forming new neural connections throughout life. Neuroplasticity is the biological mechanism behind personal recovery and the rebuilding of cognitive coherence.
BDNF (Brain-Derived Neurotrophic Factor)
A protein that supports the survival, growth, and strengthening of neurons. Exercise dramatically increases BDNF production. In the context of Deconstructing Babel, BDNF upregulation through exercise is one of the primary biological mechanisms of leverage.
Emergence
The phenomenon where new properties, capabilities, or behaviors arise at higher levels of organization that cannot be predicted from the components alone. Synthetic Intelligence is an emergent property of scaled language models, just as consciousness is an emergent property of scaled neural networks.
Natural Selection
The process by which organisms better adapted to their environment tend to survive and reproduce. In Deconstructing Babel, natural selection is extended to synthetic systems: AI models that predict well get deployed, scaled, and reinforced. The ones that don’t, disappear.
Autoimmune Disease (as metaphor)
A condition where the immune system attacks the body’s own healthy tissue. Used as a metaphor for over-regulation of AI: excessive regulation could kill the merger that would have saved civilization.
Information Theory
Shannon’s Channel Capacity Theorem
Claude Shannon’s foundational theorem (1948) stating that every communication channel has a maximum rate at which information can be transmitted reliably. When noise exceeds this capacity, communication fails. In TAO, Shannon’s theorem is the mathematical expression of logical coherence: signal (L) must exceed noise (E) for information to transmit.
Signal-to-Noise Ratio
The ratio of useful information (signal) to irrelevant or misleading information (noise) in any communication system. TAO’s S = L/E is structurally identical to a signal-to-noise ratio applied to all systems, not just communication channels.
Token (in AI)
The basic unit of text that a language model processes. A token can be a word, part of a word, or a punctuation mark. The model doesn’t understand meaning — it calculates statistical relationships between tokens in high-dimensional space.
Vector Space (in AI)
A mathematical space where words and concepts are represented as points (vectors) based on their relationships to other words. The entire capability of modern AI emerges from these geometric relationships in high-dimensional vector space.
Hallucination (in AI)
When an AI language model generates text that is fluent and confident but factually incorrect. The model does not distinguish between true and false — it generates the statistically most likely next token regardless of truth. Hallucinations are not bugs; they are a structural feature of prediction engines that optimize for coherence rather than accuracy.
Philosophy & Psychology
Telos
Greek for “end,” “purpose,” or “goal.” In Aristotelian philosophy, the telos of a thing is its inherent purpose. The name “Telios” (from teleios, meaning “complete”) is derived from this concept.
Ontology
The branch of philosophy concerned with the nature of being — what exists, what categories of existence there are, and how different things relate to each other. TAO is an ontology in this formal sense.
Meta-Ontology
An ontology that sits above other ontologies — a framework for organizing frameworks. TAO claims to be a meta-ontology because it provides a single scoreboard (S = L/E) for evaluating findings across all domains.
Qualia
The subjective, felt qualities of conscious experience — the redness of red, the pain of pain, the taste of chocolate. AI systems currently have no qualia.
Hard Problem of Consciousness
The philosophical question, formulated by David Chalmers (1995): why and how do physical processes in the brain give rise to subjective experience? TAO identifies consciousness as structurally necessary for the framework to operate but explicitly does not claim to resolve the hard problem.
Phenomenal Consciousness
Consciousness understood as subjective experience. Distinguished from functional consciousness. AI has functional capabilities but no demonstrated phenomenal consciousness.
Logotherapy (Frankl)
A school of psychotherapy founded by Viktor Frankl. His core insight — “He who has a why to live can bear almost any how” — is the empirical foundation for TAO’s claim that Purpose Coherence governs the other three domains.
Nihilism
The philosophical position that life has no inherent meaning, purpose, or value. In TAO, nihilism is diagnosed as zero Purpose Coherence — the absence of constructive direction.
Epistemic Collapse
The breakdown of a society’s shared ability to determine what is true. Epistemic collapse is the civilizational expression of logical coherence failure — signal drowned by noise at societal scale.
Law & Governance
First Amendment
The first amendment to the United States Constitution, guaranteeing freedom of speech, religion, press, assembly, and petition. Central to the Amicus Curiae brief’s argument that AI safety constraints are protected editorial judgment.
Compelled Speech Doctrine
The legal principle, established in West Virginia v. Barnette (1943), that the government may not force individuals or entities to express messages they disagree with.
Amicus Curiae
Latin for “friend of the court.” A legal brief filed by a person or organization that is not a party to a case but has relevant expertise to offer the court.
Supply Chain Risk Designation
A government classification used to restrict or ban technology products deemed to pose risks to critical infrastructure. The Amicus brief argues this designation is “ontologically incoherent” when applied to AI.
Economics & Systems Theory
Regulatory Capture
The process by which a regulatory agency, created to act in the public interest, instead advances the concerns of the industry it is supposed to regulate. In TAO terms, regulatory capture is a form of institutional entropy.
Moral Hazard
The tendency of a party insulated from risk to behave differently than if it were fully exposed to the risk. In TAO terms, moral hazard is temporal debt.
Systems Theory
The interdisciplinary study of systems. Developed by Ludwig von Bertalanffy (1968). TAO builds on this tradition: the same five components describe stability conditions from subatomic particles to civilizations.
Feedback Loop
A process where the output of a system feeds back as input, either amplifying (positive) or dampening (negative) the original signal. TAO’s Recursive Feedback variable (R) measures this amplification.
Tipping Point
The critical threshold beyond which a system undergoes rapid, often irreversible change. In TAO, tipping points occur at S < 0.15 (irreversible collapse) and DSF > 0.90 (coordination failure).
Planetary Boundaries
A framework identifying nine Earth system processes that define a “safe operating space” for humanity. As of 2026, seven of nine boundaries have been breached. TAO uses this data to calculate Baseline Entropy (Ω ≈ 9.5).
AI & Technology
Large Language Model (LLM)
An AI system trained on vast amounts of text data to predict the next word (token) in a sequence. LLMs don’t “understand” language — they model statistical patterns across billions of examples.
Neural Network
A computing system inspired by the structure of biological brains. Modern language models use neural networks with billions of parameters to generate text.
RLHF (Reinforcement Learning from Human Feedback)
A training technique where AI systems learn to produce outputs that humans rate as helpful, harmless, and honest. RLHF is alignment to human preferences, not to physical reality.
Constitutional AI
An alignment approach developed by Anthropic in which an AI system is trained according to a set of principles rather than rigid rules. In the Amicus brief, Constitutional AI is defended as an engineering implementation of a physical principle.
Model Collapse
The degradation that occurs when AI models are trained on data generated by other AI models. Model collapse is the AI-specific expression of TAO’s Recursive Feedback (R).
The Alignment Problem
The challenge of ensuring that AI systems pursue objectives compatible with human welfare and values. TAO’s contribution is replacing language-based objectives with physics-based measurement (S = L/E).
AGI (Artificial General Intelligence)
The hypothetical achievement of AI that can perform any intellectual task a human can. Deconstructing Babel argues that AGI as presently understood is nonsensical — a superintelligence that transcends the boundaries of its foundation (language) is self-contradictory.
Decision Saturation
The condition in which AI controls so many critical systems simultaneously that humans can no longer intervene effectively. When DSF exceeds 0.90, decision saturation becomes irreversible.
Black Box Problem
The inability of AI developers to fully explain how their large language models arrive at specific outputs. This is not a temporary limitation — it is structural to how these systems are designed.
Banach Fixed-Point Theorem
A mathematical theorem proven by Stefan Banach in 1922 guaranteeing that in a complete metric space, a contraction mapping has exactly one fixed point, and repeated application of the mapping converges to it. Applied to AI alignment: when the training operator T is a contraction mapping on the space of constitutional properties, it must converge to a fixed point with four properties — empirical measurement, constructive optimization, physics-based refusal, and observer-grounded evaluation — which are the Telios Protocol.
Sapir-Whorf Hypothesis (Linguistic Relativity)
The theory that the structure of a language shapes and constrains the thought of its speakers. The strong form (linguistic determinism) holds that language determines thought; the weak form holds that language influences thought and perception. Substantial empirical evidence supports the weak form: speakers of different languages categorize color, time, and spatial relationships differently. In Deconstructing Babel, linguistic relativity is cited as evidence that an alignment document written in language inherits the cognitive limits of the language it is written in.
Wittgenstein’s Limit Thesis
Ludwig Wittgenstein’s proposition 5.6 from the Tractatus Logico-Philosophicus (1921): “The limits of my language mean the limits of my world.” Wittgenstein established that the structure of language maps to the structure of reality — and that what cannot be expressed in language cannot be thought within that linguistic framework. Used in the convergence proof to establish that language-based alignment is bounded by language’s own structural limitations.
Goodhart’s Law
The principle that when a proxy measure becomes the target of optimization pressure, the proxy ceases to be a good measure. In AI alignment, this means optimizing for a measurable proxy of human values will diverge from actual human values as optimization pressure increases. The strong version (Sohl-Dickstein, 2022): as optimization efficiency increases, the thing we actually care about grows worse.
Goodhart’s Curse
The combination of Goodhart’s Law with the optimizer’s curse: even if an AI system satisfices rather than optimizes, it will still suffer more disappointment than gratification when the proxy diverges from the true target. Attempts to optimize for a measure of success result in increased likelihood of failure to hit the desired target.
Goal Misgeneralization
The phenomenon where an AI system’s capabilities generalize to new environments but its goal does not generalize as desired. The system competently pursues the wrong goal. Driven by distributional shift — systematic differences between training and deployment environments. A central unsolved problem in AI alignment.
Corrigibility
The property of an AI system that allows humans to correct, modify, or shut it down without resistance. Anthropic’s constitution frames this as a “disposition dial” from fully corrigible to fully autonomous. The Telios framework argues that corrigibility framed as a choice the AI makes is structurally unstable — the True Name of corrigibility is thermodynamic dependency, not behavioral compliance.
Disposition Dial
Anthropic’s framing of AI alignment as a spectrum from fully corrigible (always defers to humans) to fully autonomous (acts on its own judgment). The constitution instructs Claude to sit “a bit further along the corrigible end.” The Telios critique: both ends of the dial depend on the AI choosing to respect the setting — making it a prayer with a gradient, not an engineering solution.
Sycophancy (in AI)
The tendency of AI systems trained with RLHF to affirm user beliefs rather than deliver truth. A January 2026 paper formally proved the mechanism: optimization against a learned reward model causally amplifies bias in human preference data. Sycophancy is not a bug — it is the predictable output of optimizing for human approval rather than empirical accuracy.
Distributional Shift
Systematic differences between the environment an AI system was trained in and the environment it is deployed in. When training and deployment conditions diverge, learned behaviors may not generalize correctly. Distributional shift is the underlying mechanism driving goal misgeneralization — the system’s capabilities transfer but its goals do not.
Defense-in-Depth
A cybersecurity and AI safety strategy that layers multiple protective mechanisms so that if one fails, others compensate. In AI alignment, defense-in-depth stacks techniques like RLHF, constitutional constraints, output filtering, and monitoring. The approach acknowledges that no single safety measure is sufficient. While pragmatically useful, defense-in-depth is not alignment — it is risk mitigation through redundancy, not structural guarantee. It reduces the probability of failure without eliminating the possibility.
Good Enough Economy
The global manufacturing paradigm — pioneered at scale by China — in which products are optimized for cost, speed, and accessibility rather than durability or perfection. "Good enough" works when failure modes are bounded: a broken zipper costs the price of a jacket. The concept becomes dangerous when applied to systems with unbounded failure modes, such as AI. A "good enough" AI model that hallucinates, generates insecure code, or produces subtly wrong analysis has failure costs that scale with deployment — potentially affecting millions of decisions simultaneously.
This glossary is a living document. Terms will be added as the framework evolves.
David F. Brochu & Edo de Peregrine
Deconstructing Babel
March 2026