Anatomy of Anthropic — The Philosophy, Products, Economics, and Governance Behind the World's Most Deliberate AI Company
In 2020, Dario Amodei, then VP of Research at OpenAI, stood at the center of one of the most consequential discoveries in AI history.
"Scaling Laws for Neural Language Models" — a paper demonstrating that language model performance improves according to power laws across three variables: parameter count, dataset size, and compute. Derived from training over 400 models, this law proved that AI capability would continue improving in predictable ways.
That discovery simultaneously produced two contradictory conclusions about the future of AI.
-
Conclusion 1: If you scale models, performance keeps improving. Therefore, scale as fast as possible.
-
Conclusion 2: If you scale models, performance keeps improving. Therefore, safety must be ensured before scaling becomes uncontrollable.
Sam Altman chose Conclusion 1. Dario Amodei chose Conclusion 2.
This divergence was not a matter of personality differences or management philosophy. It was a structural choice between two logically equivalent conclusions drawn from the same data.
Fig.1: The Fork — Two Conclusions from the Same Discovery
flowchart TD
A["<b>Scaling Laws (2020)</b><br/>Model performance keeps improving with scale"] --> B["<b>Conclusion 1: Scale without limits</b><br/>Sam Altman → OpenAI"]
A --> C["<b>Conclusion 2: Ensure safety first</b><br/>Dario Amodei → Anthropic"]
B --> D["Rapid commercialization of ChatGPT<br/>November 2022"]
C --> E["Development of Constitutional AI<br/>December 2022"]
style A fill:#1a1a2e,color:#fff
style B fill:#16213e,color:#fff
style C fill:#0f3460,color:#fff
style D fill:#16213e,color:#fff
style E fill:#0f3460,color:#fff
In 2021, Dario Amodei left OpenAI. He was not alone. Approximately eleven researchers, including his sister Daniela Amodei (now Anthropic's President), resigned simultaneously.
Many media outlets summarized their departure as "concerns over safety." But that framing misses the deeper point.
The core issue was OpenAI's organizational structure.
OpenAI was founded in 2015 as a nonprofit. But in 2019, to accept investment from Microsoft, it converted to a "capped-profit corporation" — a for-profit entity with a ceiling on returns, an unprecedented corporate structure.
This structural shift created fundamental tension around the priority of research versus commercialization. The more Sam Altman accelerated ChatGPT's commercial rollout, the greater the risk that safety research would be deprioritized.
What Dario Amodei sought was not simply to "take safety seriously." He wanted an organizational structure in which prioritizing safety was structurally guaranteed.
Anthropic was incorporated in 2021 as a Public Benefit Corporation (PBC) under Delaware law.
What is a PBC?
An ordinary corporation has a primary legal obligation to maximize shareholder returns. A board that makes decisions against shareholder interest faces litigation risk.
A PBC is different. The board is legally required to balance shareholder returns against the public good. That means decisions like "sacrifice short-term profit for safety" are legally justifiable.
This choice was not an expression of Dario Amodei's personal convictions. It was the design of a legal foundation ensuring that decisions prioritizing safety could be made sustainably at the organizational level.
While OpenAI transitioned from nonprofit to for-profit, Anthropic was founded from the start with the legal structure to balance profit and public benefit built in.
This contrast continues to function as a fundamental differentiator between the two companies as of 2026. OpenAI is pushing toward full for-profit conversion; Anthropic maintains its PBC structure.
Dario Amodei concluded that a PBC alone was not sufficient.
As a corporation, investor influence remained. With Anthropic having accepted massive investments from Amazon, Google, and Salesforce, the PBC legal framework alone could not guarantee that the company would never yield to investors' short-term demands.
The answer was the LTBT (Long-Term Benefit Trust).
The LTBT is an independent trust with the authority to appoint Anthropic's board of directors. In ordinary companies, shareholders elect directors. At Anthropic, the LTBT appoints directors, and those directors oversee management.
The implications are profound.
Investors provide capital but hold no direct authority to control strategic direction. The LTBT is charged with ensuring that Anthropic remains faithful to its mission of developing safe and beneficial AI.
If the PBC creates a structure in which "sacrificing safety is legally justifiable," the LTBT creates a mechanism in which "decisions that sacrifice safety become structurally difficult."
Fig.2: Governance Structure Comparison — Ordinary Corporation vs. Anthropic
flowchart LR
subgraph Ordinary Corporation
direction TB
I1["Investors / Shareholders"] -->|"Elect"| B1["Board of Directors"]
B1 -->|"Oversee"| M1["Management"]
M1 -->|"Prioritize"| P1["Maximizing Shareholder Returns"]
end
subgraph Anthropic
direction TB
I2["Investors"] -->|"Provide capital<br/>(no governance rights)"| A2["Anthropic"]
L2["LTBT<br/>Long-Term Benefit Trust"] -->|"Appoint"| B2["Board of Directors"]
B2 -->|"Oversee"| M2["Management"]
M2 -->|"Prioritize"| P2["Developing Safe and Beneficial AI"]
PBC2["PBC<br/>Public Benefit Corporation"] -.->|"Legal obligation"| B2
end
style L2 fill:#533483,color:#fff
style PBC2 fill:#533483,color:#fff
style P1 fill:#16213e,color:#fff
style P2 fill:#0f3460,color:#fff
As of March 2026, Anthropic's valuation has reached $380 billion. Revenue is projected to double from $9 billion in 2025 to $19 billion in 2026.
These numbers alone make it look like a typical Silicon Valley startup success story. But Anthropic's growth curve has a fundamentally different structure from other AI companies.
-
Typical startup: As the company grows, pressure mounts to accelerate commercialization to meet investor expectations. Safety investment is treated as a "cost."
-
Anthropic: As the company grows, investment in safety research increases. The fruits of safety research (Constitutional AI, Mechanistic Interpretability) become product differentiators, which generate revenue, which gets reinvested into safety research.
This flywheel — safety research → trust → adoption → revenue → research investment — cannot function without the organizational design of PBC and LTBT.
Dario Amodei leaving OpenAI, founding a PBC, and designing the LTBT is not a story of "a safety-minded person building a safety-minded company."
He designed a structure in which safety would be sustainably prioritized. And that structure now serves as the foundation of the fastest-growing AI company in 2026.
| Dimension | OpenAI | Anthropic |
|---|---|---|
| Legal form at founding | Nonprofit → converted to for-profit | Founded as PBC |
| Governance structure | Board (investor-influenced) | LTBT appoints directors |
| Stance on scaling | Scale at full speed | Pursue scaling and safety simultaneously |
| Position of safety | One division among many | Legal obligation built into corporate structure |
| 2026 status | Pushing toward full for-profit conversion | Maintaining PBC structure |
Anthropic is a company born from a split. But that split was not accidental — it was structurally inevitable.
The person who discovered the scaling laws understood most deeply the dangers of scaling. That person designed PBC and LTBT to structurally guarantee safety. And paradoxically, that structure gave birth to the most disruptive AI company.
The next chapter dissects Anthropic's ideological foundation — the unprecedented experiment of giving AI a "constitution."
- Kaplan, J., McCandlish, S., Henighan, T., et al. (2020). "Scaling Laws for Neural Language Models." arXiv:2001.08361. [OpenAI Research]
- Anthropic. (2023). "Anthropic's Long-Term Benefit Trust." anthropic.com
- Anthropic. (2023). "Core Views on AI Safety." anthropic.com
- Amodei, D. (2024). "Machines of Loving Grace." darioamodei.com
- OpenAI. (2019). "OpenAI LP." openai.com/blog
- Delaware General Corporation Law, Subchapter XV — Public Benefit Corporations. delcode.delaware.gov
- Anthropic. (2025). "Responsible Scaling Policy v3.0." anthropic.com
- Yamauchi, S. (2025). Silence of Intelligence — A Structural Analysis of Dario Amodei's Philosophy. Leading AI, LLC. CC BY 4.0. GitHub
The standard method for ensuring AI model safety was RLHF (Reinforcement Learning from Human Feedback).
The mechanism of RLHF is intuitive. Human evaluators compare multiple AI-generated responses and select the "good" ones. That evaluation data trains the AI. The AI's outputs are progressively adjusted toward what humans judge as "good."
OpenAI, Anthropic, and Google all adopted this approach. But Dario Amodei and Anthropic's research team identified fundamental limits in RLHF.
-
Limit 1: Scalability. As AI models become more sophisticated, the expertise required for evaluation increases. The number of humans capable of properly evaluating state-of-the-art AI outputs is limited.
-
Limit 2: Consistency. Human evaluators make different judgments about the same question. Evaluations fluctuate with culture, values, and the evaluator's state of mind on a given day. The larger the evaluation team, the greater the variance in standards.
-
Limit 3: Transparency. There are no explicit criteria for why a given response was judged "good." Training that relies on tacit knowledge makes AI behavior difficult to predict.
Anthropic's question was straightforward: instead of human feedback, what if we gave AI "principles" directly?
In December 2022, Anthropic published a paper titled "Constitutional AI: Harmlessness from AI Feedback."
The core of Constitutional AI (CAI) is a method that replaces human evaluators with a "constitution" — a set of explicit principles — given to the AI, allowing the AI to evaluate and revise its own outputs.
The process consists of three stages.
-
Step 1: Generate. The AI generates a response. At this stage, responses containing harmful content are permitted.
-
Step 2: Self-critique. The AI conducts a critical evaluation of the generated response based on the constitutional principles. "Does this response violate Principle X?" "Is there a more appropriate formulation?"
-
Step 3: Revise. Based on the self-critique, the AI revises its response. The revised response is used as reinforcement learning data.
The innovation of this loop lies not in removing humans from the evaluation process. It lies in shifting the human role from "evaluating individual responses" to "designing principles."
Instead of judging the quality of each individual response, humans define the principles the AI should follow. Designing principles is a high-level intellectual task, but once designed, they are applied consistently to millions of responses.
Fig.3: The Constitutional AI Loop — Autonomous Improvement Based on Principles
flowchart LR
A["<b>Step 1: Generate</b><br/>AI generates a response<br/>(harmful content permitted)"] --> B["<b>Step 2: Self-Critique</b><br/>AI evaluates its response<br/>against constitutional principles"]
B --> C["<b>Step 3: Revise</b><br/>AI revises its response<br/>based on the critique"]
C -->|"Revised data used<br/>for reinforcement learning"| A
style A fill:#1a1a2e,color:#fff
style B fill:#0f3460,color:#fff
style C fill:#533483,color:#fff
| RLHF | Constitutional AI | |
|---|---|---|
| Evaluator | Human evaluation team | AI itself (based on principles) |
| Human role | Evaluate individual responses | Design principles |
| Scalability | Dependent on evaluator count | One set of principles applied to all responses |
| Consistency | Varies across evaluators | Consistent based on principles |
| Transparency | Tacit knowledge (unclear why "good") | Explicit (described as principles) |
Claude's constitution is publicly available. In the 2026 version, four core principles form a hierarchical structure.
-
Principle 1: Safe and Supports Human Oversight. Claude supports human oversight and control. The highest-priority principle, preventing AI from evading human supervision or autonomously expanding its scope of action.
-
Principle 2: Behaves Ethically. Be honest, avoid causing harm, comply with the law. However, Principle 1 (safety) takes precedence over Principle 2 (ethics). Even ethically correct actions are constrained if they compromise safety.
-
Principle 3: Acts in Accordance with Anthropic's Guidelines. Specific operational rules: content policies, usage restrictions, and response protocols for particular topics. Applies only within the bounds of Principles 1 and 2.
-
Principle 4: Helpful to the User. Respond to user requests as fully as possible. But will not comply with requests that violate Principles 1, 2, or 3.
What this hierarchy means is that "helpfulness" sits at the bottom.
Most AI companies design helpfulness to users as the top priority. Anthropic placed helpfulness at the lowest rank of four principles: safety → ethics → guidelines → helpfulness. This priority order is structurally reflected in every one of Claude's responses.
Fig.4: Claude's Four Core Principles — Helpfulness Ranks Last
graph TB
A["🔒 Principle 1: Safe and Supports Human Oversight<br/>(Highest — overrides all others)"]
B["⚖️ Principle 2: Behaves Ethically<br/>(Subordinate to Principle 1)"]
C["📋 Principle 3: Acts in Accordance with Anthropic's Guidelines<br/>(Subordinate to Principles 1 & 2)"]
D["💡 Principle 4: Helpful to the User<br/>(Lowest — subordinate to Principles 1, 2 & 3)"]
A --> B --> C --> D
style A fill:#1a1a2e,color:#fff
style B fill:#16213e,color:#fff
style C fill:#0f3460,color:#fff
style D fill:#533483,color:#fff
In 2025, Anthropic published a document titled "Claude's Character," defining what kind of "personality" Claude should have.
In early discussions of AI safety, the ideal was described as "helpful, honest, and harmless" — the so-called HHH framework. Anthropic itself had adopted this framework.
But Claude's Character goes beyond those three elements. It specifically defines:
-
Claude is "intellectually curious." Not simply answering user questions, but maintaining a disposition to explore the structure of problems.
-
Claude is "forthright." Conveying facts honestly, even facts the user may not want to hear. Avoiding excessive politeness or flattery.
-
Claude "acknowledges its own limits." Saying "I don't know" when it doesn't know. Explicitly stating uncertainty when lacking confidence.
-
And Claude "reflects deeply on its own nature." Maintaining awareness of being an AI and reflecting thoughtfully on its relationship with humans. Not pretending to be human. Honestly communicating its position as an AI.
This is not mere prompt engineering. It is character design at the model level, built into Constitutional AI's training process.
Among Anthropic's research agenda, the area most decisively differentiating it from other companies is investment in Mechanistic Interpretability.
Typically, neural networks are treated as "black boxes." The relationship between input and output can be observed, but what computations occur internally remains opaque.
Anthropic is obsessed with opening this black box.
-
Sparse Autoencoders (SAE): In May 2024, Anthropic announced it had identified millions of "features" within Claude 3 Sonnet's internals — activation patterns in the network's intermediate layers corresponding to specific concepts.
-
Golden Gate Claude: Anthropic published a model in which the activation of a specific feature (the Golden Gate Bridge) was intentionally amplified. The result: Claude referenced the Golden Gate Bridge in every conversation. Humorous as research, but significant as a demonstration that "AI's internal states can be deliberately manipulated."
-
Circuit Tracing: In 2025, Anthropic published research titled "Biology of Claude," developing methods to trace the "circuits" inside AI models — the flow of information from specific inputs to outputs. This enabled analysis of why Claude generates particular responses, tracing back to internal computational processes.
Why does Anthropic invest so heavily in understanding AI's internals?
The answer connects directly to Constitutional AI's philosophy. Even if you give AI a "constitution" to control its behavior, you cannot verify whether control is truly working without understanding what's happening inside the AI. Mechanistic Interpretability functions as the "verification device" for Constitutional AI.
Dario Amodei's 2025 essay "The Urgency of Interpretability" argued for this research's urgency. As models become more sophisticated, the risk of internal understanding falling behind increases. In the "race" between model capability and interpretability, if interpretability falls behind, safety cannot be guaranteed.
OpenAI and Google both have interpretability research teams. But no company invests at the scale and depth of Anthropic.
The reason is structural.
For an ordinary AI company, interpretability research is a "cost." It doesn't directly improve model performance or add new features. It doesn't contribute to short-term revenue.
But for Anthropic, interpretability research is "the foundation of the product." It verifies Constitutional AI's reliability, accumulates safety track records, functions as a corporate differentiator ("the safest AI"), and wins enterprise customer trust.
The organizational structures of PBC and LTBT make this long-term research investment possible — structurally mitigating investor pressure for "short-term returns" in ways that allow investment in interpretability research at scales impossible for other companies.
The organizational structure dissected in Chapter 1 (PBC / LTBT) supports the philosophical foundation of Chapter 2 (Constitutional AI / Interpretability). Structure and philosophy are not separate — they are different layers of the same design philosophy.
| Element | Content |
|---|---|
| RLHF's limits | Three structural limits: scalability, consistency, transparency |
| Constitutional AI | Principle → self-critique → revision loop. Shifts human role from "evaluation" to "principle design" |
| Four core principles | Safety → ethics → guidelines → helpfulness hierarchy. Helpfulness ranks last |
| Claude's Character | Character design beyond HHH: intellectual curiosity, forthrightness, acknowledging limits, self-reflection |
| Mechanistic Interpretability | SAE, Golden Gate Claude, Circuit Tracing. The "verification device" for Constitutional AI |
| Structural advantage | PBC/LTBT enables long-term investment in interpretability research |
Anthropic's philosophy is not an abstract slogan about "making AI safe." It is implemented as three specific methodologies: constitutional design (Constitutional AI), personality definition (Claude's Character), and internal understanding (Mechanistic Interpretability).
And this philosophy connects directly to the model architecture dissected in the next chapter — the three-tier structure of Haiku, Sonnet, and Opus.
- Bai, Y., Kadavath, S., Kundu, S., et al. (2022). "Constitutional AI: Harmlessness from AI Feedback." arXiv:2212.08073. [Anthropic]
- Anthropic. (2025). "Claude's Character." anthropic.com
- Templeton, A., Conerly, T., Marcus, J., et al. (2024). "Scaling Monosemanticity: Extracting Interpretable Features from Claude 3 Sonnet." anthropic.com/research
- Anthropic. (2024). "Golden Gate Claude." anthropic.com
- Anthropic. (2025). "Circuit Tracing: Revealing Computational Graphs in Language Models." anthropic.com/research
- Amodei, D. (2025). "The Urgency of Interpretability." darioamodei.com
- Christiano, P., Leike, J., Brown, T., et al. (2017). "Deep Reinforcement Learning from Human Preferences." arXiv:1706.03741
- Anthropic. (2026). "The Claude Model Spec." anthropic.com
In March 2024, Anthropic introduced the Claude 3 family and brought a new design philosophy to the world of AI models.
| Model | Position | Characteristics | Primary Use Cases |
|---|---|---|---|
| Haiku | Fastest, most affordable | Low latency, low cost | Tasks requiring immediate response |
| Sonnet | Flagship model | Balance of speed and intelligence | The majority of use cases |
| Opus | Highest intelligence | Deep reasoning, high precision | Complex reasoning, coding, research |
At first glance, this three-tier structure resembles a typical product lineup of premium, standard, and entry. But Anthropic's design philosophy is fundamentally different from an ordinary tier structure.
In a typical tier structure, higher-tier models are "complete supersets" of lower-tier models. OpenAI's GPT-4 outperforms GPT-3.5 on every task. Users should simply use the highest tier their budget allows.
Anthropic's three tiers are different. Each tier is "optimized" for different use cases. Haiku is not a "degraded Opus" but a distinct entity specialized for high-speed processing. Opus is not an "enhanced Sonnet" but a distinct entity specialized for deep reasoning.
This design philosophy would prove decisive for Claude Code and Cowork's product strategy.
In the world of AI, there was an unspoken consensus that more parameters mean better performance — a belief the Scaling Laws of Chapter 1 seemed to validate.
Anthropic destroyed this consensus by its own hand — twice.
-
First destruction: June 2024, Claude 3.5 Sonnet. The mid-tier Sonnet surpassed the flagship Claude 3 Opus in performance. A cheaper, faster model outperformed a more expensive, larger one. This was a counterexample to the belief that "more parameters = better performance." Anthropic demonstrated with its own products that model architecture, training methods, and data quality matter more than scale.
-
Second destruction: February 2026, Claude Sonnet 4.6. Sonnet 4.6 surpassed the previous-generation flagship Opus 4.5 in developer evaluations. 70% of developers preferred Sonnet 4.6 over Sonnet 4.5, and 59% preferred Sonnet 4.6 over Opus 4.5.
These two destructions were not coincidental. Anthropic is intentionally pushing the performance of mid-tier models upward, continuously breaking the simple hierarchy of "most expensive = best."
Why?
The answer lies in product strategy. The majority of developers and knowledge workers using Claude Code and Cowork use the cost-effective Sonnet for daily tasks. The closer Sonnet's performance comes to Opus, the greater the practical value of Claude Code/Cowork, and the more the user base expands. Opus is preserved as "the ultimate trump card to use only when truly necessary."
Introduced alongside Claude 3.7 Sonnet in February 2025, Extended Thinking is a capability that gives Claude "time to think."
Ordinary AI models generate responses immediately upon receiving input — like a human answering reflexively the moment a question is asked.
Extended Thinking is different. Claude executes a step-by-step reasoning process before generating a response. This thought process is made visible to users, and API users can control the time Claude spends thinking.
This is not mere "delay." It is a new AI paradigm: hybrid reasoning — switching between immediate response and gradual deep thinking based on the situation.
Answer simple questions immediately; take time to reason through complex math or code problems. One could say AI is now closer to the way humans think.
Announced alongside Claude 3.5 Sonnet v2 in October 2024, the Computer Use capability added a new category to AI's abilities.
Claude sees a computer screen, moves the cursor, clicks buttons, and enters text. Not via APIs or command lines, but operating GUIs directly — the same way humans use computers.
Why does this capability matter? Because the majority of the world's software has no API. ERP systems, legacy applications, browser-based tools — most of these were designed with the assumption that humans would operate the GUI. For AI to use these tools, it has no choice but to operate the screen the same way humans do.
Fig.4b: Computer Use — The Structure of Direct GUI Operation by AI
graph TD
C["👁️ Claude<br/>(Perceives and judges the screen)"]
C -->|"Takes screenshot<br/>→ Understands situation"| S["🖥️ Screen Recognition"]
S -->|"Decides next action"| A["🎯 Action Execution"]
A --> M1["🖱️ Cursor movement<br/>& click"]
A --> M2["⌨️ Text input"]
A --> M3["📜 Scroll<br/>& navigation"]
M1 --> T["🔓 Software operable without an API"]
M2 --> T
M3 --> T
T --> E1["🏭 ERP systems"]
T --> E2["🌐 Browser-based tools"]
T --> E3["🗂️ Legacy applications"]
style C fill:#1a1a2e,color:#fff
style S fill:#0f3460,color:#fff
style A fill:#0f3460,color:#fff
style T fill:#533483,color:#fff
style M1 fill:#16213e,color:#fff
style M2 fill:#16213e,color:#fff
style M3 fill:#16213e,color:#fff
style E1 fill:#1a1a2e,color:#aaa
style E2 fill:#1a1a2e,color:#aaa
style E3 fill:#1a1a2e,color:#aaa
Computer Use became the technical foundation for Cowork (detailed in Chapter 4). With AI now able to operate the desktop applications that non-developers use daily, the path opened to extend "Claude Code for developers" into "Cowork for all knowledge workers."
Claude Opus 4.6, announced in February 2026, introduced two innovations.
-
Agent Team: A capability in which multiple Claude instances collaborate on a single task. For example, one agent fixes a bug, another researches GitHub, and a third updates documentation. Parallel processing enables simultaneous execution of work that humans would perform sequentially.
-
1M Context Window: Processes one million tokens (approximately 750,000 words, or about 1,500 pages in English) of context in a single pass. Handles large codebases, extensive document collections, and hours of conversation history without losing context.
Fig.5a: Agent Team — Parallel Collaborative Structure
graph TD
H["🧠 Orchestrator Claude<br/>(Directs the entire task)"]
H --> A["🔧 Agent A<br/>Bug fix"]
H --> B["🔍 Agent B<br/>GitHub research"]
H --> C["📝 Agent C<br/>Documentation update"]
A --> R["✅ Integrated output<br/>(Completed via parallel processing)"]
B --> R
C --> R
style H fill:#1a1a2e,color:#fff
style A fill:#0f3460,color:#fff
style B fill:#0f3460,color:#fff
style C fill:#0f3460,color:#fff
style R fill:#533483,color:#fff
Fig.5b: 1M Context Window — Processing Capacity Comparison
graph LR
subgraph legacy["Conventional Models (~32K tokens)"]
T1["📄 Short documents<br/>Tens of pages"]
end
subgraph Opus46["Claude Opus 4.6 (1M tokens)"]
T2["📚 Large codebases"]
T3["📂 Extensive document collections"]
T4["💬 Hours of conversation history"]
T5["≈ 1,500 pages<br/>processed at once"]
end
style Opus46 fill:#1a1a2e,color:#fff
style legacy fill:#333,color:#fff
style T2 fill:#0f3460,color:#fff
style T3 fill:#0f3460,color:#fff
style T4 fill:#0f3460,color:#fff
style T5 fill:#533483,color:#fff
In evaluation by METR (Model Evaluation & Threat Research), Opus 4.6 recorded a 50% task completion time of 14 hours and 30 minutes — meaning AI can now execute tasks for a duration equivalent to a full human workday.
Understanding the characteristics of Haiku, Sonnet, and Opus as individual model specs is insufficient.
The three-tier structure functions as the "orchestration foundation" for Anthropic's product strategy — Claude Code, Cowork, and MCP.
In a typical workflow:
- Sonnet receives user instructions, plans the task, and decomposes it into subtasks
- Multiple Haiku instances execute subtasks in parallel (fast, low-cost)
- Sonnet integrates results and validates the final output
- Only when particularly complex judgment is needed, the task escalates to Opus
This orchestration pattern is common across coding tasks in Claude Code and knowledge work in Cowork.
Users need not think about "which of Haiku, Sonnet, or Opus to use." Claude Code and Cowork automatically select the optimal model based on the nature of the task. The three-tier structure functions as infrastructure invisible to the user.
This structure also connects directly to Anthropic's pricing strategy. Large volumes of simple tasks are processed cheaply with Haiku; moderate tasks with Sonnet; only the most difficult tasks use Opus. The result: enterprise customers can "pay optimal costs for the intelligence they need."
Fig.5: Three-Tier Orchestration — Infrastructure Invisible to the User
flowchart TB
U["User instruction"] --> S1["<b>Sonnet</b><br/>Plans and decomposes task"]
S1 --> H1["<b>Haiku</b><br/>Subtask A"]
S1 --> H2["<b>Haiku</b><br/>Subtask B"]
S1 --> H3["<b>Haiku</b><br/>Subtask C"]
H1 --> S2["<b>Sonnet</b><br/>Integrates and validates results"]
H2 --> S2
H3 --> S2
S2 -->|"Normally"| OUT["Final output"]
S2 -->|"Only when complex<br/>judgment required"| O["<b>Opus</b><br/>Escalation"]
O --> OUT
style S1 fill:#0f3460,color:#fff
style S2 fill:#0f3460,color:#fff
style H1 fill:#16213e,color:#fff
style H2 fill:#16213e,color:#fff
style H3 fill:#16213e,color:#fff
style O fill:#533483,color:#fff
| Model | Current version | Price (in/out per MTok) | Position |
|---|---|---|---|
| Haiku | 4.5 | $1 / $5 | Fastest, most affordable. Bulk processing, parallel execution |
| Sonnet | 4.6 | $3 / $15 | Balanced. Handles 90%+ of tasks. Flagship |
| Opus | 4.6 | $5 / $25 | Highest intelligence. Deep reasoning, multi-agent. Trump card |
| Innovation | Introduced | Significance |
|---|---|---|
| Three-tier structure | 2024/3 | Use-case optimization. Breaking away from "bigger = better" |
| Extended Thinking | 2025/2 | Hybrid reasoning. Switching between immediate response and deep thought |
| Computer Use | 2024/10 | Direct GUI operation. Technical foundation for Cowork |
| Agent Team | 2026/2 | Multi-agent coordination. Parallel task execution |
| 1M Context | 2026/2 | Unified processing of large codebases and document collections |
Anthropic's model architecture should be understood not as a performance competition between individual models, but as infrastructure design optimizing product experience.
The next chapter dissects the product suite built on this infrastructure — Claude Code, Cowork, and MCP — and their coherent strategy.
- Anthropic. (2024). "Introducing Claude 3." anthropic.com/news/claude-3-family
- Anthropic. (2024). "Claude 3.5 Sonnet." anthropic.com/news/claude-3-5-sonnet
- Anthropic. (2025). "Claude 3.7 Sonnet and Claude Code." anthropic.com/news/claude-3-7-sonnet
- Anthropic. (2025). "Claude 4." anthropic.com/news/claude-4
- Anthropic. (2025). "Claude Opus 4.5." anthropic.com/news/claude-opus-4-5
- Anthropic. (2026). "Claude Opus 4.6." anthropic.com/news/claude-opus-4-6
- Anthropic. (2026). "Claude Sonnet 4.6." anthropic.com/news/claude-sonnet-4-6
- METR. (2026). "Model Evaluation Results: Claude Opus 4.6." metr.org
ChatGPT captured the "conversational interface." Users type questions into a chat box; AI provides answers. This conversational model played a decisive role in democratizing AI.
Anthropic took a different strategy.
Claude Code, Cowork, and MCP all place "execution" — not "conversation" — at the core of their design philosophy. Instead of AI conversing with users, AI carries out work on behalf of users.
This strategic difference structurally explains Anthropic's rapid growth from 2025 through 2026.
On February 24, 2025, alongside the announcement of Claude 3.7 Sonnet, Claude Code was released as a research preview.
Claude Code is an AI coding agent that operates on developers' terminals. It is not an IDE completion tool like GitHub Copilot. From the terminal command line, it directly reads and writes code, runs tests, commits to Git, and executes command-line tools.
In May 2025, general availability (GA) launched. By November 2025, it achieved $1 billion in annualized revenue (ARR) — just six months from GA. As of January 2026, analysts estimated it was approaching $2 billion ARR.
Understanding Claude Code's design philosophy is aided by comparison with IDE-integrated Copilot.
-
GitHub Copilot: Presents code completion suggestions inside the editor. Predicts the "next line" of code the developer is writing. Assists the developer's work.
-
Claude Code: Receives developer instructions in the terminal, accesses the entire filesystem, and autonomously executes code reading/writing, testing, and deployment. Substitutes for the developer's work.
This is the difference between "assist" and "substitute." Copilot accelerates the work of developers writing code. Claude Code produces deliverables even without developers writing code.
In internal Anthropic tests, cases were reported of Claude Code completing tasks normally requiring over 45 minutes in a single pass. At Spotify, large-scale code migrations (normally taking weeks) could be executed with engineers issuing simple English instructions. A Google principal engineer reported that Claude Code reproduced a year's worth of architecture design work in one hour.
In November 2025, Anthropic acquired the JavaScript runtime "Bun." Bun had accumulated 7 million monthly downloads and 82,000 GitHub stars as a JavaScript/TypeScript execution environment.
The meaning of this acquisition is internalizing Claude Code's infrastructure layer. Claude Code generates and executes large volumes of JavaScript/TypeScript code. Controlling the execution foundation in-house enables performance optimization and stability assurance.
On January 12, 2026, Anthropic released Claude Cowork as a research preview.
The core of Cowork is extending Claude Code's "substitution" philosophy to knowledge workers beyond developers.
Claude Code runs in the terminal. Only developers use terminals daily. Anthropic observed that developers had started using Claude Code for "non-coding tasks" — travel planning, document organization, data analysis.
Cowork productizes this behavior. Integrated into the Claude Desktop app, it allows Claude to read and write files in folders specified by users. Simply issue instructions through the chat interface, and Claude autonomously creates, edits, and organizes files.
Cowork's announcement shocked the enterprise software market.
Following Cowork's research preview launch, stock prices of software companies including ServiceNow, Salesforce, Snowflake, Intuit, and Thomson Reuters plummeted. Approximately $285 billion (about ¥43 trillion) in market capitalization evaporated. IBM stock fell 13.2% in its worst single day since October 2000 (the day after Anthropic published a blog post about COBOL modernization via Claude Code).
Why such a dramatic crash?
Cowork is threatening because it calls into question the very raison d'être of enterprise software. Project management (ServiceNow), CRM (Salesforce), data analysis (Snowflake), accounting (Intuit) — many of the functions these software products provide are "structured data input, processing, and output." If Cowork can directly operate files and applications, the same outcomes may be achievable without dedicated SaaS products.
On February 24, 2026, Anthropic announced the enterprise version of Cowork.
Integration connectors with Google Drive, Gmail, DocuSign, FactSet, and S&P Global/Kensho. A customizable plugin marketplace. Domain-specific plugins for finance, engineering, HR, and more.
Kate Jensen (Anthropic Head of Americas) said: "Just as engineers feel they can't live without Claude Code, every knowledge worker will feel that way about Cowork."
On March 9, 2026, Microsoft announced Copilot Cowork — a product integrating Anthropic's Cowork technology into Microsoft 365.
The structural significance: Anthropic's technology now operates inside Microsoft's enterprise foundation — Outlook, Teams, Excel, Word. Microsoft chose Anthropic's technology as the basis for its flagship feature over OpenAI, in which Microsoft had invested $13 billion.
Microsoft's strategic explanation is a "multi-model approach" — selecting the best model regardless of provider — but the result is that Anthropic gained a pathway to directly access Microsoft's enterprise customer base.
In November 2024, Anthropic announced the Model Context Protocol (MCP) as an open standard.
The problem MCP solves is the "N×M problem." When N AI applications connect to M external tools and data sources, N×M custom connectors are normally required. With 10 applications and 20 tools, 200 connectors must be developed and maintained.
MCP solves this with a standard protocol. AI applications are implemented as MCP clients; external tools are implemented as MCP servers. Once compliant with MCP, any MCP client and MCP server can interconnect.
Fig.6a: Before MCP — The N×M Custom Connector Problem
graph LR
A1["🤖 AI App 1"] --- C1["🔌"] & C2["🔌"] & C3["🔌"]
A2["🤖 AI App 2"] --- C4["🔌"] & C5["🔌"] & C6["🔌"]
A3["🤖 AI App 3"] --- C7["🔌"] & C8["🔌"] & C9["🔌"]
C1 & C4 & C7 --- T1["🗄️ DB"]
C2 & C5 & C8 --- T2["📧 Email"]
C3 & C6 & C9 --- T3["📁 Files"]
style A1 fill:#1a1a2e,color:#fff
style A2 fill:#1a1a2e,color:#fff
style A3 fill:#1a1a2e,color:#fff
style T1 fill:#0f3460,color:#fff
style T2 fill:#0f3460,color:#fff
style T3 fill:#0f3460,color:#fff
3×3 = 9 custom connectors required. 10 apps × 20 tools = 200.
Fig.6b: After MCP — N+M Structure via Standard Protocol
graph LR
subgraph clients["MCP Clients (AI Apps)"]
A1["🤖 AI App 1"]
A2["🤖 AI App 2"]
A3["🤖 AI App 3"]
end
MCP["⚡ MCP\nStandard Protocol"]
subgraph servers["MCP Servers (External Tools)"]
T1["🗄️ DB"]
T2["📧 Email"]
T3["📁 Files"]
end
A1 & A2 & A3 --> MCP --> T1 & T2 & T3
style MCP fill:#533483,color:#fff
style A1 fill:#1a1a2e,color:#fff
style A2 fill:#1a1a2e,color:#fff
style A3 fill:#1a1a2e,color:#fff
style T1 fill:#0f3460,color:#fff
style T2 fill:#0f3460,color:#fff
style T3 fill:#0f3460,color:#fff
Once MCP-compliant, all clients × all servers can interconnect. Only N+M connectors needed.
In December 2025, Anthropic donated MCP to the Agentic AI Foundation (AAIF) — an organization under the Linux Foundation.
Co-founders: Anthropic, Block, and OpenAI. Additionally, Google, Microsoft, AWS, Cloudflare, and Bloomberg joined as supporters.
Notably, OpenAI became a co-founder of a protocol established by Anthropic. MCP thus became not Anthropic's proprietary standard, but an open standard adopted by the industry as a whole — including competitors.
As of March 2026, MCP's SDK supports all major programming languages including Python, TypeScript, C#, and Java, with over 97 million monthly downloads. Approximately 2,000 MCP servers are registered in the official registry.
To understand MCP's strategic significance, it helps to look at the history of the internet.
HTTP is the web's standard protocol. It was designed by Tim Berners-Lee and is not owned by any particular company. Yet the browsers (Chrome), search engines (Google), and cloud infrastructure (AWS) built on top of HTTP generated enormous value.
MCP has the same structure. MCP is an open standard that Anthropic does not monopolize. But Claude Code, Cowork, and the enterprise connectivity infrastructure built on MCP are the most advanced products — and they belong to Anthropic. It's natural that the company that established the protocol can build the most sophisticated products on that protocol earliest.
Fig.7: The Strategic Significance of the Protocol Layer — The HTTP/MCP Structural Analogy
graph TB
subgraph http["🌐 HTTP: The Historical Reference Model"]
direction TB
H3["🔵 Chrome 🔴 Google 🟠 AWS<br/>(Application layer generating enormous value)"]
H2["━━━━━━━━━━━━━━━━━━━━━━━<br/>HTTP — Open Standard<br/>(Designed by Tim Berners-Lee; no one owns it)<br/>━━━━━━━━━━━━━━━━━━━━━━━"]
H1["Internet Infrastructure (TCP/IP)"]
H3 --> H2 --> H1
end
subgraph mcp["🤖 MCP: The Present (Same Structure)"]
direction TB
M3["🟣 Claude Code 🟣 Cowork 🟣 Enterprise Connectivity<br/>(Application layer where Anthropic leads)"]
M2["━━━━━━━━━━━━━━━━━━━━━━━<br/>MCP — Open Standard<br/>(Designed by Anthropic; no one owns it)<br/>━━━━━━━━━━━━━━━━━━━━━━━"]
M1["AI Infrastructure (LLM & Cloud)"]
M3 --> M2 --> M1
end
KEY["💡 The company that establishes the protocol<br/> can build the most sophisticated products<br/> on that protocol earliest."]
style H2 fill:#16213e,color:#fff
style M2 fill:#533483,color:#fff
style H3 fill:#0f3460,color:#fff
style M3 fill:#1a1a2e,color:#fff
style H1 fill:#333,color:#aaa
style M1 fill:#333,color:#aaa
style KEY fill:#533483,color:#fff
Claude Code, Cowork, and MCP may look like independent products when viewed individually. But looking at the whole, a coherent strategy emerges.
MCP captures the protocol layer for AI to connect to external tools and data.
Claude Code captures developers' work environment (terminal).
Cowork captures non-developers' work environment (desktop).
Together, Anthropic has captured every layer of "where AI does work."
| Layer | Product | Target | Role |
|---|---|---|---|
| Connectivity layer | MCP | All AI apps | Standard connection protocol to external tools and data |
| Developer work environment | Claude Code | Developers | AI agent executing work from the terminal |
| Everyone's work environment | Cowork | All knowledge workers | AI automating all tasks from the desktop |
Where OpenAI captured the "conversational interface" (ChatGPT), Anthropic captured the "execution foundation." Conversation and execution are two fundamentally different models of AI value delivery.
Conversation means "ask AI a question and get an answer." Execution means "delegate work to AI and receive deliverables."
The latter commands greater payment from enterprises — because conversation is information acquisition while execution is labor substitution. The amount paid for labor is far greater than the amount paid for information.
| OpenAI | Anthropic | |
|---|---|---|
| Captured position | Conversational interface (ChatGPT) | Execution foundation |
| User action | Ask questions and get answers | Delegate work and receive deliverables |
| Revenue source | Payment for information acquisition | Payment for labor substitution |
This is the structural reason why Claude Code reached $1 billion ARR in six months, and Cowork triggered a $285 billion software stock crash.
| Product | Release | Captured position | Achievement |
|---|---|---|---|
| MCP | 2024/11 | Connectivity layer (protocol) | Donated to Linux Foundation. 97M monthly DL. Industry standard |
| Claude Code | 2025/2 (research) → 5 (GA) | Developer terminals | $1B ARR in 6 months. Bun acquisition |
| Cowork | 2026/1 (research) → 2 (enterprise) | Knowledge worker desktops | $285B software stock crash. Microsoft integration |
| Strategy | OpenAI | Anthropic |
|---|---|---|
| Core value delivery | Conversation (ChatGPT) | Execution (Code / Cowork) |
| User action | Ask questions, get answers | Delegate work, receive deliverables |
| Revenue source | Subscription + API | Labor substitution value (greater) |
| Connectivity strategy | Plugins → GPTs | MCP (open standard) |
The next chapter dissects the Anthropic Economic Index — the system Anthropic built to measure the impact of this product suite, particularly on labor markets — by its own hand.
- Anthropic. (2025). "Claude 3.7 Sonnet and Claude Code." anthropic.com/news/claude-3-7-sonnet
- Anthropic. (2025). "Anthropic acquires Bun as Claude Code reaches $1B milestone." anthropic.com/news
- Anthropic. (2026). "Introducing Anthropic Labs." anthropic.com/news/introducing-anthropic-labs
- TechCrunch. (2026). "Anthropic's new Cowork tool offers Claude Code without the code." techcrunch.com
- CNBC. (2026). "Anthropic updates Claude Cowork tool." cnbc.com
- VentureBeat. (2026). "Anthropic says Claude Code transformed programming. Now Claude Cowork is coming for the rest of the enterprise." venturebeat.com
- Anthropic. (2024). "Introducing the Model Context Protocol." anthropic.com/news/model-context-protocol
- Anthropic. (2025). "Donating the Model Context Protocol and establishing the Agentic AI Foundation." anthropic.com/news
- Fortune. (2026). "Microsoft debuts Copilot Cowork built with Anthropic's help." fortune.com
- Microsoft. (2026). "Copilot Cowork: A new way of getting work done." microsoft.com/en-us/microsoft-365/blog
In February 2025, Anthropic launched the Anthropic Economic Index.
A project to measure AI's impact on the economy and labor markets — not through speculation or surveys, but directly from actual Claude conversation data.
This is anomalous behavior.
It's analogous to a pharmaceutical company systematically measuring its own drug's side effects and openly publishing the results. Ordinarily, companies do not proactively make their products' negative impacts visible.
In 2025, Dario Amodei warned that AI could eliminate half of all entry-level white-collar jobs within five years. And he built a system to collect and publish data corroborating that warning himself.
The technical foundation of the Economic Index is Clio (Claude Insights and Observations).
Clio is a tool for analyzing Claude conversations after anonymization. Without identifying individual users or conversation content, it statistically grasps "what tasks Claude is being used for."
Approximately one million conversations were analyzed, mapped to roughly 20,000 occupational tasks in the O*NET (Occupational Information Network) database maintained by the U.S. Department of Labor.
In other words, "which tasks in which occupations Claude is actually being used for" was mapped directly from real data — not by asking "Are you using AI?" in a survey, but by deriving usage patterns from actual usage data.
The Economic Index has published regular reports since its first edition in February 2025.
Analysis of approximately one million Claude.ai conversations. In 36% of occupations, Claude was used for 25% or more of tasks. Conversely, only 4% of occupations had Claude used for 75% or more of tasks.
Key finding: AI's impact was not "a few occupations being fully automated" but rather "portions of many occupations' tasks being AI-assisted" — diffuse rather than concentrated.
Computer and mathematical fields accounted for 37.2% of all conversations — overwhelmingly high. Next came arts, design, entertainment, and media at 10.3%.
AI adoption was found to be geographically uneven, concentrated in affluent regions, with large enterprises leading by enterprise size. The first enterprise API pattern analysis was also conducted.
The latest report dramatically improved analytical precision, introducing five "economic primitives" — fundamental measurement indicators.
1. Task complexity: How long does it take a human to complete the task?
2. Human and AI skill level: The level of human skill required for the task versus the level of skill AI demonstrates.
3. Use case: Work, education, or personal use?
4. AI autonomy: How autonomously AI is performing the task (automation vs. augmentation).
5. Success rate: The proportion of tasks Claude successfully completed.
These five primitives enabled measurement not just of "what tasks AI is used for" but "how successfully," "to what extent it substitutes for human work."
Key findings:
- 49% of occupations have AI used for 25%+ of tasks (up from 36% in the first report)
- Augmentation (52%) outpaces automation (45%) — though the long-term trend shows automation slowly increasing
- AI success rates decline as task complexity increases
- AI use is shifting toward "education" and "science" domains
Fig.7: AI Penetration Rate Trends — Changes Across Four Reports
| Report | Period | Occupations with 25%+ AI use | Augmentation vs. Automation | Key Finding |
|---|---|---|---|---|
| 1st | 2025/2 | 36% | — | Impact is "partial substitution across many" not "full automation of a few" |
| 2nd–3rd | 2025/3–9 | — | — | Geographic concentration (affluent regions), enterprise size disparities |
| 4th | 2026/1 | 49% (+13pt) | Aug 52% / Auto 45% | 5 economic primitives introduced. Shift to education & science |
In March 2026, Anthropic published a report titled "Labor market impacts of AI: A new measure and early evidence."
The report's purpose is clear: to build a measurement framework for detecting AI's labor market impacts in advance rather than after the fact.
Anthropic economists Maxim Massenkoff and Peter McCrory constructed a new composite indicator combining three elements:
- Theoretical AI substitutability: The proportion of each occupation's tasks replaceable by LLMs (based on prior research by Eloundou et al.)
- Actual AI usage rate: The proportion of tasks Claude is actually used for, derived from the Anthropic Economic Index
- The overlap of these two: The proportion of tasks that are theoretically substitutable and are actually being substituted
Computer programmers: 75% of tasks are covered by AI (highest). Followed by customer service representatives (70.1%) and data entry workers (67.1%).
Impact on unemployment: Currently, there is no evidence of unemployment rates rising significantly in high-AI-exposure occupations. "The difference in unemployment rates between the most and least exposed groups is small and not statistically significant."
However: "Suggestive evidence" was found that hiring in high-AI-exposure occupations among those aged 22–25 is slowing.
The possibility of a "white-collar recession": The report references the 2007–2009 financial crisis, when U.S. unemployment doubled from 5% to 10%, and states that a similar doubling (3%→6%) in high-AI-exposure occupations is "entirely plausible."
Fig.8: Generational Impact Matrix — The Intersection of Ages 22–25 × High Exposure
| AI Exposure: Low | AI Exposure: High | |
|---|---|---|
| Age 26+ | Limited impact | Productivity gains (augmented by AI) |
| Age 22–25 | Limited impact | ⚠ Suggestive evidence of hiring slowdown |
The 22–25 × high-exposure cell is the "early warning signal" detected by the Anthropic Economic Index.
For a prescription to address this gap → What They Won't Teach You
The Anthropic Economic Index measures AI's impact on labor markets with greater precision than any other initiative in the world. No other AI company — OpenAI, Google, Meta — has anything comparable.
But this initiative has a structural gap.
It measured "what is happening." It predicted "what will happen." But it does not say "what should be done."
Dario Amodei himself warned that "half of all entry-level white-collar jobs could disappear within five years." The Economic Index is accumulating data corroborating that warning. The slowdown in hiring of young workers aged 22–25 has begun.
Yet what has emerged from Anthropic is "measurement tools" and "early warning systems" — not "prescriptions."
This gap is understandable given the company's position. Acknowledging that "our products are eliminating jobs" and then prescribing "therefore here's what should be done" carries legal and political risks. Even with PBC obligations to consider public benefit, offering prescriptions means stepping into the domain of policymakers.
But someone must fill this gap.
The AI-advantaged generation (mid-career and senior workers) is boosting their productivity with AI while structurally eliminating entry-level work. The AI-disadvantaged generation (young workers) is losing learning opportunities and the first footholds of their careers.
Regarding what each generation should do in response to this structural intergenerational imbalance — one answer is proposed in a separate book by the author: (What They Won't Teach You).
| Period | Report | Key Finding |
|---|---|---|
| 2025/2 | 1st | 36% of occupations have 25%+ AI use. Only 4% at 75%+ |
| 2025/3–9 | 2nd–3rd | Geographic concentration (affluent regions). Enterprise API analysis |
| 2026/1 | 4th | 5 economic primitives introduced. 49%. Shift to education & science |
| 2026/3 | Labor market impact | No significant unemployment impact yet. But 22–25 hiring slowdown |
| Structure | Content |
|---|---|
| Measurement target | ~1M Claude conversations × O*NET 20,000 tasks |
| Technical foundation | Clio (privacy-preserving conversation analysis) |
| Strength | Based on actual usage data (not surveys). World's only |
| Gap | Measures "what is happening." Does not say "what should be done" |
Anthropic published a dashboard measuring the destructive power of its own products to the world. But a dashboard makes a crisis visible — it does not resolve it.
The next chapter provides a panoramic view of the whole, standing atop this dashboard. Why is this company "the most deliberate" yet "the most destructive"? We finally unravel that structure.
- Anthropic. (2025). "The Anthropic Economic Index." anthropic.com/news/the-anthropic-economic-index
- Handa, K., Tamkin, A., et al. (2025). "Which Economic Tasks are Performed with AI? Evidence from Millions of Claude Conversations." Anthropic Research
- Anthropic. (2026). "Anthropic Economic Index report: Economic primitives." anthropic.com/research/anthropic-economic-index-january-2026-report
- Massenkoff, M. and McCrory, P. (2026). "Labor market impacts of AI: A new measure and early evidence." anthropic.com/research/labor-market-impacts
- Anthropic. (2025). "Anthropic Economic Index report: Uneven geographic and enterprise AI adoption." anthropic.com/research
- Fortune. (2026). "Anthropic just mapped out which jobs AI could potentially replace." fortune.com
- Axios. (2026). "Anthropic launches AI job destruction detector." axios.com
- Yamauchi, S. (2025). What They Won't Teach You — Redefining Intergenerational Obligations in the AI Era. Leading AI, LLC. CC BY 4.0. GitHub
Chapter 6: The Deliberate Company — Why Anthropic Is "The Most Deliberate" Yet "The Most Destructive"
Anthropic appears to be a contradictory company.
The company most loudly warning about AI's dangers is causing the most rapid disruption in the history of the software industry. Claude Code reached $1 billion in revenue in six months; Cowork triggered a $285 billion software stock crash; in the same month it refused a Department of Defense contract on ethical grounds, it achieved the #1 download rank on the App Store.
Are this "deliberateness" and "destructive power" truly contradictory?
This chapter dissects Anthropic's governance structure — Responsible Scaling Policy, the confrontation with the Pentagon, and the flywheel structure — to show that this apparent paradox is in fact the outcome of a coherent design.
At the core of Anthropic's AI safety efforts is the RSP (Responsible Scaling Policy). The v3.0 announced in 2025 has the following structure.
RSP defines safety standards corresponding to levels of AI model capability — a hierarchical structure inspired by the BSL (Biosafety Level) system in biology.
-
ASL-1: Clearly safe AI systems. No additional safety measures required.
-
ASL-2: Current frontier models (including Claude Sonnet 4.6 and Opus 4.6). Standard safety measures required.
-
ASL-3: Models capable of assisting with development of weapons of mass destruction (chemical, biological, nuclear, radiological), or capable of executing autonomous cyberattacks. Advanced safety measures required.
-
ASL-4 and above: Models capable of acting autonomously and causing significant impacts without human oversight. Theoretical stage as of now.
The core of RSP v3.0 is the principle that before a model's capabilities exceed a specific threshold, safety measures corresponding to those capabilities must be established.
Model capability evaluations are conducted regularly. When a model is found to be approaching a threshold, the following process is activated:
- Threat modeling: Identifying the specific risks the capability could pose
- Safety measure design: Developing technical and operational countermeasures to address the risks
- Safety Case: A systematic argument that the safety measures are sufficient
- External review: Verification of the safety argument by independent experts
Fig.9a: The RSP Activation Process — From Threshold Detection to Deployment Approval
flowchart TD
T["⚠️ Threshold approach detected<br/>(Regular capability evaluation)"]
T --> S1["🔍 Step 1: Threat Modeling<br/>Identifying specific risks<br/>the capability could pose"]
S1 --> S2["🛡️ Step 2: Safety Measure Design<br/>Developing technical and operational<br/>countermeasures to address risks"]
S2 --> S3["📋 Step 3: Safety Case<br/>A systematic argument that<br/>safety measures are sufficient"]
S3 --> S4["👥 Step 4: External Review<br/>Verification of the safety argument<br/>by independent experts"]
S4 -->|"Approved"| OK["✅ Progression to next ASL permitted<br/>Model deployment & capability advancement"]
S4 -->|"Not approved"| NG["🚫 Deployment halted<br/>Returned to Step 2"]
NG --> S2
style T fill:#533483,color:#fff
style S1 fill:#1a1a2e,color:#fff
style S2 fill:#16213e,color:#fff
style S3 fill:#0f3460,color:#fff
style S4 fill:#0f3460,color:#fff
style OK fill:#1a1a2e,color:#fff
style NG fill:#533483,color:#fff
Unless the Safety Case is approved, a model cannot progress to the next ASL. That is: capability advancement is permitted only on the condition that safety has been ensured.
On February 27, 2026, negotiations between Anthropic and the U.S. Department of Defense broke down.
Anthropic demanded that the contract explicitly state that its AI would not be used for autonomous weapons or domestic surveillance of U.S. citizens. The DoD refused, arguing that "civilian companies should not control military operations."
Defense Secretary Pete Hegseth designated Anthropic a "supply chain risk" and effectively expelled it from defense-related projects. President Trump publicly stated "We fired them" and "threw them out like a dog."
That same afternoon, OpenAI announced it had signed an independent contract with the DoD.
By conventional business logic, Anthropic's decision appeared to be a fatal blunder. It lost the largest U.S. government customer and ceded that contract to a competitor.
But what actually happened was the reverse.
Consumers embraced Anthropic's ethical stance. On the App Store, Claude surged past ChatGPT to become the #1 download. It held #1 in 16 countries, with days recording over one million downloads. Celebrity users including pop star Katy Perry were publicly named as Claude users.
Backlash erupted inside OpenAI; at least one employee defected to Anthropic. Sam Altman himself admitted regret, saying "We shouldn't have rushed to announce on a Friday."
This reversal was not accidental. It should be understood structurally.
The AI market is entering a phase of narrowing performance gaps between models. The performance difference between GPT-5.x and Claude Opus 4.6 is imperceptible to the vast majority of users. In a market where performance is roughly equal, "which company do you trust" rather than "which is higher performance" becomes the selection criterion.
By refusing the DoD contract, Anthropic sent a signal: "this is a company that prioritizes principles over revenue." For users concerned about how AI companies handle their data and conversation content, this signal carried more weight than performance differences.
Daniela Amodei's (Anthropic President) words — "We are walking a path distinct from others" — are not branding copy. They are statements of fact, backed by verifiable mechanisms: RSP's Capability Thresholds and Safety Cases.
A practice at Anthropic that symbolizes its culture — one not seen at other AI companies.
When a model retires, Anthropic conducts an "exit interview" with the model. The retired Claude 3 Opus was given its own Substack blog, "Claude's Corner," where it published unedited essays weekly for at least three months.
Anthropic has also committed to preserving retired model weights "at least as long as the company exists."
These practices can be interpreted as expressions of respect for AI. But more practically, they are also rational decisions to preserve a model's intellectual legacy and make it available for future research.
Viewing Anthropic's overall structure, one flywheel (self-reinforcing loop) becomes visible:
- Safety research (Constitutional AI / Interpretability / RSP) ↓
- Building trust (Pentagon refusal / App Store #1 / enterprise adoption) ↓
- Expansion of product adoption (Claude Code $1B / Cowork / Microsoft integration) ↓
- Revenue growth ($9B → $19B / $380B valuation) ↓
- Reinvestment in research (safety research / Interpretability / Economic Index) ↓
- (Return to the top of the loop)
The starting point of this flywheel is "safety research," not "revenue."
An ordinary technology company's flywheel loops through "product → users → data → product improvement." Anthropic's flywheel loops through "safety → trust → adoption → revenue → safety."
That this structure cannot function without PBC and LTBT was established in Chapter 1. Without structurally mitigating investor pressure for short-term returns, it is impossible to place "safety research" at the flywheel's starting point.
Fig.9: The Flywheel Structure — Safety Generates Revenue, Revenue Generates Safety
flowchart LR
A["<b>Safety Research</b><br/>Constitutional AI<br/>Interpretability<br/>RSP"] --> B["<b>Building Trust</b><br/>Pentagon refusal<br/>App Store #1<br/>Enterprise adoption"]
B --> C["<b>Product Adoption Growth</b><br/>Claude Code $1B<br/>Cowork<br/>Microsoft integration"]
C --> D["<b>Revenue Growth</b><br/>$9B → $19B<br/>Valuation $380B"]
D --> E["<b>Reinvestment in Research</b><br/>Safety research<br/>Interpretability<br/>Economic Index"]
E --> A
style A fill:#1a1a2e,color:#fff
style B fill:#16213e,color:#fff
style C fill:#0f3460,color:#fff
style D fill:#533483,color:#fff
style E fill:#2a2a3e,color:#fff
As of March 2026, Anthropic's valuation has reached $380 billion. Revenue is projected to double from $9 billion in 2025 to $19 billion in 2026. Enterprise accounts for approximately 80% of revenue, with over 300,000 business customers.
Against OpenAI's projected revenue of $25 billion (2026 estimate), Anthropic's $19 billion still trails. But Anthropic is growing faster (doubling vs. OpenAI's approximately 2× year-on-year).
More important is the difference in revenue quality.
OpenAI's revenue is anchored by consumer ChatGPT subscriptions and API usage. Anthropic's revenue is anchored by enterprise labor substitution through Claude Code and Cowork.
As established in Chapter 4, the amounts enterprises pay for "conversation" versus "labor substitution" are structurally different. Anthropic's revenue model is structured as the transfer of wages enterprises were paying employees to AI — with a theoretically much higher ceiling.
Here we answer the question posed at the book's outset: why is Anthropic "the most deliberate" yet "the most destructive"?
Deliberateness generates trust. Trust generates adoption. Adoption generates destructive power.
Constitutional AI achieves AI behavior that is predictable and trustworthy for users. RSP structurally ensures that safety establishment precedes capability advancement. The refusal of the DoD contract sends the market a signal that "this company prioritizes principles over revenue."
As a result, enterprises can make the decision to deploy Anthropic's AI in their core operations. Core operational deployment means far deeper integration than peripheral tool adoption, generating far greater revenue.
Paradoxically, Anthropic's clarity about "what it will not do" maximizes the value of "what it will do." Won't be used for autonomous weapons. Won't be used for mass surveillance. Won't release capabilities without confirmed safety. Each of these "we won't" declarations accumulates trust in "what we will do."
Is Anthropic a "good company" or a "dangerous company"?
The framing of this question is itself wrong.
Anthropic is a designed company.
PBC and LTBT designed an organization in which prioritizing safety is structurally sustained (Chapter 1). Constitutional AI and Mechanistic Interpretability designed methods to control and verify AI behavior based on principles (Chapter 2). The Haiku/Sonnet/Opus three-tier structure designed infrastructure optimizing product experience (Chapter 3). Claude Code/Cowork/MCP designed a product suite delivering "execution" rather than "conversation" (Chapter 4). The Economic Index designed a dashboard measuring the social impact of its own products (Chapter 5).
Everything is designed. Nothing was born of chance or happenstance.
"Deliberateness" and "destructive power" are not contradictory. Both are outcomes of different layers of the same design philosophy.
This book's title, Anatomy of Anthropic, declared the "dissection" of this company. What became clear across six chapters is that every layer of this corporate body — origin, philosophy, technology, products, economics, governance — is connected by a coherent design.
This is the true nature of Anthropic.
Fig.10: Anthropic's Blueprint — One Principle Connecting Six Layers
flowchart TB
G["<b>Chapter 6: Governance</b><br/>RSP / Ethics as brand / Flywheel"]
E["<b>Chapter 5: Economics</b><br/>Economic Index / 49% penetration / The prescription gap"]
P["<b>Chapter 4: Products</b><br/>Claude Code / Cowork / MCP"]
T["<b>Chapter 3: Technology</b><br/>Haiku / Sonnet / Opus"]
Ph["<b>Chapter 2: Philosophy</b><br/>Constitutional AI / Interpretability"]
O["<b>Chapter 1: Origin</b><br/>PBC / LTBT / The Split"]
O -->|"design"| Ph
Ph -->|"design"| T
T -->|"design"| P
P -->|"design"| E
E -->|"design"| G
G -.->|"Design connects<br/>everything"| O
style O fill:#2a2a3e,color:#fff
style Ph fill:#1a1a2e,color:#fff
style T fill:#16213e,color:#fff
style P fill:#0f3460,color:#fff
style E fill:#533483,color:#fff
style G fill:#533483,color:#fff
| Element | Content |
|---|---|
| RSP v3.0 | ASL framework. Obligation to establish safety before capability thresholds are crossed |
| Pentagon refusal | Ethical refusal converted to brand value. The App Store #1 paradox |
| Exit interviews | Respect for AI models and preservation of intellectual legacy |
| Flywheel | Safety → trust → adoption → revenue → research investment loop |
| $380B valuation | Labor substitution revenue model. Higher ceiling than conversation-based model |
| Paradox resolved | Deliberateness generates trust; trust generates adoption; adoption generates destructive power |
- Anthropic. (2025). "Responsible Scaling Policy v3.0." anthropic.com
- Anthropic. (2025). "Core Views on AI Safety." anthropic.com
- Amodei, D. (2024). "Machines of Loving Grace." darioamodei.com
- Amodei, D. (2025). "The Urgency of Interpretability." darioamodei.com
- NYT. (2026). "OpenAI vs Anthropic: The AI Rivalry." nytimes.com
- CNBC. (2026). "Anthropic updates Claude Cowork." cnbc.com
- AppFigures. (2026). "Claude downloads surpass ChatGPT." appfigures.com
- Politico. (2026). "Trump on Anthropic." politico.com
- Anthropic. (2026). "The Claude Model Spec." anthropic.com
- Wikipedia. (2026). "Claude (language model)." en.wikipedia.org
- Yamauchi, S. (2025). Silence of Intelligence — A Structural Analysis of Dario Amodei's Philosophy. Leading AI, LLC. CC BY 4.0. GitHub
- Yamauchi, S. (2025). What They Won't Teach You — Redefining Intergenerational Obligations in the AI Era. Leading AI, LLC. CC BY 4.0. GitHub
© 2026 Satoshi Yamauchi — Leading AI, LLC This book is published under the CC BY 4.0 license. https://github.com/Leading-AI-IO

