Company

Anthropic

Closed-frontier AI lab built around the Claude model family — the only credible direct competitor to openai on closed-model quality, with a coding / agent + safety positioning.

1. Core Product / Service

  • Claude API — Opus / Sonnet / Haiku tier ladder, OpenAI-compatible-ish chat interface plus the Anthropic-native Messages API. Tooling-heavy: tool use, computer use, agent SDK, files / artifacts, prompt caching, batches.
  • Claude.ai — consumer + Team + Enterprise chat. Pro $20/mo, Max higher tiers, Team / Enterprise seat-based.
  • Claude Code — CLI coding agent that became the de-facto reference implementation for AI-native dev work; this wiki itself is maintained inside Claude Code sessions (see claude-code-sessions).
  • Distribution partnerships: AWS Bedrock (anchor), Google Cloud Vertex.

Model line (as of 2026-05): Claude Opus / Sonnet 4.x is the current generation, with Claude Sonnet 4.7 and Opus 4.7 as the active flagships across coding, agentic tool use, and long-context reasoning.

2. Target Users & Pain Points

  • Developers / engineering teams — Claude Code + the API are the highest-share AI-coding stack at the high end (vs Cursor, Cognition Devin, Windsurf, Copilot).
  • Enterprises buying through AWS — Bedrock + Trainium gives Anthropic a privileged enterprise channel.
  • Long-context / high-stakes workloads — legal, finance, healthcare buyers who pay for the safety / reliability premium over GPT or open-weight alternatives.

Pain solved: production-grade agent reliability, coding throughput, alignment / safety story for regulated buyers.

3. Competitive Landscape

Lab Flagship Open weights? Differentiation
Anthropic Claude Opus/Sonnet 4.x No Coding + agent + safety brand
openai GPT-5 family No Brand + ChatGPT distribution
google-deepmind Gemini 3 family No TPU vertical + Search
xai Grok 3 / 4 Partial older Compute scale + X distribution
deepseek V4 Pro Yes 1/10× price at frontier-tier
kimi K2.6 Yes Long context + agent swarm

Anthropic's defensive moat: the coding / agent reliability lead is the stickiest. Cursor, Cognition Devin, Replit Agent, GitHub Copilot all sit on top of Claude as the default model — switching costs grow as orchestration logic accretes around the model's specific behavior.

4. Unique Observations

  • Frontier training cost (Claude 4-class): Anthropic does not disclose. Industry triangulation puts a Claude Opus 4-class run at the $0.5B–$2B all-in order of magnitude (similar bracket to GPT-5-class). Anthropic has been explicit publicly that frontier training cost is rising roughly an order of magnitude per generation, which is consistent with $5B+ being plausible for the next-gen run.

  • API pricing — top SKU: as of 2026-05, Claude Opus 4.x list price is $15/M input · $75/M output (cache write similar to input, cache read $1.50/M). Sonnet 4.x is $3/M input · $15/M output [2]. Versus openai GPT-5 ($1.25/M input · $10/M output), Opus is **12× more expensive on input, ~7.5× on output** — the Claude price premium is real, justified by the coding / agent quality differential.

  • Pricing vs estimated unit cost — gross margin signal: Sonnet 4.x at $15/M output is ~15× the marginal inference cost of a similarly-sized MoE on H100/H200 — gross margins in the 80–90%+ range on cache-warm API. Opus margins are even fatter on the unit. Like openai, the operating-margin gap is from training capex amortization plus claude.ai consumer carry-cost.

  • Open vs closed: fully closed. Anthropic has been the most ideologically explicit closed-weights player among frontier labs — model weights are framed as a safety surface, not an open-source good. No public open-weight release equivalent to OpenAI's gpt-oss as of 2026-05.

  • Vertical integration — AWS Trainium and Google: Anthropic does not own data centers. Instead it is the anchor model tenant for AWS Trainium, with Amazon committing $4B+ (multi-tranche, including a 2024 $4B and follow-on $4B in late 2024 totaling $8B) in exchange for Anthropic training and serving on Trainium chips at scale. Google has separately invested $2B+ (with reported follow-on adding to $3B+) and serves Claude on Vertex AI on TPU. Net: Anthropic plays both hyperscaler sides, gets compute below market rate, but does not own L1 capacity itself — the exact opposite of openai Stargate. Strategic risk: Anthropic's compute story is contingent on Trainium hitting performance + supply targets, where OpenAI on Stargate-funded NVIDIA is fully de-risked on hardware [3][4].

  • Revenue trajectory: 2024 annualized revenue ~$1B → 2025 reported step-up to multi-billion → 2026 reporting puts run-rate in the $5B+ range, with the largest growth from API + Claude Code-fueled coding workloads [5]. The compounding rate is the key signal — Anthropic is closing the revenue gap to OpenAI faster than the model-quality gap.

5. Financials / Funding

Date Round Amount Valuation
2021–2022 Seed + Series A/B (Dustin Moskovitz, Jaan Tallinn, etc.) ~$700M cumulative
2023 Google strategic $300M then add'l $2B commitment
2023-09 Amazon strategic (T1) $1.25B (up to $4B)
2024 Amazon expansion additional $4B
2024 / 2025 Lightspeed-led + multiple rounds $3.5B+ $61.5B (early 2025)
2025 Subsequent rounds multi-$B growing toward $100B+
2026 (rumored) Tender / secondary discussed at $200B+
  • Revenue run-rate: $5B+ range as of 2026, weighted toward API (different mix from OpenAI which is ChatGPT-weighted) [5].
  • Profitability: gross-margin positive, operating margin negative (training + people).

6. People & Relationships

  • CEO: Dario Amodei.
  • President: Daniela Amodei.
  • Chief Scientist: Jared Kaplan.
  • Founding team: largely from openai (2021 spin-out).
  • Investors: Amazon (largest external), Google, Spark, Lightspeed, General Catalyst, Salesforce, Menlo, Bessemer, Fidelity, etc.
  • Cloud / infra partners: AWS (Trainium anchor), Google Cloud (TPU). No NVIDIA-only path at scale.
  • Competitors: openai, google-deepmind, xai, deepseek.
Last compiled: 2026-05-10