Anthropic's 72 Hours of Identity Crisis

The end point of an identity crisis is often the disappearance of that identity.

By Ada, Deep Tide TechFlow

Tuesday, February 24. Washington, D.C., Pentagon.

Anthropic CEO Dario Amodei sat across from Defense Secretary Pete Hegseth. According to multiple media outlets including NPR and CNN, the meeting was “polite,” but the content was anything but.

Hegseth delivered a final ultimatum: by 5:01 p.m. Friday, lift restrictions on Claude’s military use, allowing the Pentagon to deploy it for “all lawful purposes,” including autonomous weapon targeting and domestic mass surveillance.

Otherwise, cancel the $200 million contract. Invoke the Defense Production Act for compulsory requisition. List Anthropic as a “supply chain risk,” effectively blacklisting it alongside hostile entities from Russia and China.

On the same day, Anthropic quietly released version 3.0 of its “Responsible Expansion Policy” (RSP 3.0), removing a core promise since the company’s founding: that they would not train more powerful models unless safety measures could be guaranteed.

Also on that day, Elon Musk posted on X: “Anthropic is mass stealing training data, that’s a fact.” Community notes on X added reports that Anthropic paid $1.5 billion in settlement for training Claude using pirated books.

Within 72 hours, this AI company, claiming to have a “soul,” played three roles simultaneously: safety martyr, intellectual property thief, and Pentagon traitor.

Which is the real one?

Maybe all of them.

Pentagon’s “Either Obey or Get Out”

The story’s first layer is simple.

Anthropic is the first AI company granted classified access by the U.S. Department of Defense. The contract, awarded last summer, has a cap of $200 million. Subsequently, OpenAI, Google, and xAI also secured similar-sized contracts.

According to Al Jazeera, Claude was used in a U.S. military operation in January this year, reportedly involving the kidnapping of Venezuelan President Maduro.

But Anthropic drew two red lines: no support for fully autonomous weapon targeting, and no support for large-scale surveillance of U.S. citizens. Anthropic believes AI is not reliable enough to control weapons, and currently there are no laws regulating AI in mass surveillance.

The Pentagon isn’t buying it.

White House AI advisor David Sacks publicly accused Anthropic on X last October of “weaponizing fear to capture regulation.”

Competitors have already capitulated. OpenAI, Google, and xAI all agreed to let the military use their AI for “all lawful scenarios.” Musk’s Grok just got approval to access classified systems this week.

Anthropic is the last standing.

As of this writing, Anthropic stated in its latest release that they have no intention to back down. But the Friday 5:01 deadline is looming.

An anonymous former DOJ and DOD liaison told CNN: “How can you simultaneously declare a company a ‘supply chain risk’ and force it to work for your military?”

Good question, but it’s not within the Pentagon’s consideration. They care about whether Anthropic will compromise—by capitulating—or be discarded by Washington.

“Distillation Attack”: A Face-Slapping Accusation

On February 23, Anthropic published a strongly worded blog accusing three Chinese AI companies of conducting an “industrial-scale distillation attack” on Claude.

The accused are DeepSeek, Moonshot AI, and MiniMax.

Anthropic claims they used over 24,000 fake accounts to initiate more than 16 million interactions with Claude, targeting its core reasoning, tool invocation, and programming capabilities.

They categorize this as a national security threat, asserting that models distilled this way are “unlikely to retain safety guardrails” and could be exploited by authoritarian governments for cyberattacks, disinformation, and mass surveillance.

The narrative is perfect, and the timing is impeccable.

It coincides precisely with the period after the Trump administration relaxed U.S. chip export controls to China, and when Anthropic was seeking ammunition for lobbying against chip export restrictions.

But Musk shot back: “Anthropic is mass stealing training data and paid billions in settlement. That’s a fact.”

Tory Green, co-founder of AI infrastructure company IO.Net, said: “You train your model on the entire internet’s data, then others learn from your public API—that’s called ‘distillation attack’?”

Anthropic calls it “attack,” but in the AI industry, this is common practice. OpenAI used it to compress GPT-4, Google to optimize Gemini, and even Anthropic itself does it. The only difference is, this time, they are the ones being distilled.

According to Erik Cambria, AI professor at Nanyang Technological University in Singapore, “The line between legal use and malicious exploitation is often blurry.”

Ironically, Anthropic paid $1.5 billion in settlement for training Claude on pirated books, yet now accuses others of using its public API to learn from it. This isn’t double standards; it’s triple standards.

Anthropic wanted to play the victim but ended up as the accused.

The Dismantling of Safety Commitments: RSP 3.0

On the same day as its confrontation with the Pentagon and public spat with Silicon Valley, Anthropic released version 3.0 of its Responsible Expansion Policy.

Anthropic Chief Scientist Jared Kaplan told the media: “We believe stopping AI training doesn’t help anyone. In the context of rapid AI development, making unilateral promises… while competitors accelerate, is pointless.”

In other words, if others don’t play fair, neither will we.

Core to RSP 1.0 and 2.0 was a strict promise: if a model’s capabilities exceed safety measures, training would be paused. This promise earned Anthropic a unique reputation in the AI safety community.

But 3.0 drops that.

Instead, it introduces a more “flexible” framework, separating safety measures that Anthropic can implement from safety recommendations requiring industry-wide collaboration. They plan to release risk reports every 3-6 months, reviewed by external experts.

Sounds responsible?

Chris Painter, an independent reviewer from nonprofit METR, said after reviewing early drafts: “This indicates that Anthropic believes it needs to enter a ‘triage mode’ because its methods for assessing and mitigating risks can’t keep pace with capability growth. It more likely reflects society’s unpreparedness for AI’s potentially catastrophic risks.”

According to TIME, it took nearly a year of internal debate for Anthropic to rewrite this policy, approved unanimously by CEO Amodei and the board. Officially, the original policy aimed to foster industry consensus, but the industry never caught up. The Trump administration’s laissez-faire attitude toward AI development, even attempts to repeal state regulations, left federal AI legislation in limbo. While a global governance framework in 2023 seemed possible, three years later, that door is closed.

A long-time AI governance researcher said more bluntly: “RSP is Anthropic’s most valuable brand asset. Removing the pause on training promises is like an organic food company secretly tearing ‘organic’ off its packaging and claiming its testing is now more transparent.”

Identity Rift Under a $380 Billion Valuation

In early February, Anthropic closed a $30 billion funding round at a $380 billion valuation, with Amazon as a cornerstone investor. Since its founding, it has achieved an annualized revenue of $14 billion, growing over tenfold each year for the past three years.

Meanwhile, the Pentagon threatened to blacklist it. Musk publicly accused it of data theft. Its core safety commitments were removed. After CTO Mrinank Sharma resigned, he posted on X: “The world is in danger.”

Contradiction?

Perhaps contradiction is in Anthropic’s DNA.

Founded by former OpenAI executives worried about OpenAI moving too fast on safety, they built their own company to develop more powerful models faster, while telling the world how dangerous these models are.

Their business model can be summarized as: we fear AI more than anyone, so you should pay us to build it.

This narrative worked perfectly in 2023-2024. AI safety was a hot topic in Washington, and Anthropic was the leading lobbyist.

By 2026, the tide turned.

“Woke AI” became a slur, state-level AI regulation bills were blocked by the White House, and the California SB 53 supported by Anthropic was signed into law but remained largely ineffective at the federal level.

Anthropic’s safety brand is slipping from a “differentiation advantage” to a “political liability.”

It is engaged in a complex balancing act: maintaining enough “safety” to uphold its brand, while staying flexible enough not to be abandoned by markets and governments. The problem is, both tolerances are shrinking.

How Much Is the Safety Narrative Worth?

Stacking these three issues together makes the picture clear.

Accusing Chinese companies of distilling Claude is to reinforce the narrative for chip export controls. Removing safety pause commitments to stay competitive. Refusing Pentagon’s autonomous weapon demands to preserve the last moral high ground.

Each step makes sense individually, but they conflict with each other.

You can’t claim Chinese companies’ “distillation” endangers national security and simultaneously remove your own safety promises to prevent model runaway. If models are truly so dangerous, you should be more cautious, not more aggressive.

Unless you’re Anthropic.

In the AI industry, identity isn’t defined by your statements but by your balance sheet. Anthropic’s “safety” narrative is essentially a brand premium.

In the early AI arms race, this premium was valuable. Investors paid higher valuations for “responsible AI,” governments greenlit “trustworthy AI,” and customers paid more for “safer AI.”

But by 2026, that premium is evaporating.

What Anthropic faces now isn’t a question of “whether to compromise,” but “who to compromise with first.” Compromise with the Pentagon damages the brand. Compromise with competitors nullifies safety promises. Compromise with investors means losing on both ends.

By 5:01 p.m. Friday, Anthropic will deliver its answer.

But whatever it is, one thing is certain: the Anthropic that once thrived on “we’re different from OpenAI” is becoming just like everyone else.

The end of an identity crisis is often the disappearance of that identity.

XAI4.34%
GROK-7.23%
IO0.62%
View Original
This page may contain third-party content, which is provided for information purposes only (not representations/warranties) and should not be considered as an endorsement of its views by Gate, nor as financial or professional advice. See Disclaimer for details.
  • Reward
  • Comment
  • Repost
  • Share
Comment
0/400
No comments
  • Pin

Trade Crypto Anywhere Anytime
qrCode
Scan to download Gate App
Community
English
  • 简体中文
  • English
  • Tiếng Việt
  • 繁體中文
  • Español
  • Русский
  • Français (Afrique)
  • Português (Portugal)
  • Bahasa Indonesia
  • 日本語
  • بالعربية
  • Українська
  • Português (Brasil)