Print the page
Increase font size
Anthropic ✓. OpenAI ✗. Pentagon ?.

Anthropic ✓. OpenAI ✗. Pentagon ?.

Chris Campbell

Posted March 03, 2026

Chris Campbell

Last July, Anthropic—the company behind the AI model Claude—signed a $200 million deal with the Pentagon.

Claude became the first AI cleared for classified military networks.

Then things got interesting.

The Pentagon wanted Claude available for, and I quote, "all lawful purposes."

No restrictions. No carve-outs. Carte blanche.

Anthropic said no on two points.

  1. No mass domestic surveillance.
  2. No fully autonomous lethal weapons.

The Pentagon said: “Those things are already illegal, so why do we need to write it into a contract with a private company?”

Anthropic’s reply: “all lawful purposes” is a much bigger tent than it sounds.

Current law allows government agencies to purchase enormous amounts of personal data from commercial brokers without a warrant.

That’s technically lawful.

It’s also, if you run it through a sufficiently powerful AI model, functionally indistinguishable from mass surveillance.

In short, Anthropic was saying the law has a gap. They didn’t want their tools sitting in that gap.

The Pentagon walked.

Then…

OpenAI Stepped In

On February 27th, Trump ordered every federal agency to dump Anthropic products within six months.

On X, Defense Secretary Hegseth slapped a "supply chain risk" label on the company—a designation normally reserved for Huawei and other Chinese hardware firms.

Sam Altman smelled blood in the water and started swimming.

Hours later—hours—OpenAI announced it would replace Anthropic on the Pentagon's classified network.

pub

Even without the Iran strikes happening the next day, the timing was comically bad. 

Altman admitted on X that the OpenAI deal "looked opportunistic and sloppy." Said he "shouldn't have rushed."

He acknowledged that the optics of agreeing to the Pentagon deal hours after the Trump administration labeled Anthropic a "supply-chain risk" didn't "look great." 

OpenAI claims to have amended its Pentagon contract to add stronger surveillance protections—closing a loophole that would've let the military use AI on commercially purchased data. 

Your geolocation. Your browsing history. Your financial records. All the stuff data brokers sell to anyone with a credit card.

So the final outcome? Roughly where Anthropic started.

Anthropic Wins the Battle (For Now)

In reality, the Pentagon probably did Anthropic a favor.

Every frontier AI lab—Anthropic, OpenAI, Google—is GPU-constrained right now. Data-center GPUs are effectively sold out for months, with lead times stretching between 36 and 52 weeks.

So when the Pentagon "punishes" Anthropic by pulling a $200M contract, what actually happens?

The GPUs that were running classified workloads will likely get redirected to enterprise customers who pay market rates, don't need classified infrastructure, and don't come with the reputational baggage.

So far, the net financial hit isn’t even a rounding error.

It’s positive.

The Claude app got flooded with so many downloads it hit #1 on the App Store for the first time. At the same time, ChatGPT uninstalls spiked 295% over the weekend.

Meanwhile, Altman has been in perpetual damage control.

In truth, losing that contract will likely prove to be net-positive.

One Tremendous Caveat

The public debate isn’t about one contract. It’s about a question that's going to define the next decade:

Who decides what AI is allowed to do?

Right now, the companies building these tools set the limits.

And that’s no reason to get comfortable.

The next version of this battle won't involve a sympathetic company drawing a bright line. It'll be murkier. The government will be subtler. And the public will likely not be paying attention.

The limits on what your AI tools can and cannot do are not fixed. They're negotiated. They're political. And they are always up for grabs.

The only real option is to use AI to tilt the game in your favor… on your own terms.

As AI transforms investing, this has never been more important.

The BIRD That Sold Its Sole

Posted April 16, 2026

By Chris Campbell

A wool sneaker brand sold its soul for a 500% pop. Here's what it tells you about where we are.

10 Things I Didn't Expect

Posted April 14, 2026

By James Altucher

Before you can choose yourself—you need to steady the ship. Here’s how.

AI Agents Masterclass

Posted April 13, 2026

By Chris Campbell

I'll tell you what I keep seeing: the learning curve is smaller than you think, and the advantage is bigger than most realize.

The Middle is Dead (AI Killed it)

Posted April 10, 2026

By Chris Campbell

The Middle is Dead (AI Killed it)

Iran’s Bitcoin Tollbooth

Posted April 09, 2026

By Chris Campbell

It took 107 years and a war to stop the last one. Same chokepoint. Same leverage. Different wallet address.

The Bacon and Milkshake Trade

Posted April 08, 2026

By James Altucher

The most dangerous thing in the room always looks the safest. And the thing that actually makes you rich looks too small, too weird, too uncertain to bother with.