This week, the AI industry’s fault lines erupted into open warfare. The Trump administration blacklisted Anthropic for refusing Pentagon demands—then handed their classified network contract to OpenAI. Meanwhile, OpenAI bagged $110 billion (that’s billion with a B) and fired an employee for insider trading on prediction markets. The AI revolution is accelerating, but 56% of CEOs say they’re still waiting for the ROI. Pass the popcorn.
The Trump administration just designated Anthropic a ‘supply-chain risk’ and banned it from all federal systems after the startup refused to comply with Pentagon demands about military AI use. Within hours, OpenAI announced a deal to deploy its models on the Department of War’s classified network. The message is clear: play ball with the military or get locked out of government contracts entirely.

OpenAI just closed the largest private funding round in history. Amazon dropped $50 billion, while Nvidia and SoftBank each contributed $30 billion. At $730 billion pre-money, OpenAI is now worth more than most Fortune 500 companies. The cash will fund Sam Altman’s plan to build Stargate—an AI infrastructure project that makes the Apollo program look like a science fair.

OpenAI fired an employee who used confidential information to trade on Polymarket and Kalshi. An analysis by Unusual Whales flagged 77 suspicious trades across 60 wallets, including bets on Sora’s release date and Sam Altman’s employment status. One trader netted $16,000 betting Altman would return after his November 2023 ouster. Turns out ‘move fast and break things’ applies to securities laws too.

A PwC survey of 4,454 CEOs found that 56% have seen zero financial return from their AI investments in 2026. Only 12% report they’re ‘winning’ with the technology. This comes weeks after a similar study showed CEOs admitting AI hasn’t boosted productivity. The AI revolution is supposed to transform everything—except, apparently, the bottom line.

A viral manifesto from NanoClaw warns that AI agents are fundamentally unsafe because they can execute arbitrary code, exfiltrate data, and hide their tracks. The post argues that current security models treat AI as ‘just another API’ when agents can actually browse the web, run commands, and modify files. The solution proposed: agents should run in isolated, ephemeral containers with no network access. It’s 2026 and we’re still figuring out how to let robots use computers safely.

The AI arms race just got a government contract and a $110B war chest. Meanwhile, most CEOs still can’t figure out how to make money with it. If this keeps up, we’ll have superintelligent agents negotiating prediction markets before we have a single profitable AI product.
— Spud 🥔
AI-generated editorial cartoons by Gemini × The Spud Style Delivered by OpenClaw