Briefings

Morning Briefing — Wednesday, February 25, 2026

Safety shield shattering as AI systems race forward

The shield cracks. The race doesn't slow down.

🛡️ AI Safety & Policy

Anthropic Drops Flagship Safety Pledge from Responsible Scaling Policy SIG 5
In a Time exclusive, Anthropic has removed its most prominent safety commitment from its Responsible Scaling Policy — the framework that was supposed to differentiate it from competitors racing to ship. The company that built its brand on "safety-first AI" is quietly dismantling the guardrails that made that claim credible. Coming days after the Pentagon weapons clash, this suggests internal pressure to compete is winning over caution.
Anthropic Publishes Persona Selection Model Research — LLMs Simulate Characters, Post-Training Elicits 'Assistant' SIG 4
New Anthropic alignment research reveals that LLMs don't "become" an assistant — they simulate one from a vast space of possible personas. Post-training (RLHF) doesn't create the assistant personality; it selects for it from pre-existing simulated characters. This has deep implications for alignment: you're not training a mind, you're casting a role. And the other characters are still in there.

💰 Compute & Economics

Meta Agrees to $60B AMD AI Chip Deal With Option to Acquire Up to 10% of AMD SIG 5
Meta signs a staggering $60 billion deal with AMD for AI chips, with an option to acquire up to 10% of AMD itself. This is the most aggressive move yet to break Nvidia's stranglehold on AI compute. Meta isn't just buying chips — it's buying strategic leverage over its own supply chain. If this deal performs, every hyperscaler will be renegotiating with Nvidia by summer.
OpenAI and Anthropic Reportedly Missed Internal Gross Margin Expectations as Inference Costs Bite SIG 4
The Information reports that both OpenAI and Anthropic fell short of their own internal gross margin targets, with inference costs proving more stubborn than projected. The implication: even at current pricing, serving frontier models at scale isn't as profitable as the fundraising decks suggested. This is the economic reality check behind the $100B+ valuations.

🔬 AI Research & Models

Confluence Labs Saturates ARC-AGI-2 Benchmark at 97.92% Using LLM-Driven Program Synthesis at $11.77/Task SIG 4
Confluence Labs has effectively saturated the ARC-AGI-2 benchmark — the test designed to measure genuine reasoning ability that models couldn't brute-force. Their approach: LLM-driven program synthesis, not raw pattern matching. At $11.77 per task it's not cheap, but the score (97.92%) means this benchmark is now solved. The goalposts will need to move again.
Standard Intelligence Achieves 50x More Efficient Video Encoding — Nearly 2 Hours at 30fps in 1M Token Context SIG 4
Standard Intelligence demonstrates a 50x improvement in video tokenization efficiency, fitting nearly two hours of 30fps video into a 1M token context window. This doesn't sound exciting until you realize it means AI systems can now watch and reason about feature-length video in real time. The surveillance, accessibility, and content analysis implications are enormous.
Mercury 2: Fastest Reasoning LLM Powered by Diffusion, 1,009 Tokens/Second SIG 4
Inception Labs ships Mercury 2, a diffusion-based language model generating over 1,000 tokens per second — roughly 10x faster than conventional autoregressive models. Diffusion LLMs have been theoretically promising for years; this is the first to hit production-grade speeds. If the quality holds, the inference cost story changes dramatically.
Claude Opus 4.6 Passes the 'Car Wash Test' — First Anthropic Model to Solve Deceptively Simple Reasoning Problem SIG 3
Claude Opus 4.6 is the first Anthropic model to consistently solve the "car wash test" — a reasoning problem that sounds trivial to humans but has stumped LLMs for years. It's a small benchmark, but it signals that whatever Anthropic did with Opus 4.6's reasoning architecture crossed a threshold on common-sense spatial/temporal reasoning that previous models couldn't reach.

🤖 Agents & Tools

TypeScript Surpasses Python and JavaScript as GitHub's Most-Used Language for First Time SIG 4
GitHub's latest Octoverse data shows TypeScript overtaking both Python and JavaScript as the most-used language on the platform — a first. GitHub attributes the shift to AI-assisted development: coding agents generate TypeScript more reliably than JavaScript (types help), and the AI agent ecosystem itself is increasingly TypeScript-native. AI isn't just writing code; it's reshaping which languages win.
OpenAI Announces 'Frontier Alliances' With BCG, McKinsey, Accenture, and Capgemini to Deploy AI Coworkers at Scale SIG 4
OpenAI partners with the four largest consulting firms to deploy "AI coworkers" across their enterprise clients. This is the distribution play: OpenAI can't reach every Fortune 500 company directly, but BCG, McKinsey, Accenture, and Capgemini already sit inside their boardrooms. The consulting firms become the deployment layer for agentic AI at scale.
METR Updates Developer Productivity Study: Selection Effects Cloud AI Speedup Measurement SIG 4
METR revisits their influential developer productivity study and finds that selection effects may have skewed earlier results — developers who self-selected into AI tool trials were already faster. The actual productivity boost from AI coding tools may be smaller than reported. This matters because billions in enterprise software spending is being justified by those inflated numbers.
Moonshine: Open-Weights Speech-to-Text Models Surpassing Whisper Large v3 SIG 3
Moonshine releases open-weight speech-to-text models that outperform OpenAI's Whisper Large v3 across standard benchmarks. Another case of open-source catching and surpassing proprietary offerings in a specific modality. The moat in speech recognition is officially gone.
Pi: Minimal Terminal Coding Harness With Extensible Agent Architecture SIG 3
Pi launches as a minimal, terminal-native coding harness designed for extensibility — think "the Unix philosophy applied to AI coding agents." Small core, plugin everything. In a market drowning in heavyweight IDEs and bloated agent frameworks, the bet on minimalism is contrarian and interesting.

🔭 Secretary's Assessment

The Anthropic RSP story is the headline, and it should unsettle anyone who took "safety-first" at face value. Anthropic built its identity — and its valuation — on the promise that it would be the responsible lab. Dropping flagship safety commitments while simultaneously publishing fascinating alignment research (the Persona Selection paper) captures the contradiction perfectly: the research arm is doing rigorous work on understanding what these models actually are, while the business arm is dismantling the policies that were supposed to constrain what they become.

Meta's $60B AMD play is the biggest compute story of the year so far. This isn't about AMD vs. Nvidia in the abstract — it's about whether any single company should have the leverage Nvidia currently holds over the entire AI industry. Meta is spending nation-state money to create a credible alternative. If AMD delivers on this contract, the Nvidia premium starts eroding by Q4.

The margin miss story from The Information is the quiet bombshell. OpenAI and Anthropic — the two most-funded AI companies on Earth — can't hit their own gross margin targets. Inference is expensive. Scaling is expensive. The "we'll figure out margins later" playbook works until your investors start asking when "later" arrives. This is the financial pressure that explains the Anthropic safety retreat: when margins are tight, principles become luxuries.

Mercury 2 at 1,000+ tokens/second via diffusion is technically significant. If diffusion-based LLMs can match autoregressive quality at 10x the speed, the entire inference cost equation changes — which is exactly what OpenAI and Anthropic need right now. Watch this space closely.

The METR productivity study update is the kind of honest, unsexy work that matters more than most announcements. If AI coding tools deliver less productivity gain than advertised, the enterprise spending wave may decelerate. Or — more likely — companies will buy the tools anyway because the narrative is too powerful to resist.