AI Safety

AI #157: Burn the Boats

Pentagon considers extreme options against Anthropic while Claude Opus 4.6 shows rapid capability escalation.

Deep Dive

This week's AI #157 newsletter by Zvi highlights two major developments: a brewing crisis between Anthropic and the Pentagon, and accelerating AI capability benchmarks. The Department of Defense is reportedly considering "extreme options" against Anthropic that could damage national security—a story mainstream media has missed. Simultaneously, Claude Opus 4.6 scored just 14.5 hours on METR's capability timeline graph, indicating AI progress is accelerating faster than anticipated. The newsletter also dissects Citrini's viral scenario where AI agents disrupt the economy by 2028, which caused actual stock market fluctuations despite its speculative nature.

Technically, the post reviews Claude Sonnet 4.6 as a cost-effective option while maintaining Claude Opus as the premium choice, and covers recent releases including Gemini 3.1 Pro (an improvement) and Grok 4.20 (a disappointment). It previews upcoming analyses of Claude Code/Codex #5 with agent capabilities and Anthropic's RSP 3.0 framework with its 100+ page risk report. The India AI Summit coverage reveals diverging narratives between AI labs and global policymakers, while Dean Ball's essay on recursive self-improvement provides crucial technical context. These developments collectively signal both regulatory pressure and technical acceleration reaching inflection points.

Key Points
  • Pentagon considering extreme national security actions against Anthropic, potentially endangering national security
  • Claude Opus 4.6 achieves 14.5-hour METR capability score, showing faster-than-expected progress escalation
  • Citrini's 2028 AI agent disruption scenario caused real stock market impacts despite being speculative

Why It Matters

Regulatory clashes with AI labs could reshape industry governance while capability benchmarks suggest faster timeline to transformative AI.