The AI Ad-Hoc Prior Restraint Era Begins
Government vetoes Anthropic's plans, setting a precedent for ad-hoc AI model control.
The White House has directly intervened in Anthropic's deployment plans for its frontier AI model Mythos, blocking the expansion of Project Glasswing that would have given more companies access. This unprecedented move signals the beginning of an ad-hoc prior restraint era in American AI policy, where the government informally decides which models can be made available to which parties. The decision comes amid concerns about cyberattacks and the need for defenders to have advanced AI tools, but the lack of formal regulatory framework raises serious concerns about arbitrariness and potential abuse.
Critics, including Neil Chilson and Dean Ball, argue that this informal approach is worse than a structured regulatory regime, as it favors insiders and creates uncertainty. The White House is now considering an executive order to establish an AI working group to examine procedures for vetting models before release. While some form of prior restraint may be necessary for frontier AI, the ad-hoc implementation risks stifling innovation and creating confusion, especially as labs like Anthropic face pressure from both domestic security concerns and European Union demands for access.
- White House vetoed Anthropic's plan to expand Mythos access via Project Glasswing
- Government considering executive order for AI model review process before release
- Ad-hoc approach criticized for lack of transparency, enabling corruption and insider favoritism
Why It Matters
Professionals must adapt to unpredictable government intervention in AI deployment, affecting innovation and security.