When a government buyer (here, the U.S. Department of Defense) labels a commercial model a supply‑chain risk or withdraws a contract over usage restrictions, AI firms face a concrete choice: keep restrictive, rights‑protecting terms that limit lucrative government business, or loosen promises to preserve market access. That dynamic creates an implicit governance lever — procurement exclusion — that can either discipline or co‑opt private safety commitments.
— This reframes AI governance as not only about law and standards but about procurement power that can force companies to choose between ethics and revenue, affecting how models are built and used at scale.
Yascha Mounk
2026.03.07
100% relevant
Anthropic’s contract with the Department of Defense, Emil Michael’s renegotiation effort, and the designation of Claude as a supply‑chain risk in 2025–2026.
← Back to All Ideas