Organize new AI‑safety organizations around heavy use of AI automation and agentic workflows (evaluations, red‑teaming, data curation, reporting) so a small, lean team can scale safety work against rapidly improving capabilities. These labs prioritize building automated tooling and agentic pipelines as the core product, not as an augmentation to large human teams.
— If successful, such labs change who can produce credible safety evaluations, accelerate the pace of safety tooling, and shift regulatory and funding questions toward provenance, auditability, and the governance of automated testing pipelines.
Scott Alexander
2026.01.05
100% relevant
ACX grantee Jacob Arbeid is soliciting a cofounder for a grant‑funded ‘automation‑first’ AI safety lab to scale evaluations and safety engineering with AI agents (article item #2).
← Back to All Ideas