In 2025, frontier AI developers started warning that their AI systems were beginning to cross risk thresholds related to cyber, chemical, and biological capabilities. This is unfortunate given how closed-weight AI systems are persistently vulnerable to prompt-injection attacks and open-weight systems are persistently vulnerable to malicious fine-tuning. This presentation will focus on tools for making frontier AI safeguards “run deep.” In particular, we will focus on technical tools for safeguarding open-weight systems. Finally, we will discuss the challenge of making AI safeguards research matter in the real world. Along the way, we will discuss what AI safety can learn from the design of lightbulbs and why you should keep a close eye on Arkansas Attorney General, Tim Griffin, in 2026.

