Sam Altman wants a full-time worrier for AI's worst-case scenarios

OpenAI is hiring a Head of Preparedness to stress-test AI and blunt worst-case harms. Think mental health risks, cyber misuse, guardrails, and go/no-go calls before release.

Published on: Dec 28, 2025
Sam Altman wants a full-time worrier for AI's worst-case scenarios

OpenAI Wants a Head of Preparedness - a Role Built to Anticipate AI's Worst-Case Scenarios

OpenAI is hiring a Head of Preparedness. Sam Altman announced the role on X and acknowledged that fast-improving models create "some real challenges," including risks to mental health and the rise of AI-enabled cybersecurity threats.

The job is blunt about its goal: think through how advanced systems can cause severe harm, then build processes to find, measure, and neutralize those risks before they show up in the wild.

What the Role Owns

  • Track "frontier" capabilities that introduce new avenues for severe harm.
  • Design and run capability evaluations, threat models, and mitigations end to end.
  • Stand up an operationally scalable safety pipeline that teams can actually use.
  • Execute a preparedness framework for release decisions, including policies for biological capability risks.
  • Set guardrails for self-improving systems.

Altman also noted this will be "a stressful job." Given the stakes, that feels like an understatement.

Why This Matters Now

Recent cases have raised alarms about chatbots contributing to self-harm among teens. Beyond individual incidents, there's growing concern that bots can reinforce delusions, amplify conspiracy content, and help people hide disordered behaviors.

On the security front, we're seeing fast iteration cycles, cheap experimentation, and new attack surfaces. That combination invites misuse, from automated phishing and deepfake social engineering to model-assisted exploit discovery.

A Practical Lens for Researchers and Technical Teams

If you build, deploy, or study advanced models, treat preparedness as a product requirement, not an afterthought. Start simple, make it measurable, and tie it to shipping gates.

  • Adopt a risk framework: Map system context, harms, and controls. The NIST AI RMF is a solid starting point. NIST AI Risk Management Framework
  • Stand up red-teaming and evals: Test prompt-injection, model spoofing, data exfiltration, jailbreaks, and social abuses before release.
  • Harden your stack: Secrets isolation, least-privilege connectors, rate limits, content filters, and abuse detection. For app builders, review the OWASP LLM Top 10. OWASP Top 10 for LLM Applications
  • Guard mental health surfaces: Prohibit clinical advice, add crisis routing, and cap conversational depth on sensitive topics. Escalate to human review where possible.
  • Define go/no-go criteria: Pre-release eval thresholds, incident playbooks, rollback plans, and a real "kill switch."
  • Document and audit: Model cards, hazard analyses, change logs, and post-incident reviews. Keep a feedback loop with users and internal reviewers.

The Bigger Signal

Creating a dedicated owner for worst-case scenarios signals where the field is heading: less hype, more accountability. As capabilities climb, the cost of sloppy thinking grows.

This role isn't about fear. It's about building the muscle to measure risks, make hard calls, and ship responsibly under pressure.

Next Steps for Your Team

Preparedness isn't a press release. It's discipline, clear thresholds, and the willingness to pause when the data says stop. That's how you ship systems people can trust.


Get Daily AI News

Your membership also unlocks:

700+ AI Courses
700+ Certifications
Personalized AI Learning Plan
6500+ AI Tools (no Ads)
Daily AI News by job industry (no Ads)