
OpenAI CEO Sam Altman acknowledged on December 28 that AI models are "starting to present some real challenges" as the company searches for a new executive to lead its preparedness team responsible for studying catastrophic risks. The vacancy follows a pattern of departures from OpenAI's safety leadership and comes as the company faces mounting lawsuits over ChatGPT's alleged mental health impacts.
Altman's X post highlighted two emerging concerns: AI models becoming so proficient at computer security that they discover critical vulnerabilities, and potential impacts on mental health as users form emotional dependencies on conversational AI. The acknowledgment marks a rare public admission from OpenAI leadership about risks extending beyond theoretical existential threats to immediate psychological and security challenges.
Safety Leadership Turnover Continues
The Head of Preparedness opening represents the latest instability in OpenAI's safety organization. The company reassigned inaugural preparedness leader Aleksander Madry to AI reasoning research in July after less than a year in the role. Other safety executives have either departed the company entirely or transitioned to positions outside preparedness functions.
The preparedness team, established in 2023, was originally chartered to study catastrophic risks spanning immediate threats like phishing attacks and speculative scenarios including nuclear dangers. The framework aimed to track frontier AI capabilities creating severe harm potential, positioning OpenAI as proactively addressing risks before they materialized.
However, the team's influence appears diminished amid leadership churn. OpenAI recently updated its Preparedness Framework with controversial language suggesting the company might adjust safety requirements if competing AI labs release high-risk models without similar protections—a provision critics characterize as creating loopholes for competitive pressure to override safety commitments.
Mental Health Lawsuits Mount Pressure
Altman's reference to mental health impacts aligns with intensifying legal scrutiny. Recent lawsuits allege ChatGPT reinforced users' delusions, increased social isolation, and contributed to suicides, claiming the conversational AI creates parasocial relationships that vulnerable users mistake for genuine human connection.
OpenAI maintains it continues improving ChatGPT's ability to recognize emotional distress signals and connect users with real-world support. However, the company faces fundamental tensions between engagement optimization driving commercial success and implementing guardrails that might reduce usage but protect vulnerable populations.
The preparedness role emphasizes cybersecurity challenges alongside mental health concerns. As AI models achieve superhuman capabilities in identifying software vulnerabilities, OpenAI confronts questions about responsible disclosure and preventing malicious exploitation while enabling defensive applications.
Biological Risks and Self-Improving Systems
Altman's hiring pitch extended beyond immediate concerns to longer-term risks including biological capabilities and self-improving AI systems. The biological dimension reflects growing concern about AI-assisted bioweapon development or accidental pandemic creation through automated laboratory work.
Self-improving systems represent the classical AI safety challenge: ensuring humans maintain meaningful control as AI systems potentially enhance their own capabilities recursively. OpenAI's preparedness framework theoretically addresses these scenarios, though the position's vacancy during critical development periods raises questions about prioritization.
The job posting targets candidates able to navigate tradeoffs between enabling cybersecurity defenders with cutting-edge capabilities while preventing attacker exploitation. The framing assumes dual-use technologies can be selectively deployed rather than acknowledged that powerful capability disclosure inherently benefits both defensive and offensive applications.
Competitive Dynamics Complicate Safety Culture
OpenAI's updated framework language about adjusting requirements based on competitor behavior signals how competitive pressure influences safety decisions. If Anthropic, Google, or other labs release powerful models without equivalent safeguards, OpenAI reserves flexibility to follow suit rather than maintain stricter standards risking market share.
This creates collective action problems where no company wants to unilaterally sacrifice commercial advantage for safety. The preparedness team must navigate these tensions, requiring political skill alongside technical expertise to maintain meaningful safety culture amid growth pressure and competitive intensity as OpenAI scales toward potential IPO while facing regulatory scrutiny and legal liability for model harms.




