
The Time Sam Altman Asked for a Countersurveillance Audit of OpenAI
Karen Hao
created: May 21, 2025, 11 a.m. | updated: May 23, 2025, 10:02 p.m.
Dario Amodei’s AI safety contingent was growing disquieted with some of Sam Altman’s behaviors.
If AI safety issues actually arose in OpenAI’s models, they worried, those commitments would make it far more difficult, if not impossible, to prevent the models’ deployment.
Within the AI safety contingent, it centered on what they saw as strengthening evidence that powerful misaligned systems could lead to disastrous outcomes.
But late one night, a researcher made an update that included a single typo in his code before leaving the RLHF process to run overnight.
During these heady discussions philosophizing about the long‑term implications of AI research, many employees returned often to Altman’s early analogies between OpenAI and the Manhattan Project.
2 weeks, 3 days ago: WIRED