Ex-OpenAI Policy Lead Slams Company for Altering AI Safety Narrative
April 10, 2025
BruceMitchell
23

A former OpenAI policy researcher, Miles Brundage, recently took to social media to call out OpenAI for what he sees as an attempt to "rewrite the history" of its approach to deploying potentially risky AI systems.
This week, OpenAI released a document detailing its current stance on AI safety and alignment, the process of ensuring AI systems act in predictable and beneficial ways. In it, OpenAI described the development of AGI, or AI systems capable of any task a human can do, as a "continuous path" that involves "iteratively deploying and learning" from AI technologies.
"In a discontinuous world [...] safety lessons come from treating the systems of today with outsized caution relative to their apparent power, [which] is the approach we took for [our AI model] GPT-2," OpenAI stated in the document. "We now view the first AGI as just one point along a series of systems of increasing usefulness [...] In the continuous world, the way to make the next system safe and beneficial is to learn from the current system."
However, Brundage argues that the cautious approach taken with GPT-2 was entirely in line with OpenAI's current iterative deployment strategy.
"OpenAI's release of GPT-2, which I was involved in, was 100% consistent [with and] foreshadowed OpenAI's current philosophy of iterative deployment," Brundage posted on X. "The model was released incrementally, with lessons shared at each step. Many security experts at the time thanked us for this caution."
Brundage, who joined OpenAI as a research scientist in 2018 and later became the company's head of policy research, focused on the responsible deployment of language generation systems like ChatGPT while on OpenAI's "AGI readiness" team.
GPT-2, announced by OpenAI in 2019, was a precursor to the AI systems behind ChatGPT. It could answer questions, summarize articles, and generate text that was sometimes indistinguishable from human writing.
Though GPT-2 might seem basic now, it was groundbreaking at the time. Due to concerns about potential misuse, OpenAI initially withheld the model's source code, instead allowing select news outlets limited access to a demo.
The decision received mixed feedback from the AI community. Some argued that the risks associated with GPT-2 were overstated, and there was no evidence to support OpenAI's concerns about misuse. The AI-focused publication The Gradient even published an open letter urging OpenAI to release the model, citing its technological significance.
OpenAI eventually released a partial version of GPT-2 six months after its announcement, followed by the full system several months later. Brundage believes this was the correct approach.
"What part of [the GPT-2 release] was motivated by or premised on thinking of AGI as discontinuous? None of it," he stated on X. "What's the evidence this caution was 'disproportionate' ex ante? Ex post, it prob. would have been OK, but that doesn't mean it was responsible to YOLO it [sic] given info at the time."
Brundage is concerned that OpenAI's document aims to establish a high burden of proof, where concerns are dismissed as alarmist unless there's overwhelming evidence of imminent dangers. He finds this mindset "very dangerous" for advanced AI systems.
"If I were still working at OpenAI, I would be asking why this [document] was written the way it was, and what exactly OpenAI hopes to achieve by poo-pooing caution in such a lop-sided way," Brundage added.
OpenAI has faced criticism in the past for prioritizing "shiny products" over safety and rushing releases to outpace competitors. Last year, the company disbanded its AGI readiness team, and several AI safety and policy researchers left for rival firms.
The competitive landscape has intensified, with Chinese AI lab DeepSeek's R1 model, which is openly available and matches OpenAI's o1 "reasoning" model on key benchmarks, drawing global attention. OpenAI CEO Sam Altman has acknowledged that DeepSeek has narrowed OpenAI's technological lead, prompting OpenAI to consider accelerating its release schedule.
With OpenAI reportedly losing billions annually and projecting losses to triple to $14 billion by 2026, a faster product release cycle could boost short-term profits but potentially compromise long-term safety. Experts like Brundage are questioning whether this trade-off is worthwhile.
Related article
Google Search Introduces 'AI Mode' for Complex, Multi-Part Queries
Google Unveils "AI Mode" in Search to Rival Perplexity AI and ChatGPTGoogle is stepping up its game in the AI arena with the launch of an experimental "AI Mode" feature in its Search engine. Aimed at taking on the likes of Perplexity AI and OpenAI's ChatGPT Search, this new mode was announced on Wed
YouTube Backs 'No Fakes Act' to Combat Unauthorized AI Replicas
Senators Chris Coons (D-DE) and Marsha Blackburn (R-TN) are once again pushing forward their Nurture Originals, Foster Art, and Keep Entertainment Safe, or NO FAKES, Act. This legislation aims to set clear rules about creating AI-generated copies of someone's face, name, or voice. After being introd
Microsoft Terminates Employee Who Labeled AI Executive as 'War Profiteer'
Microsoft Terminates Employee After Disruption at 50th Anniversary Event
Microsoft has dismissed an employee, Ibtihal Aboussad, following a disruption at the company's 50th-anniversary celebration. According to an email obtained by The Verge, Microsoft cited "acts of misconduct" as the reason for A
Comments (30)
0/200
RoySmith
April 12, 2025 at 8:02:03 PM GMT
OpenAI's attempt to rewrite history on AI safety is just shady. Miles Brundage spilling the tea on social media? Love it! But seriously, OpenAI needs to be more transparent about their safety measures. It's important, guys!
0
WillieHernández
April 11, 2025 at 7:32:08 PM GMT
OpenAIがAI安全性の歴史を書き換えようとするのは怪しいね。Miles Brundageがソーシャルメディアで暴露するなんて、最高!でも本当に、OpenAIは安全対策についてもっと透明性を持つべきだよ。重要なことなんだから!
0
LunaYoung
April 10, 2025 at 6:39:41 PM GMT
A tentativa da OpenAI de reescrever a história sobre segurança em IA é simplesmente suspeita. Miles Brundage revelando tudo nas redes sociais? Adoro! Mas sério, a OpenAI precisa ser mais transparente sobre suas medidas de segurança. É importante, pessoal!
0
WillMitchell
April 12, 2025 at 1:02:18 AM GMT
El intento de OpenAI de reescribir la historia sobre la seguridad en IA es simplemente sospechoso. ¿Miles Brundage revelando todo en las redes sociales? ¡Me encanta! Pero en serio, OpenAI necesita ser más transparente sobre sus medidas de seguridad. ¡Es importante, chicos!
0
AlbertGarcía
April 11, 2025 at 3:56:49 PM GMT
OpenAIs Versuch, die Geschichte der KI-Sicherheit umzuschreiben, ist einfach dubios. Miles Brundage, der das Ganze in den sozialen Medien aufdeckt? Liebe es! Aber im Ernst, OpenAI muss transparenter über ihre Sicherheitsmaßnahmen sein. Das ist wichtig, Leute!
0
CharlesWhite
April 12, 2025 at 2:56:18 PM GMT
I'm kinda torn on this one. On one hand, it's cool that someone's calling out OpenAI for changing their story on AI safety. On the other, it feels a bit like sour grapes from a former employee. Still, it's an interesting read! 🤔
0










OpenAI's attempt to rewrite history on AI safety is just shady. Miles Brundage spilling the tea on social media? Love it! But seriously, OpenAI needs to be more transparent about their safety measures. It's important, guys!




OpenAIがAI安全性の歴史を書き換えようとするのは怪しいね。Miles Brundageがソーシャルメディアで暴露するなんて、最高!でも本当に、OpenAIは安全対策についてもっと透明性を持つべきだよ。重要なことなんだから!




A tentativa da OpenAI de reescrever a história sobre segurança em IA é simplesmente suspeita. Miles Brundage revelando tudo nas redes sociais? Adoro! Mas sério, a OpenAI precisa ser mais transparente sobre suas medidas de segurança. É importante, pessoal!




El intento de OpenAI de reescribir la historia sobre la seguridad en IA es simplemente sospechoso. ¿Miles Brundage revelando todo en las redes sociales? ¡Me encanta! Pero en serio, OpenAI necesita ser más transparente sobre sus medidas de seguridad. ¡Es importante, chicos!




OpenAIs Versuch, die Geschichte der KI-Sicherheit umzuschreiben, ist einfach dubios. Miles Brundage, der das Ganze in den sozialen Medien aufdeckt? Liebe es! Aber im Ernst, OpenAI muss transparenter über ihre Sicherheitsmaßnahmen sein. Das ist wichtig, Leute!




I'm kinda torn on this one. On one hand, it's cool that someone's calling out OpenAI for changing their story on AI safety. On the other, it feels a bit like sour grapes from a former employee. Still, it's an interesting read! 🤔












