OpenAI Policy Chief Chris Lehane Calls Out AI "Doomers": Their Ideas Have Consequences

OpenAI's global policy chief Chris Lehane is pushing back publicly against what he calls AI "doomers" — researchers, commentators, and policy advocates who emphasize catastrophic AI risk scenarios. In a new interview, Lehane says the discussion around AI risk has "gotten out of hand" and that doomer ideas, when put into the public sphere, carry real consequences for policy, investment, and public trust in AI development.
Who Lehane Is Targeting
Lehane's comments are aimed at a loose coalition of voices — some within academia, some in the AI safety community, and some in media — who argue that advanced AI poses existential or near-existential risks to humanity. This group has been influential in shaping public narratives and, critically, regulatory proposals in the EU, UK, and US. Lehane appears to be drawing a direct line between doomer rhetoric and policy outcomes that OpenAI views as counterproductive.
The Stakes of the Narrative War
AI policy is increasingly shaped by public perception. When prominent figures argue that AI could end civilization, that framing shapes the questions regulators ask, the restrictions they propose, and the public's tolerance for AI deployment. For OpenAI — which has commercial interests in deploying AI broadly and quickly — a regulatory environment shaped by existential risk thinking is a material business risk. Lehane's pushback is as much strategic as philosophical.
The Counter-Argument
AI safety researchers would argue that Lehane has it exactly backwards: the consequences of underestimating AI risk are far worse than the consequences of overestimating it. Dismissing doomer concerns as harmful speech, they would say, is itself a form of narrative capture — using policy influence to silence legitimate scientific concern. The tension between OpenAI's commercial interests and the AI safety community's independence has been building for years.
Lehane's Political Background
Chris Lehane brings an unusual background to AI policy — he was a political operative for the Clinton White House and a high-profile crisis communications expert before joining OpenAI. His instinct to reframe narratives and go on offense against critics is a political skill, not a scientific one. That context matters when evaluating the substance of his claims.
The Bottom Line
Lehane's comments mark a new phase in OpenAI's public posture: from cautious acknowledgment of AI risks to active pushback against risk-focused critics. Whether this is principled optimism or commercial self-interest depends heavily on who's watching — and what they stand to gain or lose.
Related Articles
- White House Prepares Agencies to Use Anthropic's Mythos Model
- a16z: AI Is an Industrial Revolution-Scale Event