'We don't understand the consequences' -  Why I quit OpenAI - The Sunday Story

“我们不理解后果”——我为何离开OpenAI——周日故事

The Story

2026-02-15

34 分钟
PDF

单集简介 ...

AI researcher Zoe Hitzig quit her job at OpenAI this week over “deep reservations” about the company’s strategy, including their decision to pilot running ads on ChatGPT. In a rare interview, she gives her reasons for leaving the industry - a world with access to "an unprecedented archive of human candour". Meanwhile, another AI researcher, this time at Anthropic, also quit with a the stark warning that "the world is in peril". They just the latest in a series of high profile resignations. So what's going on - and what are the big fears for a world increasingly dominated by AI? This podcast was brought to you thanks to the support of readers of The Times and The Sunday Times. Subscribe today: http://thetimes.com/thestory Guests: Zoe Hitzig, former AI researcher at OpenAI.Mark Sellman,Technology Correspondent for the Times. Host: Manveen Rana.  Producer: Dave Creasey. We want to hear from you - email: thestory@thetimes.com Read more: ‘The world is in peril’: AI researchers quit with public warnings Photo: Getty Images. This podcast was brought to you thanks to subscribers of The Times and The Sunday Times. To enjoy unlimited digital access to all our journalism subscribe here. Hosted on Acast. See acast.com/privacy for more information.
更多

单集文稿 ...

  • From The Times and The Sunday Times, this is the story on Sunday.

  • I'm Manvi Rana.

  • Something seems to have changed.

  • People are suddenly waking up to the dangers that AI could be about to unleash on the world.

  • And in the tech world, there's new concerns over the safety of artificial intelligence,

  • sapped the head of Anthropics' AI safety, resigned this Monday.

  • Fears were heightened by two very prominent resignations from within the world of AI this week.

  • Marinak Sharma writing in a post on X that the world is in peril.

  • There's also a creeping sense of alarm about what the technology itself might be capable of doing.

  • If you tell the model it's going to be shut off, it has extreme reactions.

  • It can blackmail the engineer that's going to shut it off if given the opportunity to do so.

  • It's ready to kill someone, wasn't it?

  • I'm not sure if it was Claude or someone else.

  • And then...

  • There are the dangers of whether the AI industry might be about to bring the global economy to its knees.

  • A lot of the things we worry about when we talk about AI's sort of societal dangers kind of pale in comparison to the economic crash that's going to make 2008 look like the best day of your life.

  • So, what are the real risks of AI?

  • We spoke to one of the researchers who resigned this week over fears of where the industry is headed.

  • I just don't think we understand the potential consequences yet for AI.

  • We have some initial signs that it might not be good.