Eliezer Yudkowsky — Why AI will kill us, aligning LLMs, nature of intelligence, SciFi, & rationality

伊莱泽·尤德科夫斯基——为什么人工智能会杀死我们,对齐大型语言模型,智能的本质,科幻小说与理性

Dwarkesh Podcast

2023-04-06

4 小时 3 分钟
PDF

单集简介 ...

For 4 hours, I tried to come up reasons for why AI might not kill us all, and Eliezer Yudkowsky explained why I was wrong. We also discuss his call to halt AI, why LLMs make alignment harder, what it would take to save humanity, his millions of words of sci-fi, and much more. If you want to get to the crux of the conversation, fast forward to 2:35:00 through 3:43:54. Here we go through and debate the main reasons I still think doom is unlikely. Watch on YouTube. Listen on Apple Podcasts, Spotify, or any other podcast platform. Read the full transcript here. Follow me on Twitter for updates on future episodes. Timestamps (0:00:00) - TIME article (0:09:06) - Are humans aligned? (0:37:35) - Large language models (1:07:15) - Can AIs help with alignment? (1:30:17) - Society’s response to AI (1:44:42) - Predictions (or lack thereof) (1:56:55) - Being Eliezer (2:13:06) - Othogonality (2:35:00) - Could alignment be easier than we think? (3:02:15) - What will AIs want? (3:43:54) - Writing fiction & whether rationality helps you win Get full access to Dwarkesh Podcast at www.dwarkesh.com/subscribe
更多

单集文稿 ...

  • No, no, no.

  • Misaligned!

  • Misaligned!

  • No, no, no.

  • Not yet.

  • Not now.

  • Nobody's been careful and deliberate now.

  • But maybe at some point in the indefinite future, people will be careful and deliberate.

  • Sure.

  • Let's grant that premise.

  • Keep going.

  • If you try to rouse your planet, there are the idiot disaster monkeys who are like, ooh, ooh.

  • Like, if this is dangerous, it must be powerful, right?

  • I'm going to, like, be first to grab the poison banana.

  • And it's not a coincidence that I can like zoom in and poke at this and ask questions like this and that you did not ask these questions of yourself.

  • You are imagining nice ways you can get the thing,

  • but reality is not necessarily imagining how to give you what you want.

  • Should one remain silent?

  • Should one let everyone walk directly into the whirling racer blades?

  • Like continuing to play out a video game you know you're going to lose.