2023-10-31
3 小时 7 分钟Okay, today I have the pleasure of interviewing Paul Cristiano,
who is the leading AI safety researcher.
He's the person that labs and governments turn to when they want feedback and advice on their safety plans.
He previously led the language model alignment team at OpenAI, where he led the invention of RLHF.
And now he is the head of the Alignment Research Center.
And they've been working with the big labs to identify when these models will be too unsafe to keep scaling.
Paul, welcome to the podcast.
Thanks for having me.
Looking forward to talking.
Okay, so first question, and this is a question I've asked,
Holden, Ilya, Dario, and none of that had given me a satisfying answer.
Give me a concrete sense of what a post-AGI world that would be good would look like.
Like, how are humans interfacing with the AI?
What is the economic and political structure?
Yeah, I guess this is a tough question for a bunch of reasons.
Maybe the biggest one is concrete, and I think it's just,
if we're talking about really long spans of time.
then a lot will change and it's really hard for someone to talk completely about what that will look like without saying really silly things.
But I can venture some guesses or fill in some parts.
I think this is also a question of how good is good.