Lex Fridman Podcast
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
It's, I don't think it's as smart as human yet.
Lex Fridman Podcast
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
I do, like back in the day, I went around saying, like, I do not think that just stacking more layers of transformers is going to get you all the way to AGI.
Lex Fridman Podcast
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
And I think that GPT-4 is past where I thought this paradigm was going to take us.
Lex Fridman Podcast
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
And I, you know, you want to notice when that happens.
Lex Fridman Podcast
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
You want to say like, whoops, well, I guess I was incorrect about what happens if you keep on stacking more transformer layers.
Lex Fridman Podcast
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
And that means I don't necessarily know what GPT-5 is going to be able to do.
Lex Fridman Podcast
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
I'd rather not be wrong next time.
Lex Fridman Podcast
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
It's a bit ambitious to go through your entire life never having been wrong.
Lex Fridman Podcast
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
one can aspire to be well-calibrated, like not so much think in terms of like, was I right, was I wrong, but like when I said 90% that it happened nine times out of 10.
Lex Fridman Podcast
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
Yeah, like oops is the sound we make, is the sound we emit when we improve.
Lex Fridman Podcast
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
I suppose that's the objective function.
Lex Fridman Podcast
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
The name less wrong was, I believe, suggested by Nick Bostrom, and it's after someone's epigraph, I actually forget whose, who said, like, we never become right, we just become less wrong.
Lex Fridman Podcast
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
What's the something?
Lex Fridman Podcast
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
Something that's easy to confess, just error and error and error again, but less and less and less.
Lex Fridman Podcast
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
I mean, the beauty does interact with the screaming horror.
Lex Fridman Podcast
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
Is the beauty in the horror?
Lex Fridman Podcast
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
But like beautiful moments, well, somebody asked GPT,
Lex Fridman Podcast
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
Bing Sidney to describe herself and fed the resulting description into one of the stable diffusion things, I think.
Lex Fridman Podcast
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
She's pretty, and this is something that should have been an amazing moment.
Lex Fridman Podcast
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
The AI describes herself.