Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Sam Altman

๐Ÿ‘ค Speaker
See mentions of this person in podcasts
3374 total appearances

Appearances Over Time

Podcast Appearances

Lex Fridman Podcast
#367 โ€“ Sam Altman: OpenAI CEO on GPT-4, ChatGPT, and the Future of AI

You know, it wasn't the underlying model that mattered.

Lex Fridman Podcast
#367 โ€“ Sam Altman: OpenAI CEO on GPT-4, ChatGPT, and the Future of AI

It was the usability of it, both the RLHF and the interface to it.

Lex Fridman Podcast
#367 โ€“ Sam Altman: OpenAI CEO on GPT-4, ChatGPT, and the Future of AI

So we train these models on a lot of text data.

Lex Fridman Podcast
#367 โ€“ Sam Altman: OpenAI CEO on GPT-4, ChatGPT, and the Future of AI

And in that process, they learn the underlying something about the underlying representations of what's in here.

Lex Fridman Podcast
#367 โ€“ Sam Altman: OpenAI CEO on GPT-4, ChatGPT, and the Future of AI

are in there and they can do amazing things but when you first play with that base model that we call it after you finish training it can do very well on evals it can pass tests it can do a lot of you know there's knowledge in there but it's not very useful

Lex Fridman Podcast
#367 โ€“ Sam Altman: OpenAI CEO on GPT-4, ChatGPT, and the Future of AI

or at least it's not easy to use, let's say.

Lex Fridman Podcast
#367 โ€“ Sam Altman: OpenAI CEO on GPT-4, ChatGPT, and the Future of AI

And RLHF is how we take some human feedback.

Lex Fridman Podcast
#367 โ€“ Sam Altman: OpenAI CEO on GPT-4, ChatGPT, and the Future of AI

The simplest version of this is show two outputs, ask which one is better than the other, which one the human raters prefer, and then feed that back into the model with reinforcement learning.

Lex Fridman Podcast
#367 โ€“ Sam Altman: OpenAI CEO on GPT-4, ChatGPT, and the Future of AI

And that process works remarkably well with, in my opinion, remarkably little data to make the model more useful.

Lex Fridman Podcast
#367 โ€“ Sam Altman: OpenAI CEO on GPT-4, ChatGPT, and the Future of AI

So RLHF is how we

Lex Fridman Podcast
#367 โ€“ Sam Altman: OpenAI CEO on GPT-4, ChatGPT, and the Future of AI

align the model to what humans want it to do.

Lex Fridman Podcast
#367 โ€“ Sam Altman: OpenAI CEO on GPT-4, ChatGPT, and the Future of AI

Maybe just because it's much easier to use.

Lex Fridman Podcast
#367 โ€“ Sam Altman: OpenAI CEO on GPT-4, ChatGPT, and the Future of AI

It's much easier to get what you want.

Lex Fridman Podcast
#367 โ€“ Sam Altman: OpenAI CEO on GPT-4, ChatGPT, and the Future of AI

You get it right more often the first time, and ease of use matters a lot, even if the base capability was there before.

Lex Fridman Podcast
#367 โ€“ Sam Altman: OpenAI CEO on GPT-4, ChatGPT, and the Future of AI

To be fair, we understand the science of this part at a much earlier stage than we do the science of creating these large pre-trained models in the first place, but yes, less data.

Lex Fridman Podcast
#367 โ€“ Sam Altman: OpenAI CEO on GPT-4, ChatGPT, and the Future of AI

Much less data.

Lex Fridman Podcast
#367 โ€“ Sam Altman: OpenAI CEO on GPT-4, ChatGPT, and the Future of AI

That's so interesting.

Lex Fridman Podcast
#367 โ€“ Sam Altman: OpenAI CEO on GPT-4, ChatGPT, and the Future of AI

We spend a huge amount of effort pulling that together from many different sources.

Lex Fridman Podcast
#367 โ€“ Sam Altman: OpenAI CEO on GPT-4, ChatGPT, and the Future of AI

There are open source databases of information.

Lex Fridman Podcast
#367 โ€“ Sam Altman: OpenAI CEO on GPT-4, ChatGPT, and the Future of AI

We get stuff via partnerships.