Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

John Schulman

๐Ÿ‘ค Speaker
528 total appearances

Appearances Over Time

Podcast Appearances

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

And we have a pretty good handle on what it's going to do, what it can do.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

So I think...

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

yeah we would have to be we'd have to be very careful um if it happened way sooner than expected because i think uh our understanding is rudimentary in a lot of ways still

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

Yeah, I would say just like maybe not training the even smarter version, not being really careful when you do train it, that it's not, it's like properly sandboxed and everything.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

Maybe not deploying it at scale or, yeah, being careful

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

Yeah, being careful about what scale you deploy it.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

Yeah, the game theory is a little tough to think through.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

So first of all, I don't think this is gonna happen next year, but it's still useful to have the conversation and maybe it's like two or three years instead.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

But yeah, I guess- Two or three years is still pretty soon.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

Yeah, still pretty soon.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

I do think you probably need some coordination.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

Everyone needs to agree on some reasonable limits to deployment or to further training for this to work.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

Otherwise you have the race dynamics where everyone's trying to stay ahead and that might require compromising on safety.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

So I think you would probably need some coordination

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

uh larger entities that are doing this kind of training and so you're coordinating to um i guess pause deployment until until what exactly like until you figure out what's happening in the models uh either uh further training pause deployment uh like uh avoid certain types of training that we think might be riskier

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

So just like setting up some reasonable rules for like what everyone should do to, yeah, having everyone somewhat limit these things.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

Yeah, that's... I don't have a good answer to that.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

I mean, I would say...

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

um if we can if everyone is gonna coordinate like that uh i think we would be that would be an okay scenario that would be a pretty good scenario because i do think uh like um building these models is very capital intensive and uh there are a lot of complex pieces so it's not like everyone's gonna go and recreate this stuff at home

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

So I think it is possible to do, given the relatively small number of entities who could train the largest models, it does seem possible to coordinate.