Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

John Schulman

๐Ÿ‘ค Speaker
528 total appearances

Appearances Over Time

Podcast Appearances

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

So I guess there might be some practical questions like that that would also determine how things play out.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

Maybe if you just require people to be accountable for various liability, this would also change the incentives a bit.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

So if it turned out that...

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

AIs are better at running everything and they're also completely benevolent and we've totally solved alignment and they're better at being accountable to people than people are, then I would say maybe it's okay having the AIs run the firms, but I think that might be pretty far out.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

we're more likely to be in a situation where they look better like in the short term, but they still have some problems, like the AI run entities still have some serious problems.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

And it's actually like practical considerations that push you more towards having humans in the loop, at least for the near future.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

If the models are being used for these higher stakes use cases, then we would have to think about RLHF in a much different way than we are right now.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

So I would say we're not quite ready for that, or the current methods might not be completely sufficient, but I would say...

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

I would say we would need to make compromises between the needs of the different stakeholders involved.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

So we have this document that we're releasing called the model spec, and it's about how we want our models to behave in the API and in ChatGBT.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

And we try to talk about this issue where there are different stakeholders involved, and sometimes there are conflicts between what they might want, like the...

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

uh in our case we were thinking of the stakeholders as uh the user uh or the end user that means like someone sitting in front of chat gbt or or some other app um the developer so this is like someone using the api um who might be serving other end users with their app like the um

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

the platform, which is open AI.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

We don't want the models to expose us to legal risk and so forth.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

And then the rest of humanity, including people who might not be users or customers or anything.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

So obviously, the user might ask,

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

ask the model to do something that we think is like actively harmful to other people.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

And so we might have to refuse that.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

By the way, this isn't the order of priority necessarily.

Dwarkesh Podcast
John Schulman (OpenAI Cofounder) - Reasoning, RLHF, & Plan for 2027 AGI

So this is just like, we have these four or so classes of stakeholder.