Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Ryan Kidd

๐Ÿ‘ค Speaker
958 total appearances

Appearances Over Time

Podcast Appearances

Future of Life Institute Podcast
Can AI Do Our Alignment Homework? (with Ryan Kidd)

Your stock price is going to plummet.

Future of Life Institute Podcast
Can AI Do Our Alignment Homework? (with Ryan Kidd)

What do you do?

Future of Life Institute Podcast
Can AI Do Our Alignment Homework? (with Ryan Kidd)

Do you revert to an older system?

Future of Life Institute Podcast
Can AI Do Our Alignment Homework? (with Ryan Kidd)

That's safer, probably.

Future of Life Institute Podcast
Can AI Do Our Alignment Homework? (with Ryan Kidd)

So I think, yeah, we should definitely be tracking this stuff.

Future of Life Institute Podcast
Can AI Do Our Alignment Homework? (with Ryan Kidd)

And I wouldn't say that we are in the clear by a long shot.

Future of Life Institute Podcast
Can AI Do Our Alignment Homework? (with Ryan Kidd)

I would say that we are in a better world, by my estimation, than Bostrom and Meary predicted 10-something years ago.

Future of Life Institute Podcast
Can AI Do Our Alignment Homework? (with Ryan Kidd)

But I don't know.

Future of Life Institute Podcast
Can AI Do Our Alignment Homework? (with Ryan Kidd)

They would say I'm very wrong about that.

Future of Life Institute Podcast
Can AI Do Our Alignment Homework? (with Ryan Kidd)

But I don't know.

Future of Life Institute Podcast
Can AI Do Our Alignment Homework? (with Ryan Kidd)

I think that it's useful that we can get some work out of these things that looks like it is actually quite likely to accelerate AI safety work.

Future of Life Institute Podcast
Can AI Do Our Alignment Homework? (with Ryan Kidd)

It's a very good question.

Future of Life Institute Podcast
Can AI Do Our Alignment Homework? (with Ryan Kidd)

And I'll preface by saying that all safety work is capabilities work.

Future of Life Institute Podcast
Can AI Do Our Alignment Homework? (with Ryan Kidd)

Fundamentally, people like to distinguish these things in terms of like, oh, capabilities work is about the engine.

Future of Life Institute Podcast
Can AI Do Our Alignment Homework? (with Ryan Kidd)

It's about making the plane go faster.

Future of Life Institute Podcast
Can AI Do Our Alignment Homework? (with Ryan Kidd)

And safety work is about the directionality.

Future of Life Institute Podcast
Can AI Do Our Alignment Homework? (with Ryan Kidd)

But as you've pointed out, ROHF, which was intended as safety work to help the directionality steer it to where you want to go, also made people realize, oh, wait, this thing is useful.

Future of Life Institute Podcast
Can AI Do Our Alignment Homework? (with Ryan Kidd)

I can actually hop in this plane now because it's going to land where I want.

Future of Life Institute Podcast
Can AI Do Our Alignment Homework? (with Ryan Kidd)

which made them want to make the engine go faster so they could get there faster, right?

Future of Life Institute Podcast
Can AI Do Our Alignment Homework? (with Ryan Kidd)

And that whole feedback loop started.