Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Steven Byrnes

๐Ÿ‘ค Speaker
266 total appearances

Appearances Over Time

Podcast Appearances

LessWrong (Curated & Popular)
"Why we should expect ruthless sociopath ASI" by Steven Byrnes

So that's consequentialism, one possible answer for how an AI might accomplish impressive feats, and it's an answer that brings in ruthlessness by default.

LessWrong (Curated & Popular)
"Why we should expect ruthless sociopath ASI" by Steven Byrnes

And then there's a second, different possible answer to how an AI might accomplish impressive feats.

LessWrong (Curated & Popular)
"Why we should expect ruthless sociopath ASI" by Steven Byrnes

Imitative learning from humans.

LessWrong (Curated & Popular)
"Why we should expect ruthless sociopath ASI" by Steven Byrnes

You train an AI to predict what actions a skilled human would take in many different contexts and then have the AI take that same action itself.

LessWrong (Curated & Popular)
"Why we should expect ruthless sociopath ASI" by Steven Byrnes

I claim that LLMs get their impressive capabilities almost entirely from imitative learning.

LessWrong (Curated & Popular)
"Why we should expect ruthless sociopath ASI" by Steven Byrnes

By contrast, true imitative learning is entirely absent and impossible in humans and animals.

LessWrong (Curated & Popular)
"Why we should expect ruthless sociopath ASI" by Steven Byrnes

Imitative learning AIs do not have ruthless sociopathy by default because of course the thing they're imitating is non-ruthless humans.

LessWrong (Curated & Popular)
"Why we should expect ruthless sociopath ASI" by Steven Byrnes

Optimist.

LessWrong (Curated & Popular)
"Why we should expect ruthless sociopath ASI" by Steven Byrnes

Who?

LessWrong (Curated & Popular)
"Why we should expect ruthless sociopath ASI" by Steven Byrnes

Wait.

LessWrong (Curated & Popular)
"Why we should expect ruthless sociopath ASI" by Steven Byrnes

So you're an optimist about superintelligence, ASI, being non-ruthless, as long as people stick to LLMs?

LessWrong (Curated & Popular)
"Why we should expect ruthless sociopath ASI" by Steven Byrnes

Me?

LessWrong (Curated & Popular)
"Why we should expect ruthless sociopath ASI" by Steven Byrnes

Alas, no.

LessWrong (Curated & Popular)
"Why we should expect ruthless sociopath ASI" by Steven Byrnes

I think that the full power of consequentialism is super dangerous by default, and I think that the full power of consequentialism is the only way to get ASI, and so AI researchers are going to keep working until they eventually learn to fully tap that power.

LessWrong (Curated & Popular)
"Why we should expect ruthless sociopath ASI" by Steven Byrnes

In other words, I see a disjunction.

LessWrong (Curated & Popular)
"Why we should expect ruthless sociopath ASI" by Steven Byrnes

Either, LLMs will always get their powers primarily from imitative learning, as I claim they do today, in which case they will never be able to figure things out way beyond the human-created training data, and will thus never reach ASI.

LessWrong (Curated & Popular)
"Why we should expect ruthless sociopath ASI" by Steven Byrnes

And then eventually we'll get ASI via a different AI paradigm, one that can rocket arbitrarily far past any human data.

LessWrong (Curated & Popular)
"Why we should expect ruthless sociopath ASI" by Steven Byrnes

And that paradigm will have to draw its powers from consequentialism, which brings in ruthlessness by default.

LessWrong (Curated & Popular)
"Why we should expect ruthless sociopath ASI" by Steven Byrnes

OR, someone will figure out how to get LLMs themselves to rocket arbitrarily far past human training data and into ASI.

LessWrong (Curated & Popular)
"Why we should expect ruthless sociopath ASI" by Steven Byrnes

But the only way to do that is to somehow modify LLMs to draw on the full powers of consequentialism.