Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Ihor Kendiukhov

๐Ÿ‘ค Speaker
515 total appearances

Appearances Over Time

Podcast Appearances

LessWrong (Curated & Popular)
"The Case for Low-Competence ASI Failure Scenarios" by Ihor Kendiukhov

In his framing, the big failure mode of early transformative AGI is that it does not actually solve the alignment problems of stronger AI, and if early AGI makes us think we can handle stronger AI, that is a central path by which we die.

LessWrong (Curated & Popular)
"The Case for Low-Competence ASI Failure Scenarios" by Ihor Kendiukhov

Wentworth's argument maps two main failure channels, one, intentional scheming by a deceptive AGI, and, two, slop where the problem is simply too hard to verify and we convince ourselves we have solved it when we have not.

LessWrong (Curated & Popular)
"The Case for Low-Competence ASI Failure Scenarios" by Ihor Kendiukhov

I want to point at a third channel.

LessWrong (Curated & Popular)
"The Case for Low-Competence ASI Failure Scenarios" by Ihor Kendiukhov

Moderately superhuman AIs that are not particularly capable of doing anything singularity level but are still capable of defeating humanity because of humanity's incompetence.

LessWrong (Curated & Popular)
"The Case for Low-Competence ASI Failure Scenarios" by Ihor Kendiukhov

These AIs are not producing slop.

LessWrong (Curated & Popular)
"The Case for Low-Competence ASI Failure Scenarios" by Ihor Kendiukhov

It ain't much, but it's honest work, they say, as they cooperate with human sympathizers on the development of a super virus.

LessWrong (Curated & Popular)
"The Case for Low-Competence ASI Failure Scenarios" by Ihor Kendiukhov

the research goes slowly, it requires extensive experimentation, to some extent the process is even being documented in public blog posts or on forums.

LessWrong (Curated & Popular)
"The Case for Low-Competence ASI Failure Scenarios" by Ihor Kendiukhov

But no one particularly cares, or rather, the people who care lack the institutional power to do anything about it, and the people who have institutional power are busy with other things, or have been convinced by interested parties that the concern is overblown, or are themselves collaborating.

LessWrong (Curated & Popular)
"The Case for Low-Competence ASI Failure Scenarios" by Ihor Kendiukhov

This is, to some degree, what Andrew Critch describes in What Multipolar Failure Looks Like and Robust Agent-Agnostic Processes, IAPs.

LessWrong (Curated & Popular)
"The Case for Low-Competence ASI Failure Scenarios" by Ihor Kendiukhov

A world where no single system does a theatrical betrayal, but competitive automation yields an interlocking production web where each subsystem is locally acceptable to deploy, governance falls behind the speed and opacity of machine-mediated commerce, and the system's implicit objective gradually becomes alien to human survival.

LessWrong (Curated & Popular)
"The Case for Low-Competence ASI Failure Scenarios" by Ihor Kendiukhov

The difference in my framing is that the AIs in question do not need to be particularly alien or incomprehensible in their goals.

LessWrong (Curated & Popular)
"The Case for Low-Competence ASI Failure Scenarios" by Ihor Kendiukhov

They may have straightforwardly bad goals that are recognizable as bad, and they may be pursuing those goals through channels that are recognizable as dangerous, and the response may still be inadequate.

LessWrong (Curated & Popular)
"The Case for Low-Competence ASI Failure Scenarios" by Ihor Kendiukhov

It is also somewhat similar to what is depicted in A Country of Alien Idiots in a data center, again with one important difference.

LessWrong (Curated & Popular)
"The Case for Low-Competence ASI Failure Scenarios" by Ihor Kendiukhov

Although the AIs in my scenario are not particularly super smart, they are definitely not idiots either.

LessWrong (Curated & Popular)
"The Case for Low-Competence ASI Failure Scenarios" by Ihor Kendiukhov

They are, let us say, slightly above human level in relevant domains, capable of doing cool novel scientific work but not capable of the kind of rapid recursive self-improvement or decisive strategic advantage that most takeover scenarios assume.

LessWrong (Curated & Popular)
"The Case for Low-Competence ASI Failure Scenarios" by Ihor Kendiukhov

They are the kind of system that, in a competent civilization, would be caught and contained.

LessWrong (Curated & Popular)
"The Case for Low-Competence ASI Failure Scenarios" by Ihor Kendiukhov

In the actual civilization we live in, they may not be.

LessWrong (Curated & Popular)
"The Case for Low-Competence ASI Failure Scenarios" by Ihor Kendiukhov

In other words, we do not need to posit for de-chess when ordinary chess is sufficient against an opponent who keeps forgetting the rules.

LessWrong (Curated & Popular)
"The Case for Low-Competence ASI Failure Scenarios" by Ihor Kendiukhov

Heading Undignified AGI disaster scenarios deserve more careful treatment.

LessWrong (Curated & Popular)
"The Case for Low-Competence ASI Failure Scenarios" by Ihor Kendiukhov

As examples, I am talking about such things.