Stephen McAleese
๐ค SpeakerAppearances Over Time
Podcast Appearances
Instead, they predict that the first ASI would have a strange, alien goal that is not compatible with human survival despite the best efforts of its designers to align its motivations with human values.
Quote
All of what we've described here, a bleak universe devoid of fun, in which earth-originating life has been annihilated, is what a sufficiently alien intelligence would most prefer.
We've argued that an AI would want a world where lots of matter and energy was spent on its weird and alien ends, rather than on human beings staying alive and happy and free.
Just like we, in our own ideal worlds, would be spending the universe's resources on flourishing people leading fun lives, rather than on making sure that all our houses contained a large prime number of pebbles.
A misaligned ASI would reshape the world and the universe to achieve its strange goal and its actions would cause the extinction of humanity since humans are irrelevant for the achievement of most strange goals.
For example, a misaligned ASI that only cared about maximizing the number of paperclips in the universe would prefer to convert humans to paperclips instead of helping them have flourishing lives.
The next question is why the authors believe that ASI alignment would be so difficult.
To oversimplify, I think there are three underlying beliefs that explain why the authors believe that ASI alignment would be extremely difficult.
1.
Human values are very specific, fragile, and a tiny space of all possible goals.
2.
Current methods used to train goals into AIs are imprecise and unreliable.
3.
The ASI alignment problem is hard because it has the properties of hard engineering challenges.
One analogy the authors have used before to explain the difficulty of AI alignment is landing a rocket on the moon.
Since the target is small, hitting it successfully requires extremely advanced and precise technology.
In theory this is possible, however the authors believe that current AI creators do not have sufficient skill and knowledge to solve the AI alignment problem.
If aligning an ASI with human values is a narrow target and we have a poor aim, consequently there is a low probability that we will successfully create an aligned ASI and a high probability of creating a misaligned ASI.
Quote