Ilya Sutskever
๐ค SpeakerAppearances Over Time
Podcast Appearances
It will continue to improve, but it will also not be it.
So the it, we don't know how to build.
And I think that a lot hinges on understanding reliable generalization.
And I'll say another thing, which is like...
You know, one of the things that you could say is what would that cause alignment to be difficult is that human value that it's it's your ability to learn human values is fragile, then your ability to optimize them is fragile, you will you actually learn to optimize them?
And then can't you say are these not all instances of unreliable generalization?
Why is it that human beings appear to generalize so much better?
What if generalization was much better?
What would happen in this case?
What would be the effect?
But those questions are right now still unanswerable.
So one reason why I liked the AI that cares for sentient life, and we can debate on whether it's good or bad, but if the first N of these dramatic systems actually do care for love humanity or something, care for sentient life, obviously this also needs to be achieved.
This needs to be achieved.
So if this is achieved by the first N of those systems,
then I can see it go well, at least for quite some time.
And then there is the question of what happens in the long run.
What happens in the long run?
How do you achieve a long-run equilibrium?
And I think that there is an answer as well.
And I don't like this answer, but it needs to be considered.