Dario Amodei
๐ค SpeakerAppearances Over Time
Podcast Appearances
How can you imagine that they're not going to take over?
And my intuition is somewhere in the middle, which is that, look, these, you know, you can't just give instructions.
I mean, we try, but you can't just...
have these things do exactly what you want to do.
They're more like growing a biological organism.
But there is a science of how to control them.
Like early in our training, these things are often unpredictable, and then we shape them.
We address problems one by one.
So I have more of a...
not a fatalistic view that these things are uncontrollable, not a, what are you talking about?
What could possibly go wrong?
But I like, this is a complex engineering problem.
And I think something will go wrong with someone's AI system, hopefully not ours, not because it's an insoluble problem, but again, this, and this is the constant challenge because we're moving so fast.
So a couple of points.
Right now, the agents don't learn continuously, right?
And so we just deploy these agents and they have a fixed set of weights.
And so the problem is only that they're interacting in a million different ways.
And so there's a large number of situations and therefore a large number of things that could go wrong, but it's the same agent.
It's like it's the same person.
So the alignment is a constant thing.