Rob Wiblin
๐ค SpeakerAppearances Over Time
Podcast Appearances
So it's like AI for...
coordination, compromise, negotiation, truth-seeking, that cluster of things.
Because I think something like the question of space governance, how do we divide up the resources of space if there are some existing factions that have an existing distribution of power?
No one really wants the sort of destruction that comes from everybody racing as hard as possible to get there first.
But there's like a complicated space of like negotiated options beyond that.
And I think AIs could potentially help a lot with that sort of thing.
Yeah, I would think so.
I think if you look at public communications from at least OpenAI, Anthropic, and Google DeepMind, this sort of jumps out more or less in these different cases.
But
in all of their stated safety plans, you see this element of as AIs get better and better, they're going to incorporate the AIs themselves into their safety plans more and more.
And I think some are more explicit than others about expecting some sort of specific crunch time that occurs when AI is rapidly accelerating AI R&D.
But everybody is picturing AIs playing a heavy role in the safety of future AIs.
Yeah, I think fundamentally,
you need it to be the case that there exists a window of opportunity where before AIs are uncontrollably powerful or have created unacceptable levels of risk, where they are really capable and really change the game for AI safety research, and that there's some meaningful window of time where you can notice as you're approaching it and
Even by default, without like crazy slowdown, it lasts at least six months or lasts a year.
If you think instead that once your AI sort of hits upon some generality threshold, it like within a matter of days or weeks becomes crazy super intelligent, this plan doesn't work because like, you know, you wouldn't even notice probably before it's too late.
So, and then I think there's also, there can also be unlucky orderings of capabilities where this plan wouldn't work.
Where you could have AIs that are like really specifically good at AI R&D, and they're really not good at anything else.
Not even AI safety research that's very similar to AI R&D.
They're just like extremely good at AI R&D.