Matt Clifford
๐ค SpeakerAppearances Over Time
Podcast Appearances
But it's worth saying, isn't it, going back to your earlier discussion of the technology, I think one thing that is not obvious to a lot of people is that these are not computer programs in the sense that I think most of us traditionally thought of them.
You can't go and say, well, here are the lines of code.
One metaphor, and it is a metaphor, but it's quite helpful, is these are computer programs that are grown.
This is a really hard technical problem, even if we just take out the risk question for a second.
Understanding why a large neural network has done a particular thing is just a very hard technical problem.
This thing is too- It's actually a little bit like neuroscience in the sense that one way of thinking about this is what these layers that Yoshua is talking about doing is building representations of ideas which may or may not map to human formulations of those ideas.
within AI called mechanistic interpretability, which is really trying to almost be the neurosurgeon saying like, if we turn this bit off, does the behavior change?
But it's almost at a very, very basic level, right?
You mean OpenAI's O1 model, the thinking model?
This is really recent, right?
This is the last year we've had thinking models.
You're the godfather of this.
I've worked a lot on these issues, and I also find them troubling, as you know.