Joscha Bach
๐ค SpeakerAppearances Over Time
Podcast Appearances
And for me, it would be very interesting to experiment with this, to basically build a system like a cat, probably should be careful at first, build something that's small, that's limited, has limited resources that we can control and study how systems notice a self-model, how they become self-aware in real time.
And for me, it would be very interesting to experiment with this, to basically build a system like a cat, probably should be careful at first, build something that's small, that's limited, has limited resources that we can control and study how systems notice a self-model, how they become self-aware in real time.
And for me, it would be very interesting to experiment with this, to basically build a system like a cat, probably should be careful at first, build something that's small, that's limited, has limited resources that we can control and study how systems notice a self-model, how they become self-aware in real time.
And I think it might be a good idea to not start with a language model, but to start from scratch using principles of self-organization.
And I think it might be a good idea to not start with a language model, but to start from scratch using principles of self-organization.
And I think it might be a good idea to not start with a language model, but to start from scratch using principles of self-organization.
My intuition is that the language models that we are building are golems. They are machines that you give a task and they're going to execute the task until some condition is met. And there's nobody home. And the way in which nobody is home leads to that system doing things that are undesirable in a particular context.
My intuition is that the language models that we are building are golems. They are machines that you give a task and they're going to execute the task until some condition is met. And there's nobody home. And the way in which nobody is home leads to that system doing things that are undesirable in a particular context.
My intuition is that the language models that we are building are golems. They are machines that you give a task and they're going to execute the task until some condition is met. And there's nobody home. And the way in which nobody is home leads to that system doing things that are undesirable in a particular context.
So you have that thing talking to a child and maybe it says something that could be shocking and traumatic to the child. Or you have that thing writing a speech and it introduces errors in the speech that your human being would ever do if they were responsible. But The system doesn't know who's talking to whom. There is no ground truth that the system is embedded into.
So you have that thing talking to a child and maybe it says something that could be shocking and traumatic to the child. Or you have that thing writing a speech and it introduces errors in the speech that your human being would ever do if they were responsible. But The system doesn't know who's talking to whom. There is no ground truth that the system is embedded into.
So you have that thing talking to a child and maybe it says something that could be shocking and traumatic to the child. Or you have that thing writing a speech and it introduces errors in the speech that your human being would ever do if they were responsible. But The system doesn't know who's talking to whom. There is no ground truth that the system is embedded into.
And of course, we can create an external tool that is prompting our language model always into the same semblance of ground truth. But it's not like the internal structure is causally produced by the needs of a being to survive in the universe. It is produced by imitating structure on the internet.
And of course, we can create an external tool that is prompting our language model always into the same semblance of ground truth. But it's not like the internal structure is causally produced by the needs of a being to survive in the universe. It is produced by imitating structure on the internet.
And of course, we can create an external tool that is prompting our language model always into the same semblance of ground truth. But it's not like the internal structure is causally produced by the needs of a being to survive in the universe. It is produced by imitating structure on the internet.
Maybe it's sufficient to use the transformer with the different loss function that optimizes for short-term coherence rather than next token prediction over the long run. We had many definitions of intelligence and history of AI. Next token prediction was not very high up on the list. And there are some similarities, like cognition as data compression is an old trope.
Maybe it's sufficient to use the transformer with the different loss function that optimizes for short-term coherence rather than next token prediction over the long run. We had many definitions of intelligence and history of AI. Next token prediction was not very high up on the list. And there are some similarities, like cognition as data compression is an old trope.
Maybe it's sufficient to use the transformer with the different loss function that optimizes for short-term coherence rather than next token prediction over the long run. We had many definitions of intelligence and history of AI. Next token prediction was not very high up on the list. And there are some similarities, like cognition as data compression is an old trope.
Solomonov induction, where you are trying to understand intelligence as predicting future observations from past observations, which is intrinsic to data compression. Mm-hmm. And predictive coding is a paradigm with this boundary between neuroscience and physics and computer science.
Solomonov induction, where you are trying to understand intelligence as predicting future observations from past observations, which is intrinsic to data compression. Mm-hmm. And predictive coding is a paradigm with this boundary between neuroscience and physics and computer science.