Nathan Lambert
๐ค SpeakerAppearances Over Time
Podcast Appearances
You know, chatbots, right? Character AI has shown that they can, like, you know, talk to kids or adults. And, like, it will, like, people feel a certain way, right? And that's unintentional alignment. But, like, what happens when there's intentional alignment deep down on the open source standard? It's a backdoor today for, like, Linux, right?
You know, chatbots, right? Character AI has shown that they can, like, you know, talk to kids or adults. And, like, it will, like, people feel a certain way, right? And that's unintentional alignment. But, like, what happens when there's intentional alignment deep down on the open source standard? It's a backdoor today for, like, Linux, right?
right, that we discover, or some encryption system, right? China uses different encryption than NIST defines, the US NIST, because there's clearly, at least they think there's backdoors in it, right? What happens when the models are backdoors not just to computer systems, but to our minds?
right, that we discover, or some encryption system, right? China uses different encryption than NIST defines, the US NIST, because there's clearly, at least they think there's backdoors in it, right? What happens when the models are backdoors not just to computer systems, but to our minds?
right, that we discover, or some encryption system, right? China uses different encryption than NIST defines, the US NIST, because there's clearly, at least they think there's backdoors in it, right? What happens when the models are backdoors not just to computer systems, but to our minds?
Because once it's open weights, it doesn't like phone home. It's more about, like, if it recognizes a certain system, it could... Now, it could be a backdoor in the sense of, like, hey, if you're building a software, you know, something in software, all of a sudden it's a software agent. Oh, program this backdoor that only we know about.
Because once it's open weights, it doesn't like phone home. It's more about, like, if it recognizes a certain system, it could... Now, it could be a backdoor in the sense of, like, hey, if you're building a software, you know, something in software, all of a sudden it's a software agent. Oh, program this backdoor that only we know about.
Because once it's open weights, it doesn't like phone home. It's more about, like, if it recognizes a certain system, it could... Now, it could be a backdoor in the sense of, like, hey, if you're building a software, you know, something in software, all of a sudden it's a software agent. Oh, program this backdoor that only we know about.
Or it could be, like, subvert the mind to think that, like, XYZ opinion is the correct one.
Or it could be, like, subvert the mind to think that, like, XYZ opinion is the correct one.
Or it could be, like, subvert the mind to think that, like, XYZ opinion is the correct one.
There's this very good quote from Sam Altman who, you know, he can be a hype beast sometime, but one of the things he said, and I think I agree, is that superhuman persuasion will happen before superhuman intelligence. Yeah. And if that's the case, then these things before we get this AGI-ASI stuff, we can embed superhuman persuasion towards our ideal or whatever the ideal of the model maker is.
There's this very good quote from Sam Altman who, you know, he can be a hype beast sometime, but one of the things he said, and I think I agree, is that superhuman persuasion will happen before superhuman intelligence. Yeah. And if that's the case, then these things before we get this AGI-ASI stuff, we can embed superhuman persuasion towards our ideal or whatever the ideal of the model maker is.
There's this very good quote from Sam Altman who, you know, he can be a hype beast sometime, but one of the things he said, and I think I agree, is that superhuman persuasion will happen before superhuman intelligence. Yeah. And if that's the case, then these things before we get this AGI-ASI stuff, we can embed superhuman persuasion towards our ideal or whatever the ideal of the model maker is.
And again, today, I truly don't believe DeepSeek has done this. But it is a sign of what could happen.
And again, today, I truly don't believe DeepSeek has done this. But it is a sign of what could happen.
And again, today, I truly don't believe DeepSeek has done this. But it is a sign of what could happen.
Yeah, recommendation systems hack the dopamine induced reward circuit, but the brain is a lot more complicated. And what other sort of circuits, quote unquote, feedback loops in your brain can you hack slash subvert in ways like recommendation systems are purely just trying to do? you know, increased time and ads and et cetera.
Yeah, recommendation systems hack the dopamine induced reward circuit, but the brain is a lot more complicated. And what other sort of circuits, quote unquote, feedback loops in your brain can you hack slash subvert in ways like recommendation systems are purely just trying to do? you know, increased time and ads and et cetera.
Yeah, recommendation systems hack the dopamine induced reward circuit, but the brain is a lot more complicated. And what other sort of circuits, quote unquote, feedback loops in your brain can you hack slash subvert in ways like recommendation systems are purely just trying to do? you know, increased time and ads and et cetera.