#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
I think that if, you know, like half of today's physicists stop wasting their lives on string theory or whatever and go off and study what goes on inside transformer networks, then in, you know, like 30, 40 years, we'd probably have a pretty good idea.
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
And interestingly, people have started to get preliminary results showing that reinforcement learning by human feedback has made the GPT series worse in some ways.
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
And if you apply reinforcement learning from human feedback, the nice graph of 70% 7 out of 10 sort of flattens out into the graph that humans use, where there's some very improbable stuff.
#368 โ Eliezer Yudkowsky: Dangers of AI and the End of Human Civilization
So it's like it used to be able to use probabilities, but if you apply, but if you try to teach it to talk in a way that satisfies humans, it gets worse at probability in the same way that humans are.