Jaeden Schaefer
๐ค SpeakerAppearances Over Time
Podcast Appearances
It is out on the market.
And so I'm really excited by that.
And I'm, you know, I'm stoked that Google has been pushing so hard and has made
a big upgrade and on on speed specifically now the other thing that i think is funny which is completely not important but the naming convention i appreciate that they're going gemini 3.1 i kind of like opening eyes doing gpt 5.1 5.2 because these aren't like completely new trained models they just have fine-tuned it a lot and the one thing that i'll say is some people be like well why are they having to fine-tune it why don't they just do that before they release it i appreciate getting gemini
three pro out getting to play with that for a couple months, and then having them come up with all of their tweaks.
And a lot of times when they make these tweaks, I'll say when they come up with like, let's say Gemini, Gemini four, a lot of the tweaks that they put into Gemini 3.1, they're kind of these like software integrations, these upgrades, if you've noticed with chat GPT recently, when you ask it some sort of math question actually pulls up a literal calculator inside of chat GPT, where it computes your question for you and shows you on a calculator.
Those are kind of like nice things that you would see, for example, in GPT 5.2, like that calculator feature came out.
And what's nice is when GPT 5.3 comes out, or even when GPT 6 comes out, that calculator tool is getting built in.
So what's exciting to me is when they make these incremental updates, the 3.1, 3.2, 3.3,
All of the little features, all of the little nice to haves and the things that they're kind of building in are going to get rolled over when the whole model gets an entire overhaul.
So that's what I'm excited about.
They were sharing a bunch of the results from some independent evaluations, a bunch of the benchmarks, especially humanity's last exam.
This is kind of, it feels like the AI models cooked a lot of the benchmarks and just kind of beat them.
And they weren't basically hard enough or built well enough for the AI models.
And so now we've kind of come up with some more challenging ones.
One of the more challenging ones is humanity's last exam.
Gemini 3.1 Pro outperformed Gemini 3.
I mean, obviously, if it didn't, I don't think they'd be releasing it to us, but it did it by a huge margin.
So the model is also coming up, climbing on a bunch of real world performance leaderboards.
This is what I think is actually the most important.