Taylor Mullen
๐ค SpeakerAppearances Over Time
Podcast Appearances
It's because we have a massive open source community that's helping us build this out.
There's always a where does it lose track and how much do you allow?
So first of all, one of the things that I felt very strongly about is
We're Google.
We're building Gemini CLI.
You'd darn well be able to use all one million of those tokens if you want to.
And so this is super important because a lot of products will just restrict the boundary because it's either more expensive or for varying reasons.
We allow users to restrict it further if they want to.
But we'll never restrict you not to be able to do it, if that makes sense.
You always have the a million token context open to you.
And so where Gemini falls over from our experiments is it's not actually a clear-cut answer.
It's a little different for every single scenario.
So for some coding tasks, it could fall over super quickly.
For others, it could go forever and you'll never see a difference at all.
If you give it several books of information, because you can easily give it several books of information, a lot of the times it can be coherent.
I think my personal hot take here is I think,
In industry, a lot of people look at the context window and see these artificial limits and think, oh, it falls over, it gets not lucid after a certain amount of tokens.
When in reality, there's been so much back and forth.
in the conversation, like if you were to give this conversation to a human and say, okay, like what guidelines do you want to follow?
Would the human even be successful?