Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Corey Knowles

๐Ÿ‘ค Speaker
See mentions of this person in podcasts
8326 total appearances
Voice ID

Voice Profile Active

This person's voice can be automatically recognized across podcast episodes using AI voice matching.

Voice samples: 1
Confidence: Medium

Appearances Over Time

Podcast Appearances

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

We just saw a model drop a couple of days ago where in order to get it two and a half times faster, the cost six X.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

Well, I've got to say at a dollar per million output tokens, you seem to be doing okay with that.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

I keep thinking when we talk like GPUs and what it takes to do the autoregressive approach, you know, this kind of...

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

in a lot of ways could be a smart way to sort of side skirt things like the current memory supply issue, the need to go acquire brink trucks of money and back them up at Jensen Wong's patio door.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

You know, I really think that this is an interesting approach at a prime time for that.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

That makes sense.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

That makes sense.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

So how does this behave with long context?

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

Are we looking at it getting specifically any more expensive or is parallelism playing more of a role as they grow?

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

You know, I would say probably outside of most, you know, maybe enterprise and software applications, but you're dealing with what say the average worker uses a hundred K context is, is plenty for most things.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

You can really do a lot in that range.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

Yeah.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

I was, I was kind of wondering how coherence.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

plays together by not going left to right.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

And I guess that's me thinking of it through how my human brain works.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

Same as with, if I'm running, you know, stable diffusion and comfy UI or something, and you're doing your denoising runs.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

I've been eyeballing how to connect it to my open clock.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

Oh my gosh.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

Excellent.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

Excellent, excellent.