Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Andy Halliday

๐Ÿ‘ค Speaker
3893 total appearances

Appearances Over Time

Podcast Appearances

The Daily AI Show
From DeepSeek to Desktop Agents

So all of those things are improving on the efficiency scale.

The Daily AI Show
From DeepSeek to Desktop Agents

The second dimension has to do with memory.

The Daily AI Show
From DeepSeek to Desktop Agents

And this is where the new deep seek technique comes in.

The Daily AI Show
From DeepSeek to Desktop Agents

We know that models just left as a dense model and being injected with your prompt and some additional context that you type in at the time of inference, they can be subject to hallucinations.

The Daily AI Show
From DeepSeek to Desktop Agents

And so we like to ground that with a retrieval augmented generation model where you have an external memory, a database that is going to be referenced as context based.

The Daily AI Show
From DeepSeek to Desktop Agents

And semantic relevance is used to selectively retrieve the relevant components of the grounding truth data that's in that retrieval augmented generation, typically a vector database, in order to achieve that semantic retrieval.

The Daily AI Show
From DeepSeek to Desktop Agents

So that's this outboard memory that's used to inform the inference process in an efficient way.

The Daily AI Show
From DeepSeek to Desktop Agents

So those are the two big dimensions of advancement.

The Daily AI Show
From DeepSeek to Desktop Agents

One has to do with sparsity and the other has to do with memory.

The Daily AI Show
From DeepSeek to Desktop Agents

Okay, so memory being like external caching or placement of and then retrieval of static knowledge that doesn't really change and isn't subject to reasoning and manipulation by the computational process in inference.

The Daily AI Show
From DeepSeek to Desktop Agents

Okay, so what did DeepSeek do?

The Daily AI Show
From DeepSeek to Desktop Agents

DeepSeq introduced this thing called Ngram.

The Daily AI Show
From DeepSeek to Desktop Agents

It's a novel module that's added to their LLM that provides conditional memory.

The Daily AI Show
From DeepSeek to Desktop Agents

And here's the jargon.

The Daily AI Show
From DeepSeek to Desktop Agents

It's a complementary axis of sparsity.

The Daily AI Show
From DeepSeek to Desktop Agents

that adds to the conditional computation paradigm of mixture of experts models in large language models.

The Daily AI Show
From DeepSeek to Desktop Agents

So what is this doing?

The Daily AI Show
From DeepSeek to Desktop Agents

What it's doing is it's efficiently identifying the things that are static knowledge in the input context and putting those into a file, in effect, that's sort of a scratchpad kind of memory.

The Daily AI Show
From DeepSeek to Desktop Agents

And this then frees up.