Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Stefano Ermon

๐Ÿ‘ค Speaker
359 total appearances

Appearances Over Time

Podcast Appearances

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

So diffusion is a type of generative AI model.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

It's the kind of model that is commonly used to generate images, video, music.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

And you're probably familiar with the, you know, ChatGPTs or Geminis or Cloud, where you kind of like see the models generate text, kind of like left to right, one token at a time.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

A diffusion model works very differently in the sense that it generates the full object from the beginning, and then it refines it by kind of like

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

fixing mistakes, making it sharper, making it look better and better.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

And it's a very different kind of like solution that it's more parallel in the sense that the neural network is able to modify many components of the image or the text at the same time.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

And that's why diffusion models tend to be a lot faster than traditional autoregressive models that kind of like work left to right one token at a time.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

Yeah.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

That's a great question.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

Yeah, that's a great question.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

And it really stems from the way the models are trained.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

A traditional autoregressive model, like a GPT model, is trained to, there is a neural network and it's trained to predict the next token, the next word.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

And that's how you use it at inference time.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

You give it a question and then it will try to predict the answer left to right, one token at a time.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

The fusion language model, it's trained to remove mistakes, fix mistakes.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

So you kind of like start with clean text or clean code.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

You artificially add mistakes and then you train the model to fix those mistakes.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

And that's how the model is also used at inference time.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

You start with kind of like a full answer and then you refine it.

The Neuron: AI Explained
Diffusion for Text: Why Mercury Could Make LLMs 10x Faster

And so it's a very different way of training the models.