Menu
Sign In Pricing Add Podcast

Ilya Shumailov

Appearances

Short Wave

When AI Cannibalizes Its Data

136.718

Quite a lot of these models, especially back at the time, they're relatively low quality. So there are errors and there are biases. There are systematic biases inside of those models. And thus, you can kind of imagine the case where rather than learning useful contexts, and useful concepts, you can actually learn things that don't exist. They are purely hallucinations.

Short Wave

When AI Cannibalizes Its Data

176.363

In simple theoretical setups, we consider it, you're guaranteed to collapse.

Short Wave

When AI Cannibalizes Its Data

235.583

So there are three sources, three primary sources of error that we still have. So the very first one is basically just data-associated errors. And usually those are questions along the lines of, do we have enough data to approximate a given process? So if some things happen very infrequently in your underlying distribution, your model may get a wrong perception that, like,

Short Wave

When AI Cannibalizes Its Data

263.939

that some things are impossible wait what do you mean by they are impossible like an example i've seen on twitter was uh if you google for a baby peacock you'll discover pictures of birds that look relatively realistic but they are not peacocks at all they are completely generated and you will not find a real picture but if you try learning anything from it of course you're

Short Wave

When AI Cannibalizes Its Data

289.558

I've got to be absorbing this bias.

Short Wave

When AI Cannibalizes Its Data

305.337

Exactly. And those are the kinds of errors that you don't normally see that often because they are so improbable, right? And if people are going to start reporting things to you and saying, oh, your model is wrong here, they're likely to notice things that on average are wrong.

Short Wave

When AI Cannibalizes Its Data

320.068

But if they're wrong in some small part of the internet that nobody really cares about, then it's very unlikely that you will even notice that you're making a mistake. And usually this is the problem because... As the number of dimensions grow, you will discover that the volume in the tails is going to grow disproportionately.

Short Wave

When AI Cannibalizes Its Data

342.688

Yeah, exactly. So as a result, you'll discover that you need to capture quite a bit.

Short Wave

When AI Cannibalizes Its Data

352.222

On top of it, we have errors that come from learning regimes and from the models themselves. So on learning regimes, we are all training our models. All of them are structurally biased. So basically to say that your model is going to be good, But it's unlikely to be optimal. So it's likely to have some errors somewhere. And this was the error source number two.

Short Wave

When AI Cannibalizes Its Data

376.454

And error source number three is that the actual model design, what shape and form your model should be taking, is very much alchemy. Nobody really knows why stuff works. We kind of just know empirically stuff works.

Short Wave

When AI Cannibalizes Its Data

398.365

Yeah, which parts of the model are responsible for what? We don't know the fundamental underlying bias of a given model architecture. What we observe is that there is always some sort of an error that is introduced by those architectures.

Short Wave

When AI Cannibalizes Its Data

425.181

Exactly. And then we also have empirical errors from, for example, hardware. So we also have practical limitations of hardware with which we work. And those errors also exist.

Short Wave

When AI Cannibalizes Its Data

447.526

Yes, certainly. So what we observe in simple theoretical models is that two main phenomena happen. The very first phenomena that happens is it's really hard to approximate improbable events, in part because you don't encounter them very often. So you may discover that you're collecting more and more data, and a lot of this data looks very similar to what you already possess.

Short Wave

When AI Cannibalizes Its Data

470.953

So you're not discovering too much information. But importantly, you're not discovering those infrequent data points. So those tail events, they kind of disappear. And then the other thing that happens is that the first time you made this error and underestimated your improbable events... When you hit the model on top of this, it's unlikely to recover from this taking place.

Short Wave

When AI Cannibalizes Its Data

506.331

Originally improbable events are even more improbable for the subsequent model, and it kind of like snowballs out of control until the whole thing just collapses fully to near zero variance.

Short Wave

When AI Cannibalizes Its Data

523.446

Exactly. And the thing is, you can theoretically describe this. It's actually very simple. And you can run these experiments however many times you want. And you'll discover that even if you have a lot of data, if you keep on repeating this process, and the rate at which this collapses, you can also bound, you end up always in a state where your improbable events kind of disappear.

Short Wave

When AI Cannibalizes Its Data

545.469

In practice, when we grab large language models, we observe that they become more confident in the predictions that they are making. So basically, the improbable events here are going to be things that the model is not very confident about, and normally it would not make predictions about it.

Short Wave

When AI Cannibalizes Its Data

563.465

So when you're trying to generate more data out of a language model in order for another language model to learn from it, over time, basically, it becomes more and more confident. And then it basically, during the generation setup, it gets stuck very often in these repetitive loops.

Short Wave

When AI Cannibalizes Its Data

588.056

a person the same two sentences and then like the next person says the same two sentences and it usually gets like more and more garbled as it goes down the line i think this is a comparison kind of kind of works yes so this is the first thing it's the improbable events and then the second thing that happens is your models are going to produce errors so misunderstandings of the underlying phenomenon right and as a result

Short Wave

When AI Cannibalizes Its Data

59.105

Large language models are statistical beasts that learn from example of human written text and learn to produce text that is similar to the ones that the model was taught.

Short Wave

When AI Cannibalizes Its Data

612.779

what you will see is that those errors start propagating as well. And they are relatively correlated. If all of your models are using the same architecture, then it's likely to be correlatedly wrong in the same kinds of way. So whenever it sees errors, it may amplify the same errors that it's observing.

Short Wave

When AI Cannibalizes Its Data

659.045

Yeah, so approximations of approximations of approximations end up being very imprecise. As long as you can bound the errors of your approximations, it's okay, I guess. But yeah, in practice, because machine learning is very empiric, quite often we can't.

Short Wave

When AI Cannibalizes Its Data

679.473

Yeah. So an important thing to say here is that... The settings we talk about here are relatively hypothetical in a sense that we are not in the world in which, you know, today we can build a model and tomorrow they disappear. That is not going to happen. We already have very good models and the way forward is having even better models. And there is no doubts about this.

Short Wave

When AI Cannibalizes Its Data

719.353

I mean, there are many different solutions. You'll find a lot of different papers that are exploring what are the most effective mitigations. And it's mostly data filtering of different kinds. And basically making sure that the data that ends up being ingested by the models is representative of the underlying data distribution.

Short Wave

When AI Cannibalizes Its Data

740.811

And whenever we hit this limit and we see that our model diverges into some sort of a training direction,

Short Wave

When AI Cannibalizes Its Data

746.896

trajectory that is making the model worse i promise you people will stop training of the models retract back a couple of steps maybe add additional data of certain kind and keep on training right because we can always go back to previous models nothing stopping us and then we can always spend more effort getting high quality data paying more people to create high quality data

Short Wave

When AI Cannibalizes Its Data

770.842

Yeah, so model collapse is not going to magically kill the models tomorrow. We just need to change the way we build stuff. So this is not all doom and gloom. I am quite confident we'll solve this problem.

Short Wave

When AI Cannibalizes Its Data

790.487

Thank you very much for having me. It was a pleasure. Thank you.

Short Wave

When AI Cannibalizes Its Data

97.791

If you were today to sample data from internet randomly, I'm sure you'll find that a bigger proportion of it is generated by machines. But this is not to say that the data itself is bad. The main question is how much of this data is generated