Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Andrew Ilyas

๐Ÿ‘ค Speaker
638 total appearances

Appearances Over Time

Podcast Appearances

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

on clean pictures of dogs and cats.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And so what that means is that this useful versus useless features dichotomy is not enough to explain what happened here.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

There has to be something else at play.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And so by doing this experiment, we basically gave support for this model of non-robust features, whereby not only are there useless features and useful features,

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

But there are actually, within the class of useful features, there are features that are genuinely useful for the classification task at hand, but just happen to be not robust to small perturbations.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And so a very toy example of what that would look like is imagine if there were a small pixel in the top right of every single example that told you exactly what the class was.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

Then you wouldn't have to learn anything about the image.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

You could just get 100% accuracy by looking at this pixel.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

But if you did that, an adversary could come in and by changing just that one pixel completely flip your classification.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And so what we're showing is that obviously this one pixel scenario isn't real, but there are features like this in sort of standard image data sets.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

That's a great question.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

I think it gets back to this idea of disambiguating different notions of robustness.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And so ideally, we want our models to be robust in every sense of the word robustness.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

If they were truly learning human-level features, then we'd want them to A, be adversarially robust,

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

because humans are adversarially robust.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

So we'd like that.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

B, we'd like them to be like sort of data space robust.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And that's sort of what we were talking about earlier that like, you know, your prediction should not hinge on like one or two examples being removed from your training set, because that feels like kind of this exemplar based, not really abstract,

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

like this memorization behavior that we were just talking about.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And then three, which is something that I haven't explored, but also I know is being explored right now, you also want them to learn features that are causally robust, in the sense that you want it to learn to avoid spurious features and