Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Andrew Ilyas

๐Ÿ‘ค Speaker
638 total appearances

Appearances Over Time

Podcast Appearances

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

published experiments from social sciences and stuff.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And they found that by removing just a couple of data points from these surveys analyses, they could flip the conclusions.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And so we basically were able to do the same thing in the context of deep learning by using data models.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

So figuring out like which exact training examples do you need to drop to flip a model's prediction on a given test example.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

So I'd say those were sort of all of the

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

original applications of data models.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

I think the ones that I'm more excited about now, as I was mentioning earlier, the ones that I'm more excited about now are more of the flavor of you have some property that you want from your model.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And that property can be written as a function of the dataset that it's trained on as a function of model predictions.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And so for example, like, you know, dataset selection or even machine teaching.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

But if you think about dataset selection, you can sort of write like, okay, I want, you know, max over dataset of model performance.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And so, you know, that's obviously a very expensive problem to solve manually because, you know, inside this maximization, you have like the entire model training process.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

But if you can take that model training process and replace it with a data model prediction, then it becomes a much more tractable optimization problem.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

Yeah, so I think one very nice thing that's a little bit specific to these linear data models is the fact that after you estimate them, what you have is sort of a vector whose length is the size of whatever pool of data you're sampling data sets from.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And the nice thing about these vectors is that each index in the vector corresponds to a single training example.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And the value of that index is exactly how important that training example was to a prediction.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And so what that means is that you can actually estimate these vectors for one learning algorithm.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And you can estimate them for another learning algorithm.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And the resulting representations that you get mean the same thing by default.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

So you can component-wise compare these representations.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And we actually had some follow-up work led by a great student in our lab where we used this exact property to compare different learning algorithms, for example.