Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Andrew Ilyas

๐Ÿ‘ค Speaker
638 total appearances

Appearances Over Time

Podcast Appearances

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And we basically showed that by plugging in this class of techniques with a super rich history and optimization and plugging that in for what we were doing in the query limited setting, we could significantly accelerate black box attacks.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

Yeah, so I think if you sort of step back and view what we've been talking about so far, I think the data models, data attribution stuff is really focused on understanding this data set and learning algorithm stages of the pipeline.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

A lot of my work on adversarial examples is about understanding this final stage of the pipeline and have some other work on other aspects of the final stage of the pipeline.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

But really, the missing piece is what if your data collection process itself is biased?

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And so I've done a lot of work both with Alexander on the more empirical side, but also with my other advisor, Kostas Daskalakis, on the more theoretical side of what happens when you're doing statistical inference or you're doing machine learning and the way that you collect data is biased in some way.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

Can we find the ways in which that data is biased?

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

and can we correct for it?

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

So I can maybe give you like one highlight from each side, on the empirical side and the theoretical side.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

On the empirical side, probably my favorite one is this work that was led by two students who are now graduated, Dimitris and Shivani, Dimitris Tsipras and Shivani Santaker.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And it was this really, really cool work that I was happy to be a part of that was about studying the ImageNet dataset, which like we think of as this sort of static, you know, like represented, or we used to think of at least at the time of this paper as like this representative benchmark of image classification.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And so they, you know, we went in there and tried to uncover sort of like how the ImageNet dataset was collected, you know, what possible biases could emerge.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And, you know, in general we found, so like the way ImageNet was collected just as a,

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

As a note, the way it was actually collected is the original authors scraped Flickr with a bunch of tags.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

So for a given search term, they turned that class into a bunch of search terms.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

They searched Flickr for the search terms.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And then they uploaded those images to Mechanical Turk and asked people, does this image contain this class?

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

Yes or no.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And so as a result, there are a bunch of really interesting biases that kind of creep in unexpectedly from doing this.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

So for example, for confusing classes, like if there are two breeds of dogs that look very similar, because you're not actually asking people, is this breed A or breed B?

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

You're just asking them, is this an example of breed A?