Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Andrew Ilyas

๐Ÿ‘ค Speaker
638 total appearances

Appearances Over Time

Podcast Appearances

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And so what you'd expect, if that was your mental model,

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

what you would expect is for things to generally get worse as the projection dimension goes down, because you're progressively losing more and more information from these random projections.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

What we find, though, is that there's actually sort of like this peaking behavior, where for a while, making the random projection smaller actually gets better and better, and then it starts getting worse.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And so what that indicates is that there's actually some non-compression effect of the random projections that we're doing.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

They're adding something.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And so there's a lot of really interesting work out of a couple of groups in the statistics community that suggests that actually doing the influence function on randomly projected vectors is equivalent to doing some sort of regularized influence function on the original vectors.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And so there's some regularization element or something like that happening that we didn't originally account for when we were writing the track paper.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

Yeah, so that's probably my favorite application from the track paper is that we took this data set, which is this brilliantly designed data set out of MIT.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

I think Ekin Akurek is the lead author on that, of Jacob Andreas' group.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

But they designed this really cool data set called Ftrace.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And the idea behind Ftrace is that they construct this artificial data set where you have a test set that contains a bunch of facts and a train set of Wikipedia abstracts.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And they have annotations for which abstracts entail which facts.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

So for example, in your training set, if you have, you know, like France is a country whose capital is Paris, blah, blah, blah, blah, blah.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And then in your test set, it says like, what's the capital of France?

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

Paris.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

Then like that would be labeled with a logical entailment.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And so the really nice thing about this data set is it sort of gives you this technique for evaluating different data attribution methods by saying, OK, if I take this test example where the test example says what's the capital of France, Paris, and I look at sort of like what the most important examples are for that test example, it should highlight the things that entail that fact.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

Yeah, so in practice, what they found in their paper, which came out a while before TRACKT, is that an information retrieval system beat every single data attribution baseline they tried.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

In particular, the data attribution

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

did not work well.