Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Andrew Ilyas

๐Ÿ‘ค Speaker
638 total appearances

Appearances Over Time

Podcast Appearances

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

So the thing that inspired them, the right way to interpret them, is as coefficients of a model that tries to predict these data counterfactuals.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And so we were like, OK, if what these things are doing are trying to predict data counterfactuals, then the way we should evaluate them is basically we should try to tell how good they are at evaluating data counterfactuals.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

So the LDS is exactly generating a bunch of data counterfactuals, seeing what these different methods predict will happen under those data counterfactuals, and then looking at the correlation between predictions and reality.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

Yeah, of course.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

So like I was saying, the key idea behind TRAQ is really this approximation of your neural network as a linear model in parameter space.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And so if you take a step back and you think about, what if we were just doing logistic regression?

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

What if there were no neural network or anything?

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

It turns out that in this case, like we were saying earlier, the influence function is a really good data model.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

It's a very good approximation for getting data counterfactuals.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

The problem is that we're not doing logistic regression.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

We have this crazy big neural network or whatever.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

So let's start with thinking about a two-class neural network.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

So this is still a binary classification problem, but now we have a neural network instead of logistic regression.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

What we can do is we can train that neural network, or we can apply our learning algorithm once, and we get a neural network.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

Once we do that, that neural network has a corresponding parameter vector theta, which are just the weights of all of the neural network.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

What we're going to do then is treat the output of that neural network as a linear function in theta by doing a Taylor approximation around those final parameters.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

So what that looks like is if your normal neural network was like f of, you know,

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

you're going to make a new neural network that looks like f hat of theta, which is now a linear function in theta times your gradient evaluated at theta star.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

And so this is a very classical trick, especially in deep learning theory.

Machine Learning Street Talk (MLST)
Adversarial Examples and Data Modelling - Andrew Ilyas (MIT)

It's called the empirical neural tangent kernel.