Mark Zuckerberg
👤 PersonAppearances Over Time
Podcast Appearances
In terms of how we're going to update the content policies, changing the content filters to have to require higher confidence and precision is actually going to be the thing that reduces the vast majority of the censorship mistakes that we make. Removing the fact checkers and replacing them with community notes, I think it's a good step forward.
like a very small percent of content is fact checked in the first place. So it's, is that going to make the hugest difference? I'm not sure. Um, I think it'll be a positive step though. Um, and we, we like opened up some content policies. So some stuff that was restricted before we opened up, okay, that's good.
like a very small percent of content is fact checked in the first place. So it's, is that going to make the hugest difference? I'm not sure. Um, I think it'll be a positive step though. Um, and we, we like opened up some content policies. So some stuff that was restricted before we opened up, okay, that's good.
like a very small percent of content is fact checked in the first place. So it's, is that going to make the hugest difference? I'm not sure. Um, I think it'll be a positive step though. Um, and we, we like opened up some content policies. So some stuff that was restricted before we opened up, okay, that's good.
It'll mean that some set of things that might've been censored before or not, but by far the biggest set of issues we have, and you and I have talked about a bunch of issues like this over the years is like, it's just, um, Okay, you have some classifier that's trying to find, say, like drug content, right? People decide, okay, it's like the opioid epidemic is a big deal.
It'll mean that some set of things that might've been censored before or not, but by far the biggest set of issues we have, and you and I have talked about a bunch of issues like this over the years is like, it's just, um, Okay, you have some classifier that's trying to find, say, like drug content, right? People decide, okay, it's like the opioid epidemic is a big deal.
It'll mean that some set of things that might've been censored before or not, but by far the biggest set of issues we have, and you and I have talked about a bunch of issues like this over the years is like, it's just, um, Okay, you have some classifier that's trying to find, say, like drug content, right? People decide, okay, it's like the opioid epidemic is a big deal.
We need to do a better job of cracking down on drugs and drug sales. I don't want people dealing drugs on our networks. So we build a bunch of systems that basically go out and try to automate finding people who are dealing drugs. And then you basically have this question. Which is how precise do you want to set the classifier?
We need to do a better job of cracking down on drugs and drug sales. I don't want people dealing drugs on our networks. So we build a bunch of systems that basically go out and try to automate finding people who are dealing drugs. And then you basically have this question. Which is how precise do you want to set the classifier?
We need to do a better job of cracking down on drugs and drug sales. I don't want people dealing drugs on our networks. So we build a bunch of systems that basically go out and try to automate finding people who are dealing drugs. And then you basically have this question. Which is how precise do you want to set the classifier?
So do you want to make it so that the system needs to be 99% sure that someone is dealing drugs before taking them down? Do you want it to be 90% confident, 80% confident? And then those correspond to amounts of... I guess the statistics term would be recall. What percent of the bad stuff are you finding?
So do you want to make it so that the system needs to be 99% sure that someone is dealing drugs before taking them down? Do you want it to be 90% confident, 80% confident? And then those correspond to amounts of... I guess the statistics term would be recall. What percent of the bad stuff are you finding?
So do you want to make it so that the system needs to be 99% sure that someone is dealing drugs before taking them down? Do you want it to be 90% confident, 80% confident? And then those correspond to amounts of... I guess the statistics term would be recall. What percent of the bad stuff are you finding?
So if you require 99% confidence, then maybe you only actually end up taking down 20% of the bad content. Whereas if you reduce it and you say, okay, we're only going to require 90% confidence. Now maybe you can take down 60% of the bad content. But let's say you say, no, we really need to find everyone who is doing this bad thing. And it doesn't need to be as severe as dealing drugs.
So if you require 99% confidence, then maybe you only actually end up taking down 20% of the bad content. Whereas if you reduce it and you say, okay, we're only going to require 90% confidence. Now maybe you can take down 60% of the bad content. But let's say you say, no, we really need to find everyone who is doing this bad thing. And it doesn't need to be as severe as dealing drugs.
So if you require 99% confidence, then maybe you only actually end up taking down 20% of the bad content. Whereas if you reduce it and you say, okay, we're only going to require 90% confidence. Now maybe you can take down 60% of the bad content. But let's say you say, no, we really need to find everyone who is doing this bad thing. And it doesn't need to be as severe as dealing drugs.
It could just be – I mean it could be any kind of content of – any kind of category of harmful content. You start getting to some of these classifiers might have 80%, 85% precision in order to get 90% of the bad stuff down. But the problem is if you're at 90% precision, that means one out of 10 things that the classifier takes down is not actually problematic.
It could just be – I mean it could be any kind of content of – any kind of category of harmful content. You start getting to some of these classifiers might have 80%, 85% precision in order to get 90% of the bad stuff down. But the problem is if you're at 90% precision, that means one out of 10 things that the classifier takes down is not actually problematic.
It could just be – I mean it could be any kind of content of – any kind of category of harmful content. You start getting to some of these classifiers might have 80%, 85% precision in order to get 90% of the bad stuff down. But the problem is if you're at 90% precision, that means one out of 10 things that the classifier takes down is not actually problematic.
And if you kind of multiply that across the billions of people who use our services every day, that is... millions and millions of posts that are basically being taken down that are innocent.