In the absence of fact-checking, false and misleading content can spread like wildfire, distorting public perception and undermining the integrity of information available to users. It becomes even easier for malicious actors to manipulate public opinion, potentially fueling social and political unrest, which is quite concerning. Meta should really reconsider this decision and work towards implementing more robust and effective fact-checking mechanisms to safeguard the quality of information on its platform.
I believe the problem with fact checkers lies in the systems in place to ensure their accuracy. We all have implicit bases and it’s difficult or impossible to not have them occasionally influence our own judgment.
During the COVID pandemic, fact checkers prevented the spread of demonstrable false and dangerous claims, such as claiming ingesting bleach could lessen COVID symptoms. Conversely, these expert fact checkers suppressed reasonable articles that suggested or questioned COVID might have originated from a lab leak. Additionally, fact checkers suppressed any articles suggesting Hunter Biden’s laptop was genuine and contained real evidence. These articles were labeled as misinformation because they contradicted the fact checkers’ political leanings. There was no way to challenge the fact checkers on platforms that matter.
In contrast, community notes allow for the presentation of multiple viewpoints, enabling individuals to form their own informed opinions. Someone needs to fact-check the fact checkers.
Fact-checking needs to be absolutely unbiased and based upon verifiable facts. However the ideal of rigorous peer-review in the sciences has shown that it often fails to eliminate unsound conclusions.
"You and I, if we say a lie we are held responsible for it, so people can trust us. Well, Facebook made a system where the lies repeated so often that people can't tell."
"Both United Nations and Meta came to the same conclusion, which is that this platform Facebook actually enabled genocide that happened in Myanmar. Think about it as, when you say it a million times... it is not just the lie but also it is laced with fear, anger and hate. This is what was prioritized in the design and the distribution on Facebook. It keeps us scrolling, but in countries like Myanmar, in countries like Philippines, in countries where institutions are weak, you saw that online violence became real world violence."
"Fear, anger, hate, lies, salaciousness, this is the worst of human nature... and I think that's what Big Tech has been able to do through social media... the incentive structure is for the worst of who we are because you keep scrolling, and the longer you keep scrolling the more money the platform makes."
"Without a shared reality, without facts, how can you have a democracy that works?"
In the absence of fact-checking, false and misleading content can spread like wildfire, distorting public perception and undermining the integrity of information available to users. It becomes even easier for malicious actors to manipulate public opinion, potentially fueling social and political unrest, which is quite concerning. Meta should really reconsider this decision and work towards implementing more robust and effective fact-checking mechanisms to safeguard the quality of information on its platform.
I believe the problem with fact checkers lies in the systems in place to ensure their accuracy. We all have implicit bases and it’s difficult or impossible to not have them occasionally influence our own judgment.
During the COVID pandemic, fact checkers prevented the spread of demonstrable false and dangerous claims, such as claiming ingesting bleach could lessen COVID symptoms. Conversely, these expert fact checkers suppressed reasonable articles that suggested or questioned COVID might have originated from a lab leak. Additionally, fact checkers suppressed any articles suggesting Hunter Biden’s laptop was genuine and contained real evidence. These articles were labeled as misinformation because they contradicted the fact checkers’ political leanings. There was no way to challenge the fact checkers on platforms that matter.
In contrast, community notes allow for the presentation of multiple viewpoints, enabling individuals to form their own informed opinions. Someone needs to fact-check the fact checkers.
Fact-checking needs to be absolutely unbiased and based upon verifiable facts. However the ideal of rigorous peer-review in the sciences has shown that it often fails to eliminate unsound conclusions.
Partial transcript:
"You and I, if we say a lie we are held responsible for it, so people can trust us. Well, Facebook made a system where the lies repeated so often that people can't tell."
"Both United Nations and Meta came to the same conclusion, which is that this platform Facebook actually enabled genocide that happened in Myanmar. Think about it as, when you say it a million times... it is not just the lie but also it is laced with fear, anger and hate. This is what was prioritized in the design and the distribution on Facebook. It keeps us scrolling, but in countries like Myanmar, in countries like Philippines, in countries where institutions are weak, you saw that online violence became real world violence."
"Fear, anger, hate, lies, salaciousness, this is the worst of human nature... and I think that's what Big Tech has been able to do through social media... the incentive structure is for the worst of who we are because you keep scrolling, and the longer you keep scrolling the more money the platform makes."
"Without a shared reality, without facts, how can you have a democracy that works?"
— Nobel Prize-winning journalist Maria Ressa