ADVERTISEMENT

Facebook Is Expanding Efforts to Block False Information

The company said that the Associated Press will expand its role as part of Facebook’s third-party fact-checking program.

Facebook Is Expanding Efforts to Block False Information
People write on a chalk board wall at the new Facebook Inc. Frank Gehry-designed MPK 21 office building in Menlo Park, California, U.S. (Photographer: Michael Short/Bloomberg)

(Bloomberg) -- Facebook Inc. said it’s rolling out a slew of new and expanded ways to rein in the spread of misinformation across its websites and apps, amid heightened global scrutiny of social networks’ actions to remove false and violent content.

The company said Wednesday that the Associated Press will expand its role as part of Facebook’s third-party fact-checking program. Facebook also will reduce the reach of Groups that repeatedly share misinformation, such as anti-vaccine views; make Group administrators more accountable for violating content standards; and allow people to remove posts and comments from Facebook Groups even after they’re no longer members.

Facebook’s executives for years have said they’re uncomfortable choosing what’s true and false. Under pressure from critics and lawmakers in the U.S. and elsewhere, especially since the flood of misinformation during the 2016 U.S. presidential campaign, the social media company with 2 billion users has been altering its algorithms and adding human moderators to combat false, extreme and violent content.

“There simply aren’t enough professional fact-checkers worldwide and, like all good journalism, fact-checking takes time,” Guy Rosen, Facebook’s vice president of integrity, and Tessa Lyons, head of news feed integrity, wrote in a blog post. “We’re going to build on those explorations, continuing to consult a wide range of academics, fact-checking experts, journalists, survey researchers and civil society organizations to understand the benefits and risks of ideas like this.”

While Facebook has updated its policies and efforts, content that violates the company’s standards persists. Most recently, the social network was criticized for not quickly removing the video of the mass shooting in New Zealand that was live streamed.

The U.S. 2020 elections will be a test for the new efforts, which come after the platform was used by Kremlin-linked trolls in the leadup to voting in 2016 and 2018. The scope of election integrity problems is "vast," ranging from misinformation designed to suppress voter turnout to sophisticated activity "trying to strategically manipulate discourse on our platforms," said Samidh Chakrabarti, a product management director at Facebook.

Facebook is looking to crack down on fake accounts run by humans. "The biggest change since 2016 is that we’ve been tuning our machine learning systems to be able to detect these manually created fake accounts," Chakrabarti said, adding that the platform removes millions of accounts -- run by both bots and humans -- each day.

The Menlo Park, California-based company has made progress in detecting and removing misinformation designed to suppress the vote -- content ranging from fake claims that U.S. Immigration and Customs Enforcement agents were monitoring the polls to the common tactic of misleading voters about the date of an election. Facebook removed 45,000 pieces of voter-suppression content in the month leading up to the 2018 elections, 90 percent of which was detected before users reported it.

"We continue to see that the vast majority of misinformation around elections is financially motivated," said Chakrabarti. As a result, efforts to remove clickbait benefit election integrity, he said.

To contact the reporters on this story: Selina Wang in San Francisco at swang533@bloomberg.net;Alyza Sebenius in Washington at asebenius@bloomberg.net

To contact the editors responsible for this story: Jillian Ward at jward56@bloomberg.net, Andrew Pollack

©2019 Bloomberg L.P.