The report, released Tuesday, aims to show how Facebook is taking action against content that violates its standards. The staggering number of fake accounts it disabled in the period fell from 694 million in the fourth quarter of 2017. The report didn’t reveal earlier data.
The first-quarter report also said Facebook acted on 836 million pieces of spam content, 2.5 million pieces of hate speech content, 1.9 million pieces of terrorist propaganda content, 21 million pieces of adult nudity and sexual activity content and 3.4 million pieces of graphic violence content.
Facebook executives vowed to increase transparency in the wake of recent controversies involving the spread of fake news and the and the unauthorized harvesting of personal data.
“It’s a good move and it’s a long time coming,” Jillian York, director for international freedom of expression at the Electronic Frontier Foundation, told The New York Times of the new report. “But it’s also frustrating because we’ve known that this has needed to happen for a long time. We need more transparency about how Facebook identifies content, and what it removes going forward.”
The report said Facebook increasingly relies on AI to flag unsavory content. AI tools detected 98.5 percent of the fake accounts that were shut down, according to the report, and almost all of the spam content acted upon.
“Technology isn’t going to solve all of it, but we will make progress,” Guy Rosen, who heads Facebook’s team policing community standards, told The Financial Times.
The report acknowledged that Facebook’s metrics tracking its response to content that violates standards are still being refined.
“This is the start of the journey and not the end of the journey and we’re trying to be as open as we can,” said Richard Allan, Facebook’s vice president of public policy for Europe, the Middle East and Africa.
Facebook a day earlier announced it had suspended about 200 apps while it investigates whether any of them contributed to the misuse of data.