- Facebook‘s chief technology officer, Mike Schroepfer, has been in charge of trying to rid the social platform of false and dangerous content.
- In a recent interview with The New York Times, Schroepfer teared up while discussing the pressure he’s under to fix the problem.
- Facebook has upped its efforts to censor offensive and violent content since the shootings in Christchurch, New Zealand, were livestreamed on the platform.
- Visit Business Insider’s homepage for more stories.
The pressure Facebook has faced trying to eliminate violent and offensive content from the platform is enough to make a grown person cry — literally, if you ask Facebook executive Mike Schroepfer.
Schroepfer, Facebook’s chief technology officer, teared up several times during a series of interviews with The New York Times about the platform’s recent policing efforts. Criticism of the platform has ramped up since the terrorist attacks in March on Christchurch, New Zealand, during which the shooter livestreamed his attack on Facebook.
Schroepfer “choked up” when talking about “the scale of the issues that Facebook was confronting and his responsibilities in changing them,” The Times reported.
“It won’t be fixed tomorrow,” Schroepfer said about Facebook’s efforts. “But I do not want to have this conversation again six months from now. We can do a much, much better job of catching this.”
The CTO is known for “often” letting his feelings shows, “many” people told The Times. A former Facebook employee, Jocelyn Goldfein, a venture capitalist, said she’d seen Schroepfer cry in the office when she worked for the social platform.
Schroepfer has been tasked with building artificial-intelligence tools for Facebook that will better work to detect harmful content, and can prevent something like the Christchurch shooting from being broadcasted on Facebook again.
To figure out how Facebook’s technology can best identify the next terrorist-related video, Schroepfer had to watch the gruesome footage of the shooting “several times,” according to The Times.
“I wish I could unsee it,” Schroepfer said.
Facebook has taken some steps to avoid an incident like the New Zealand shooting livestream from repeating itself. The platform has implemented a “one strike” policy that blocks users immediately from livestreaming if they violate Facebook’s “most serious” rules.
The company has invested $7.5 million into research on better techniques for detecting videos that have been manipulated, which is how millions of repostings of the Christchurch shooting got past Facebook’s automated system and spread online.
Schroepfer told The Times that his task of removing harmful posts is a complex one without an “endgame.”
He said the number of posts is “never going to go to zero.”