Skip to main content

Ahead of the 2020 presidential election, Facebook says it’s banning deepfakes

Two days before Facebook is set to appear in front of lawmakers at a House Energy and Commerce hearing on manipulated media, the social network has announced it’s banning all forms of deepfakes. The announcement represents a significant step forward for Facebook, which has been struggling to mend its ailing image with the 2020 presidential election right around the corner.

In a blog post, Monika Bickert, Facebook’s vice president of global policy management, said the company will take down videos that have been “edited or synthesized in ways that aren’t apparent to an average person” or are the “product of artificial intelligence or machine learning that merges, replaces or superimposes content onto a video, making it appear to be authentic.”

Recommended Videos

However, the policy doesn’t apply to parodies, satire, or clips “that has been edited solely to omit or change the order of words.” This means it also doesn’t apply to awidely viewed doctored video of House Speaker Nancy Pelosi, as it’s not forged by A.I. and instead, is merely the result of readily available speech software.

Please enable Javascript to view this content

In addition to keeping an eye on A.I.-generated content, Bickert says Facebook is also “partnering with academia, government, and industry to expose people behind these efforts.”

As A.I. tech continues to advance, deepfakes — most of which are nearly impossible to identify — have emerged as one of the most critical challenges for governments and platforms. With a looming presidential election, companies have scrambled to come up with ways to crack down on and spot deepfake videos at scale. In November, Twitter also announced a new draft for deepfake policies, although it has yet to come into effect.

Unlike its peers, Facebook has been in the public crosshairs as it refuses to ban and fact-check political ads.

Last month, a number of organizations and companies including Facebook, Microsoft, the Massachusetts Institute of Technology, University of Oxford, and more joined hands to launch the Deepfake Detection Challenge, which aims to catalyze the development in deepfake detection through large data sets, awards, and research grants.

Shubham Agarwal
Former Digital Trends Contributor
Shubham Agarwal is a freelance technology journalist from Ahmedabad, India. His work has previously appeared in Firstpost…
What the biggest tech companies are doing to make the 2020 election more secure
what the biggest tech companies are doing to secure 2020 election i voted sticker

As we gear up for the November election, all eyes are on tech companies to ensure there’s no spread of misinformation about the voting process or other false claims.

In 2016, Russian aids used Facebook to target Americans on the platform, and the manipulated content spreading fake news reached as many as 126 million Americans, according to The New York Times.

Read more
Facebook says it has helped 2.5 million people register to vote
facebook voter registration 25 million nrp national voters day banner 1

Ahead of National Voter Registration Day tomorrow, Facebook says it is playing its part in getting people registered to vote in the 2020 presidential election.

Facebook announced it has helped 2.5 million people register to vote, and that it aims to get 4 million eligible voters registered. "It’s a promising start," the company announced Monday, "but we have more work to do."

Read more
TikTok says it’s here to stay; ban from U.S. app stores delayed to September 27
tiktok logo

TikTok claims that it is here to stay, shortly after President Donald Trump said that he has approved Oracle's proposed deal to acquire the video-sharing app's U.S. assets.

TikTok said on Twitter said that it is not going anywhere, while also uploading a short video message from interim CEO Vanessa Pappas.

Read more