Deepfakes Could Finally Bring Accountability to Big Tech Companies

Lawmakers see these fake videos as potential justification to crack down on companies’ legal immunity.

An AFP journalist views a "deepfake" video on January 25, 2019, in Washington, DC. Alexandra Robinson/AFP/Getty Images

Long-frustrated efforts to crack down on harmful content on social media sites may finally get some momentum thanks to an unlikely source: counterfeit videos known as “deepfakes” that experts worry could undermine democracies around the world.

The House Intelligence Committee held its first-ever hearing on deepfakes Thursday, probing the implications of AI technology that can create realistic, counterfeit videos. At the hearing, members of Congress, including committee chairman Adam Schiff (D-Calif.), argued that deepfakes could be justification for altering laws that exempt technology companies from legal liability for harmful content on their platforms.

“If the social media companies can’t exercise the proper standard of care when it comes to a variety of fraudulent or illicit content, then we have to think about whether that immunity still makes sense,” Schiff said. He was referring to Section 230 of the Communications Decency Act of 1996, which gives internet companies immunity from liability for harmful content on their sites. Section 230 was a bipartisan effort, intended to “encourage the unfettered and unregulated development of free speech on the internet,” as one judge said later.

If lawmakers significantly alter Section 230, it could be one of the first cracks in the dam of relative inaction on regulating the tech industry.

Lawmakers have expressed concern over the potential damage deepfakes could do. With the technology, someone could make a fake video of Jeff Bezos saying something he never actually said that would cause the share price of Amazon to tank. Groups could undermine elections by issuing fake statements from candidates and politicians. Across the world, in countries like Myanmar and Gabon, the mere potential of deepfakes is already causing confusion.

On Thursday, several members of Congress asked expert witnesses to weigh in on changing the regulation. “I just want to ask each of you whether we should get back to the requirement of an editorial function and a reasonable standard of care by treating these platforms as publishers,” said Rep. Peter Welch (D-Vt.) “That seems like one fundamental question that we would have to ask because that would be a legislative action.”

Schiff told reporters after the hearing that he’s open to working with other committees on these potential legislative changes.

“The Intelligence Committee has a unique perspective and concern about this issue because of the prospect of foreign interference and manipulation,” he said, noting that other committees will have different but relevant interests in reforming Section 230. “So I think this is an oversight issue for several committees—a legislative issue for several committees. And we’ll be working with our colleagues to determine the right response.”

While Republicans and Democrats in Congress agree that technology companies need to reform their content moderation practices (though they don’t necessarily agree on how), a small number are concerned about unraveling it in potentially damning ways. That includes Sen. Ron Wyden (D-Ore.), an original sponsor of the legislation that included Section 230. Wyden opposed 2018 legislation that reformed Section 230 to make technology companies liable for sex trafficking content. The measure ended up hurting sex workers by taking away avenues for them to advertise online and forcing many onto the streets, into the hands of exploitative pimps. Wyden has said he’s worried future reforms to Section 230 could cause other unforeseen problems.

$500,000 MATCHING GIFT

In 2014, before Donald Trump announced his run for president, we knew we had to do something different to address the fundamental challenge facing journalism: how hard-hitting reporting that can hold the powerful accountable can survive as the bottom falls out of the news business.

Being a nonprofit, we started planning The Moment for Mother Jones: A special campaign to raise $25 million for key investments to make Mother Jones the strongest watchdog it can be. Five years later, readers have stepped up and contributed an astonishing $23 million in gifts and future pledges. This is an incredible statement from the Mother Jones community in the face of the huge threats—both economic and political—against the free press.

Read more about The Moment and see what we've been able to accomplish thanks to readers' incredible generosity so far, and please join them today. Your gift will be matched dollar for dollar, up to $500,000 total, during this critical moment for journalism.

We Recommend

Latest

Sign up for our newsletters

Subscribe and we'll send Mother Jones straight to your inbox.

Get our award-winning magazine

Save big on a full year of investigations, ideas, and insights.

Subscribe

Support our journalism

Help Mother Jones' reporters dig deep with a tax-deductible donation.

Donate

We have a new comment system! We are now using Coral, from Vox Media, for comments on all new articles. We'd love your feedback.