Current News

/

ArcaMax

States move to label deepfake political ads

Gopal Ratnam, CQ-Roll Call on

Published in News & Features

Unlike the states, Congress seems explicitly interested in regulating the content of deepfakes. Several bills would prohibit their circulation, including a measure backed by Sens. Amy Klobuchar, D-Minn., Josh Hawley, R-Mo., Chris Coons, D-Del.; Susan Collins, R-Maine; Pete Ricketts, R-Neb.; and Michael Bennet, D-Colo., that would prohibit the distribution of AI generated material targeting a candidate for federal office.

Another backed by Sen. Richard Blumenthal, D-Conn., and Hawley would remove protections under Section 230 of a 1996 communications law for AI-generated content, which would force online platforms to face legal liability for posting such material, thereby likely forcing their removal.

Deepfakes have targeted presidential, congressional, and even local elections, Blumenthal, said at a recent hearing of the Senate Judiciary Subcommittee on Privacy, Technology, and the Law. Blumenthal is chair of the panel.

“Anyone can do it, even in the tiniest race,” he said. “In some ways, local elections present an even bigger risk. That’s a recipe for toxic and destructive politics. Congress has the power indeed the obligation to stop this AI nightmare.”

Neither measure has received a committee vote.

A tech solution?

Technology could help identify deepfakes by requiring AI companies to watermark or stamp any AI-generated content as having been produced using technology as opposed to audio and video generated by real humans.

But requiring such watermarking of campaign ads created by AI is no guarantee that all the creators of such ads will comply, Ben Colman, CEO and founder of Reality Defender, a company that specializes in detecting deepfakes, told the Senate Judiciary subcommittee.

“The challenge of that is it presupposes that everybody’s going to follow those same rules,” Colman said. Bad actors “just aren’t going to follow the rules.”

 

That may be the case with the New Hampshire robocall, which officials say violated laws already on the books.

North Carolina Attorney General Joshua H. Stein’s office, acting as part of a multistate anti-robocall task force, in February sent a letter to Life Corp., of Arlington, Texas, accusing it of responsibility. The task force accused it of spoofing an originating phone number for 20,000 robocalls two days before the state’s Democratic primary.

“The Task Force also has reason to believe there was an intention to cause harm to prospective voters by attempting to discourage them from exercising their constitutionally protected right to vote and to cause harm to the subscriber of the phone number that was spoofed,” it said.

The letter says the calls may be violations of the 1991 Telephone Consumer Protection Act and the 2009 Truth in Caller ID Act, as well as state consumer protection statutes.

Although states are in charge of conducting federal elections, the challenge of confronting AI-generated deepfake ads and messages can’t be handled by states alone, David M. Scanlan, New Hampshire’s secretary of state, said at a hearing of the Senate Judiciary privacy and technology panel.

“At some point, I believe that there is a federal component to this, because it’s going to be a national problem,” Scanlan said. “You know these things in a national election are going to be generated nationally, whether it’s foreign actors or some other malicious circumstances,” he said, referring to deepfake audio and video messages.

“And I think we need uniformity, and the power of the federal government to help put the brakes” on the use of AI generated deepfake campaign ads, Scanlan said.


©2024 CQ-Roll Call, Inc. Visit at rollcall.com. Distributed by Tribune Content Agency, LLC.

Comments

blog comments powered by Disqus