© 2024 Ideastream Public Media

1375 Euclid Avenue, Cleveland, Ohio 44115
(216) 916-6100 | (877) 399-3307

WKSU is a public media service licensed to Kent State University and operated by Ideastream Public Media.
Play Live Radio
Next Up:
0:00
0:00
0:00 0:00
Available On Air Stations

AI fakes raise election risks as lawmakers and tech companies scramble to catch up

Voters wait to cast their ballots on Jan. 23 in Loudon, New Hampshire. Shortly before voting began, some voters in the state got calls from a faked version of President Biden's voice urging them not to vote, a sign of the potential that deepfakes could have on the electoral process.
Tasos Katopodis
/
Getty Images
Voters wait to cast their ballots on Jan. 23 in Loudon, New Hampshire. Shortly before voting began, some voters in the state got calls from a faked version of President Biden's voice urging them not to vote, a sign of the potential that deepfakes could have on the electoral process.

"What a bunch of malarkey." That's what thousands of New Hampshire voters heard last month when they received a robocall purporting to be from President Biden. The voice on the other end sounded like the president, and the catchphrase was his. But the message that Democrats shouldn't vote in the upcoming primary election didn't make sense.

"Your vote makes a difference in November, not this Tuesday," the voice said.

It quickly emerged that the voice wasn't Biden at all. It was the product of artificial intelligence. Bloomberg reported that ElevenLabs, maker of the AI voice-cloning software believed to have made the digital voice, banned the account involved. On Tuesday, New Hampshire's attorney general said a Texas telemarketing company was behind the call and was being investigated for illegal voter suppression.

On Thursday, the Federal Communications Commission ruled robocalls using AI-generated voices illegal under federal telecoms law, opening the door to fines and lawsuits against violators.

Faking a robocall is not new. But making a persuasive hoax has gotten easier, faster and cheaper thanks to generative AI tools that can create realistic images, video and audio depicting things that never happened.

As AI-generated deepfakes are being used to spread false information in elections around the world, policymakers, tech companies and governments are trying to catch up.

"We don't really think of [AI] as a free-standing threat but as more of a threat amplifier," said Dan Weiner, director of the elections and government program at the Brennan Center for Justice at New York University School of Law.

He worries that AI will turbocharge efforts to discourage voters or spread bogus claims, especially in the immediate run-up to an election, when there's little time for journalists or campaigns to fact-check or debunk.

That's what appears to have happened last fall in Slovakia, just days before voters went to the polls. Faked audio seeming to show one candidate discussing rigging votes and raising the cost of beer started to spread online. His pro-Western party ended up losing to one led by a pro-Russian politician.

Because the stakes were high and the deepfake came at a critical moment, "there is a plausible case that that really did impact the outcome," Weiner said.

While high-profile fakes like the Biden robocall get a lot of attention, Josh Lawson, director of AI and democracy at the Aspen Institute, is focused on how AI could be used for personalized targeting.

"We are quickly advancing towards a point in the technology, likely before the election itself, when you can have real-time synthetic audio conversations," said Lawson, a former election lawyer who previously worked on elections at Facebook owner Meta.

He imagines a scenario where a bad actor deploys AI, sounding like a real person, to call a voter and give false information about their specific polling place. That could be repeated for other voters in multiple languages.

He's also worried about AI fakes targeting lower-profile elections, especially given the collapse of local news.

"The concern ... is not the big, bad deepfake of somebody at the top of the ticket, where all kinds of national press is going to be out there to verify it," Lawson said. "It's about your local mayor's race. It's about misinformation that's harder and harder for local journalists to tackle, when those local journalists exist at all. And so that's where we see synthetic media being something that will be particularly difficult for voters to navigate with candidates."

Deceiving voters, including spreading false information about when and where to vote, is already illegal under federal law. Many states prohibit false statements about candidates, endorsements or issues on the ballot.

But growing concerns about other ways that AI could warp elections are driving a raft of new legislation. While bills have been introduced in Congress, experts say states are moving faster.

In the first six weeks of this year, lawmakers in 27 states have introduced bills to regulate deepfakes in elections, according to the progressive advocacy group Public Citizen.

"There's huge momentum in the states to address this issue," Public Citizen President Robert Weissman said. "We're seeing bipartisan support ... to recognize there is no partisan interest in being subjected to deepfake fraud."

Many state-level bills focus on transparency, mandating that campaigns and candidates put disclaimers on AI-generated media. Other measures would ban deepfakes within a certain window — say 60 or 90 days before an election. Still others take aim specifically at AI-generated content in political ads.

These cautious approaches reflect the need to weigh potential harms against free speech rights.

"It is important to remember that under the First Amendment, even if something is not true, generally speaking you can't just prohibit lying for its own sake," Weiner said. "There is no truth-in-advertising rule in political advertising. You need to have solutions that are tailored to the problems the government has identified."

Just how prominent a role deepfakes end up playing in the 2024 election will help determine the shape of further regulation, Weiner said.

Tech companies are weighing in too. Meta, YouTube and TikTok have begun requiring people to disclose when they post AI content. Meta said on Tuesday that it is working with OpenAI, Microsoft, Adobe and other companies to develop industrywide standards for AI-generated images that could be used to automatically trigger labels on platforms.

But Meta also came under fire this week from its own oversight board over its policy prohibiting what it calls "manipulated media." The board, which Meta funds through an independent trust, said the policy is "incoherent" and contains major loopholes, and it called on the company to overhaul it.

"As it stands, the policy makes little sense," said Michael McConnell, the board's co-chair. "It bans altered videos that show people saying things they do not say, but does not prohibit posts depicting an individual doing something they did not do. It only applies to video created through AI, but lets other fake content off the hook. Perhaps most worryingly, it does not cover audio fakes, which are one of the most potent forms of electoral disinformation we're seeing around the world."

The moves to develop laws and guardrails reining in AI in elections are a good start, said Lawson, but they won't stop determined bad actors.

He said voters, campaigns, lawmakers and tech platforms have to adapt, creating not just laws but social norms around the use of AI.

"We need to get to a place where things like deepfakes are looked at almost like spam. They're annoying, they happen, but they don't ruin our day," he said. "But the question is, this election, are we going to have gotten to that place?"

Copyright 2024 NPR. To see more, visit https://www.npr.org.

Shannon Bond is a business correspondent at NPR, covering technology and how Silicon Valley's biggest companies are transforming how we live, work and communicate.