AI

Generative AI poses threat to election security, federal intelligence agencies warn


Generative artificial intelligence could threaten election security this November, intelligence agencies warned in a new federal bulletin.

Generative AI uses images, audio, video and code to create new content, like so-called “deep fake” videos in which a person is made to look like they’re saying something they never said. 

Both foreign and domestic actors could harness the technology to create serious challenges heading into the 2024 election cycle, according to the analysis compiled by the Department of Homeland Security and sent to law enforcement partners nationwide. Federal bulletins are infrequent messages to law enforcement partners, meant to call attention to specific threats and concerns. 

“A variety of threat actors will likely attempt to use generative artificial intelligence (AI) – augmented media to influence and sow discord during the 2024 U.S. election cycle, and AI tools could potentially be used to boost efforts to disrupt the elections,” the bulletin, shared with CBS News, stated. “As the 2024 election cycle progresses, generative AI tools likely provide both domestic and foreign threat actors with enhanced opportunities for interference by aggravating emergent events, disrupting election processes, or attacking election infrastructure.”

Director of National Intelligence Avril Haines also warned Congress about the perils of generative AI during a Senate Intelligence Committee hearing last week, saying AI technology can create realistic “deepfakes” whose origin can be concealed. 

“Innovations in AI have enabled foreign influence actors to produce seemingly authentic and tailored messaging more efficiently, at greater scale,” she testified, while insisting the U.S. is better prepared for an election than ever. 

Director of National Intelligence Avril Haines testifies before the Senate Armed Services Committee on May 2, 2024, in Washington, DC. 

Win McNamee/Getty Images


One example the DHS cited in the bulletin was a fake robocall impersonating the voice of President Joe Biden on the eve of the New Hampshire primary in January. The fake audio message was circulated, encouraging recipients of the call to “save your vote” for the November general election instead of participating in the state’s primary. 

The “timing of election-specific AI-generated media can be just as critical as the content itself, as it may take time to counter-message or debunk the false content permeating online,” the bulletin said. 

The memo also noted the lingering threat overseas, adding that in November 2023, an AI video encouraged a southern Indian state to vote for a specific candidate on election day, giving officials no time to discredit the video.

The bulletin goes on to warn about the potential use of artificial intelligence to target election infrastructure. 

“Generative AI could also be leveraged to augment attack plotting if a threat actor, namely a violent extremist, sought to target U.S. election symbols or critical infrastructure,” the bulletin read. “This may include helping to understand U.S. elections and associated infrastructure, scanning internet-facing election infrastructure for potential vulnerabilities, identifying and aggregating a list of election targets or events, and providing new or improved tactical guidance for an attack.”

Some violent extremists have even experimented with AI chatbots to fill gaps in tactical and weapons guidance, DHS said, although the department noted it has not yet observed violent extremists using that technology to supplement election-related target information.



Source

Related Articles

Back to top button