WASHINGTON (AP) β With only five months before voters head to the polls, the U.S. may be more vulnerable to foreign disinformation aimed at influencing voters and undermining democracy than it was before the 2020 election, the leader of the Senate Intelligence Committee said Monday.
Sen. Mark Warner, a Virginia Democrat, based his warning on several factors: improved disinformation tactics by and , the rise of domestic candidates and groups who are themselves , and the arrival of programs that allow the rapid creation of images, audio and video difficult to tell from the real thing.
In addition, tech companies have rolled back their efforts to protect users from misinformation even as the government's own attempts to combat the problem have become mired in debates about surveillance and censorship.
As a result, the U.S. could face a greater threat of foreign disinformation ahead of the 2024 election than it did in the 2016 or 2020 presidential election cycles, Warner said.
βWe may be less prepared 155 days out in 2024 than we were under President Trump (in 2020),β Warner told The Associated Press in an interview Monday.
Noting similar campaigns in and , security officials, democracy activists and disinformation researchers have warned for years that Russia, China, Iran and domestic groups within the U.S. will use online platforms to spread false and polarizing content designed to influence the race between Trump, a Republican, and President Joe Biden, a Democrat.
Warner's assessment of America's vulnerability comes just weeks after top security officials told the Intelligence Committee that the U.S. has to combat foreign disinformation.
Several new challenges, however, will make safeguarding the 2024 election different than past cycles.
AI programs have already been used to generate misleading content, such as a robocall that telling New Hampshire voters not to cast a ballot in that state's primary. Deceptive deepfakes created with AI programs have also popped up ahead of elections in , , .
Attempts by federal agencies to communicate with tech companies about disinformation campaigns have been complicated by and over the role of government in monitoring political discourse.
Tech platforms have largely moved away from aggressive policies prohibiting election misinformation. X, formerly Twitter, in favor of a hands-off approach that now allows Neo-Nazi hate speech, .
Last year its policy prohibiting debunked election claims and now allows videos that argue the 2020 election was the result of widespread fraud.
Questions about China's influence over TikTok prompted Congress to pass a law that in the U.S. if its Beijing-based owner refuses to divest.
Meta, the owner of Facebook, WhatsApp and Instagram, prohibits information that interferes with election processes and regularly removes foreign influence operations . The platform also says it will label content made with AI. But the company is also allowing political advertisements that , which critics say undercuts its promises.
βI'm not sure that these companies, other than the press release, have done anything in a meaningful way,β Warner said.
Representatives from X and TikTok did not immediately respond to messages on Monday.
David Klepper, The Associated Press