╌>

Russian propaganda is using deepfakes, sham websites and social media swarms

  

Category:  News & Politics

Via:  hallux  •  one week ago  •  8 comments

By:   Shannon Bond - NPR

Russian propaganda is using deepfakes, sham websites and social media swarms

S E E D E D   C O N T E N T


A deepfake video of a State Department official falsely claiming a Russian city is a legitimate target for Ukrainian strikes using U.S. weapons.

Pro-Russia social media accounts amplifying stories about divisive political topics such as   immigration   and campus protests over the war in Gaza.

Sham news sites spoofing real publications or   posing as legitimate-sounding outlets   with names like D.C. Weekly, the Boston Times and Election Watch.

Russian propaganda is ramping up in a   busy global election year , targeting American voters as well as   elections in Europe   and the Paris Olympics, according to intelligence officials, internet researchers and tech companies.

“Russia remains the most active foreign threat to our elections,”   Director of National Intelligence Avril Haines told senators   last month at a briefing about election risks.

Influence operations linked to Russia take aim at a disparate range of targets and subjects around the world. But their hallmarks are consistent: attempting to erode support for Ukraine, discrediting democratic institutions and officials, seizing on existing political divides and   harnessing new artificial intelligence tools .

"They're often producing narratives that feel like they're throwing spaghetti at a wall," said Andy Carvin, managing editor at the Atlantic Council's Digital Forensic Research Lab, which tracks online information operations. "If they can get more people on the internet arguing with each other or trusting each other less, then in some ways their job is done."

Some efforts have been linked directly to the Kremlin, including a   network of fake accounts and phony news websites given the name Doppelganger , whose operators have been sanctioned by both the U.S. and the European Union.

The origins of others are still unknown, such as the fabricated video of State Department spokesman Matthew Miller, in which reporters' questions and Miller's response about U.S. policy in the Ukraine war were faked, likely with the help of artificial intelligence. The video circulated on Russian Telegram channels and was picked up by Russian state media and government officials, according to   The New York Times .

Russia employed tactics such as exacerbating existing divisive issues and creating fake accounts posing as Americans in its 2016 and 2020 election-meddling efforts, researchers say. Since Russia invaded Ukraine in early 2022, discrediting Ukraine and amplifying voices in the U.S. and other countries that oppose aid to Ukraine and support for NATO has become a dominant theme of the Kremlin's efforts.

"What you can see is they are referencing politics in a certain country, and they are generally tying that to what is going on in Ukraine. The underlying message is, 'Here's why people should not support Ukraine,'" said Ben Nimmo, principal investigator on OpenAI’s intelligence and investigations team, who previously led global threat intelligence at Facebook's owner, Meta.

Fake accounts, phony websites


The Kremlin relies on what Haines called "a vast multimedia influence apparatus, which consists of its intelligence services, cyber-actors, state media proxies and social media trolls" to pump out propaganda, launder fake and misleading news articles and circulate conspiracy theories.

Since the invasion of Ukraine, the European Union has banned Russian media outlets including RT, Sputnik, Voice of Europe and RIA Novosti from publishing or broadcasting within the bloc. That hasn’t stopped RT articles from proliferating across hundreds of other websites widely available in Europe, according to a recent  report  from the German Marshall Fund of the United States, the University of Amsterdam and the Institute for Strategic Dialogue.

"We discovered RT articles reposted to third-party websites targeting audiences from Iraq to Ethiopia to New Zealand, often without any indication that the content was sourced from a Russian propaganda outlet," the researchers wrote.


Perhaps the most persistent and prevalent Russian online operation is Doppelganger. First identified by researchers at the   EU DisinfoLab   in 2022, the campaign has impersonated news outlets including the U.K.'s   The Guardian , Germany's   Der Spiegel ,   The Washington Post   and Fox News, and it has posed as NATO, the Polish and Ukrainian governments, the German police and the French Foreign Ministry.

In addition to operating fake accounts and phony websites, the operation purchased Facebook ads targeting French and German audiences with messages about aid to Ukraine, farmers' protests and the war in Gaza, according to the European nonprofit   AI Forensics .

Doppelganger has also set its sights on the Paris Olympics,   Microsoft   said in a report this week. It used fake French-language news sites to push claims of corruption in the Games' organizing body and to warn of potential violence.

In March, the U.S. Treasury sanctioned two Russian companies identified as being behind Doppelganger — Social Design Agency and Structura — as well as their founders, saying they carried out the campaign "at the direction of the Russian Presidential Administration."

The misinformation-tracking company   NewsGuard   has connected a separate network of 167 websites "masquerading as independent local news publishers in the U.S." to a former deputy sheriff from Florida who now lives in Moscow.

Using AI tools to create propaganda


The volume of posts, articles and websites that Russian-linked operations produce is being boosted by artificial intelligence — another new factor that sets 2024 apart from previous election cycles.

Covert influence campaigns based in Russia, as well as in China, Iran and Israel, have begun using AI in their attempts to manipulate public opinion and shape politics,   according to recent reports from OpenAI, Meta and Microsoft .

A Russian operation that Microsoft calls Storm-1679 used AI to fake actor Tom Cruise's voice narrating a phony Netflix documentary disparaging the International Olympic Committee.

According to OpenAI, Doppelganger has used its AI tools, which include ChatGPT, to translate articles into other languages and generate social media posts and comments. Another Russian operation, dubbed Bad Grammar, used AI to debug code for a program that automatically posted on Telegram.


The question remains: How effective are Russia's attempts to influence public opinion and democratic elections?

Many online operations that have been publicly identified haven't reached large audiences of real people, researchers say, and AI hasn't made them any more convincing — at least not yet.

"It's absolutely true that when you look at an individual campaign, it's just as likely as not that it hasn't had a huge amount of influence, which is why Russia just does it again and again, or in a different form, or targeting a different group," the Digital Forensic Research Lab's Carvin said. "It's almost like producing cheaply manufactured goods and just getting it out there in the world, hoping that maybe one particular gadget ends up becoming the popular toy of the season, even if the others completely fail."

Many researchers who study disinformation   warn against   seeing the hand of Russia as an all-powerful puppeteer, especially since so much of what its mouthpieces amplify is homegrown.

"Any potential narrative that's being argued in a given political environment is fodder for Russian operations — which in itself can sound a little crazy and conspiracy-ish," Carvin said. "And in some ways you risk creating a … situatio n where absolutely everything that's happening online is all Russia's fault."

But, he added, "at the same time, Russia has a lot of resources at its disposal and it's willing to experiment in different ways to see which things stick. … Why not try all of the above and see where it takes you?"


Tags

jrDiscussion - desc
[]
 
Hallux
PhD Principal
1  seeder  Hallux    one week ago

"If they can get more people on the internet arguing with each other or trusting each other less, then in some ways their job is done."

Seems to be working here ...

 
 
 
evilone
Professor Guide
1.1  evilone  replied to  Hallux @1    one week ago
Seems to be working here ...

I'm waiting for one of the useful idiots to seed from a fake Russian site.

 
 
 
Trout Giggles
Professor Principal
1.1.1  Trout Giggles  replied to  evilone @1.1    one week ago

I don;t think we;ll have long to wait

 
 
 
Sean Treacy
Professor Principal
1.1.2  Sean Treacy  replied to  evilone @1.1    one week ago
ng for one of the useful idiots to seed from a fake Russian site.

already been done, a couple times. Of course, the seed and seeder probably didn't fit the narrative you'd like.

Remember, the most successful operation of russia's  2016 campaign was an anti-Trump rally.  

 
 
 
Ronin2
Professor Quiet
2  Ronin2    one week ago

Russia doesn't need to do a damn thing.

Democrats and their sycophants in the media have already done it for them.

 
 
 
Sean Treacy
Professor Principal
2.1  Sean Treacy  replied to  Ronin2 @2    one week ago
Russia doesn't need to do a damn thing.

Yep, when threatened with exposure,  Biden has learned to cry "russian disinformation" and his lackeys in government and media will use it censor legitimate information. 

 
 
 
CB
Professor Principal
3  CB    one week ago

This is precisely one of the prime drivers of monies spent on dealing with BULLSHIT in this world we all live. Every thing that is good is countered by its equal and opposite BULLSHIT. Monies wasted in the millions, billions, and maybe even trillions on fighting back BULLSHIT. And relatively few seem to ever take stock of the "savings" that could be had if we, they, cut out the retributive BULLSHIT.

 
 
 
Nerm_L
Professor Expert
4  Nerm_L    one week ago

Does this mean Russians are more sophisticated and technology savvy than Americans?  We know that the two political parties would never, ever try to dupe voters (even though we've seen it on YouTube).  And law enforcement never, ever entice people to commit crimes.  And the intelligence agencies of the Federal government would never, ever lie to the American public.  And foreign governments would never, ever try to generate fear and outrage toward their enemies.

Yep, its only the Russians who are trying to divide the country and spook voters.  And New York really is selling the Brooklyn bridge.   

 
 

Who is online

Texan1211
CB
arkpdx


60 visitors