Russians brag that only 1 per cent of fake social media profiles are caught, leak shows

Save articles for later

Add articles to your saved list and come back to them any time.

San Francisco: The Russian government has become far more successful at manipulating social media and search engine rankings than previously known, boosting lies about Ukraine’s military and the side effects of vaccines with hundreds of thousands of fake online accounts, according to documents recently leaked on the chat app Discord.

A World War II-era T-34 Soviet tank rehearses for a Victory Day military parade. Russian patriotism and nationalism is a powerful brew.Credit: Alexander Zemlianichenko

The Russian operators of those accounts boast that they are detected by social networks only about 1 per cent of the time, one document says.

That claim, described here for the first time, drew alarm from former government officials and experts inside and outside social media companies contacted for this article.

“Google and Meta and others are trying to stop this, and Russia is trying to get better. The figure that you are citing suggests that Russia is winning,” said Thomas Rid, a disinformation scholar and professor at Johns Hopkins University’s School of Advanced International Studies.

The undated analysis of Russia’s effectiveness at boosting propaganda on Twitter, YouTube, TikTok, Telegram and other social media platforms cites activity in late 2022 and was apparently presented to US military leaders in recent months. It is part of a trove of documents circulated in a Discord chatroom and obtained by The Washington Post.

The building in St Petersburg where Russian trolls worked to interfere in the 2016 US election.Credit: AP

Air National Guard technician Jack Teixeira was charged on Friday with taking and transmitting the classified papers, charges for which he faces 15 years in prison.

Many of the 10 current and former intelligence and tech safety specialists interviewed for this article cautioned that the Russian agency whose claims helped form the basis for the leaked document may have exaggerated its success rate.

But even if Russia’s fake accounts escaped detection only 90 per cent of the time instead of 99 per cent, that would indicate Russia has become far more proficient at disseminating its views to unknowing consumers than in 2016 when it combined bot accounts with human propagandists and hacking to try to influence the course of the US presidential election, the experts said.

“If I were the US government, I would be taking this seriously but calmly,” said Ciaran Martin, former head of the United Kingdom’s cyber defence agency. “I would be talking to the major platforms and saying ‘let’s have a look at this together to see what credence to give these claims.’”

“Don’t automatically equate activity with impact,” Martin said.

The Department of Defence declined to comment. TikTok, Twitter and Telegram, all named in the document as targets of Russian information operations, did not respond to a request for comment.

In a statement, YouTube owner Google said, “We have a strong track record detecting and taking action against botnets. We are constantly monitoring and updating our safeguards.”

With the average internet user spending more than two hours daily on social media, the internet has become perhaps the leading venue for conversations on current events, culture and politics, raising the importance of influencing what is seen and said online.

But little is known about how a specific piece of content gets shown to users. The big tech companies are secretive about the algorithms that drive their sites, while marketing companies and governments use influencers and automated tools to push messages of all kinds.

The possible presence of disguised propaganda has evoked widespread concern in recent months about TikTok, whose Chinese ownership has prompted proposed bans in Congress, and Twitter, whose former trust and safety chief Yoel Roth told Congress in February that the site still harboured thousands or hundreds of thousands of Russian bots.

The document offers a rare candid assessment by US intelligence of Russian disinformation operations. The document indicates it was prepared by the Joint Chiefs of Staff, US Cyber Command and Europe Command, the organisation that directs American military activities in Europe. It refers to signals intelligence, which includes eavesdropping, but does not cite sources for its conclusions.

It focuses on Russia’s Main Scientific Research Computing Centre, also referred to as GlavNIVTs. The centre performs work directly for the Russian presidential administration. It said the Russian network for running its disinformation campaign is known as Fabrika.

The centre was working in late 2022 to improve the Fabrika network further, the analysis says, concluding that “The efforts will likely enhance Moscow’s ability to control its domestic information environment and promote pro-Russian narratives abroad.”

The analysis said Fabrika was succeeding even though Western sanctions against Russia and Russia’s own censorship of social media platforms inside the country had added difficulties.

“Bots view, ‘like,’ subscribe and repost content and manipulate view counts to move content up in search results and recommendation lists,” the summary says. It adds that in other cases, Fabrika sends content directly to ordinary and unsuspecting users after gleaning their details such as email addresses and phone numbers from databases.

The intelligence document says the Russian influence campaigns’ goals included demoralising Ukrainians and exploiting divisions among Western allies.

After Russia’s 2016 efforts to interfere in the US presidential election, social media companies stepped up their attempts to verify users, including through phone numbers. Russia responded, in at least one case, by buying SIM cards in bulk, which worked until companies spotted the pattern, employees said. The Russians have now turned to front companies that can acquire less detectable phone numbers, the document says.

A separate top secret document from the same Discord trove summarised six specific influence campaigns that were operational or planned for later this year by a new Russian organisation, the Centre for Special Operations in Cyberspace. The new group is mainly targeting Ukraine’s regional allies, that document said.

Those campaigns included one designed to spread the idea that US officials were hiding vaccine side effects, intended to stoke divisions in the West. Another campaign claimed that Ukraine’s Azov Brigade was acting punitively in the country’s eastern Donbas region.

Others, aimed at specific countries in the region, push the idea that Latvia, Lithuania and Poland want to send Ukrainian refugees back to fight; that Ukraine’s security service is recruiting UN employees to spy; and that Ukraine is using influence operations against Europe with help from NATO.

A final campaign is intended to reveal the identities of Ukraine’s information warriors – the people on the opposite side of a deepening propaganda war.

Washington Post

Get a note directly from our foreign correspondents on what’s making headlines around the world. Sign up for the weekly What in the World newsletter here.

Most Viewed in World

From our partners

Source: Read Full Article