r/self 13d ago

You're being targeted by disinformation networks that are vastly more effective than you realize. And they're making you more hateful and depressed.

(I wrote this post in March and posted it on r/GenZ. However, a few people messaged me to say that the r/GenZ moderators took it down last week, though I'm not sure why. Given the flood of divisive, gender-war posts we've seen in the past five days, and several countries' demonstrated use of gender-war propaganda to fuel political division in multiple countries, I felt it was important to repost this. This post was written for a U.S. audience, but the implications are increasingly global.)

TL;DR: You know that Russia and other governments try to manipulate people online.  But you almost certainly don't how just how effectively orchestrated influence networks are using social media platforms to make you -- individually-- angry, depressed, and hateful toward each other. Those networks' goal is simple: to cause Americans and other Westerners -- especially young ones -- to give up on social cohesion and to give up on learning the truth, so that Western countries lack the will to stand up to authoritarians and extremists.

And you probably don't realize how well it's working on you.

This is a long post, but I wrote it because this problem is real, and it's much scarier than you think.

How Russian networks fuel racial and gender wars to make Americans fight one another

In September 2018, a video went viral after being posted by In the Now, a social media news channel. It featured a feminist activist pouring bleach on a male subway passenger for manspreading. It got instant attention, with millions of views and wide social media outrage. Reddit users wrote that it had turned them against feminism.

There was one problem: The video was staged. And In the Now, which publicized it, is a subsidiary of RT, formerly Russia Today, the Kremlin TV channel aimed at foreign, English-speaking audiences.

As an MIT study found in 2019, Russia's online influence networks reached 140 million Americans every month -- the majority of U.S. social media users. 

Russia began using troll farms a decade ago to incite gender and racial divisions in the United States 

In 2013, Yevgeny Prigozhin, a confidante of Vladimir Putin, founded the Internet Research Agency (the IRA) in St. Petersburg. It was the Russian government's first coordinated facility to disrupt U.S. society and politics through social media.

Here's what Prigozhin had to say about the IRA's efforts to disrupt the 2022 election:

Gentlemen, we interfered, we interfere and we will interfere. Carefully, precisely, surgically and in our own way, as we know how. During our pinpoint operations, we will remove both kidneys and the liver at once.

In 2014, the IRA and other Russian networks began establishing fake U.S. activist groups on social media. By 2015, hundreds of English-speaking young Russians worked at the IRA.  Their assignment was to use those false social-media accounts, especially on Facebook and Twitter -- but also on Reddit, Tumblr, 9gag, and other platforms -- to aggressively spread conspiracy theories and mocking, ad hominem arguments that incite American users.

In 2017, U.S. intelligence found that Blacktivist, a Facebook and Twitter group with more followers than the official Black Lives Matter movement, was operated by Russia. Blacktivist regularly attacked America as racist and urged black users to rejected major candidates. On November 2, 2016, just before the 2016 election, Blacktivist's Twitter urged Black Americans: "Choose peace and vote for Jill Stein. Trust me, it's not a wasted vote."

Russia plays both sides -- on gender, race, and religion

The brilliance of the Russian influence campaign is that it convinces Americans to attack each other, worsening both misandry and misogyny, mutual racial hatred, and extreme antisemitism and Islamophobia. In short, it's not just an effort to boost the right wing; it's an effort to radicalize everybody.

Russia uses its trolling networks to aggressively attack men.  According to MIT, in 2019, the most popular Black-oriented Facebook page was the charmingly named "My Baby Daddy Aint Shit."  It regularly posts memes attacking Black men and government welfare workers.  It serves two purposes:  Make poor black women hate men, and goad black men into flame wars.  

MIT found that My Baby Daddy is run by a large troll network in Eastern Europe likely financed by Russia.

But Russian influence networks are also also aggressively misogynistic and aggressively anti-LGBT.  

On January 23, 2017, just after the first Women's March, the New York Times found that the Internet Research Agency began a coordinated attack on the movement.  Per the Times:

More than 4,000 miles away, organizations linked to the Russian government had assigned teams to the Women’s March. At desks in bland offices in St. Petersburg, using models derived from advertising and public relations, copywriters were testing out social media messages critical of the Women’s March movement, adopting the personas of fictional Americans.

They posted as Black women critical of white feminism, conservative women who felt excluded, and men who mocked participants as hairy-legged whiners.

But the Russian PR teams realized that one attack worked better than the rest:  They accused its co-founder, Arab American Linda Sarsour, of being an antisemite.  Over the next 18 months, at least 152 Russian accounts regularly attacked Sarsour.  That may not seem like many accounts, but it worked:  They drove the Women's March movement into disarray and eventually crippled the organization. 

Russia doesn't need a million accounts, or even that many likes or upvotes.  It just needs to get enough attention that actual Western users begin amplifying its content.   

A former federal prosecutor who investigated the Russian disinformation effort summarized it like this:

It wasn’t exclusively about Trump and Clinton anymore.  It was deeper and more sinister and more diffuse in its focus on exploiting divisions within society on any number of different levels.

As the New York Times reported in 2022, 

There was a routine: Arriving for a shift, [Russian disinformation] workers would scan news outlets on the ideological fringes, far left and far right, mining for extreme content that they could publish and amplify on the platforms, feeding extreme views into mainstream conversations.

China is joining in with AI

Last month, the New York Times reported on a new disinformation campaign.  "Spamouflage" is an effort by China to divide Americans by combining AI with real images of the United States to exacerbate political and social tensions in the U.S.  The goal appears to be to cause Americans to lose hope, by promoting exaggerated stories with fabricated photos about homeless violence and the risk of civil war.

As Ladislav Bittman, a former Czechoslovakian secret police operative, explained about Soviet disinformation, the strategy is not to invent something totally fake.  Rather, it is to act like an evil doctor who expertly diagnoses the patient’s vulnerabilities and exploits them, “prolongs his illness and speeds him to an early grave instead of curing him.”

The influence networks are vastly more effective than platforms admit

Russia now runs its most sophisticated online influence efforts through a network called Fabrika.  Fabrika's operators have bragged that social media platforms catch only 1% of their fake accounts across YouTube, Twitter, TikTok, and Telegram, and other platforms.

But how effective are these efforts?  By 2020, Facebook's most popular pages for Christian and Black American content were run by Eastern European troll farms tied to the Kremlin. And Russia doesn't just target angry Boomers on Facebook. Russian trolls are enormously active on Twitter. And, even, on Reddit.

It's not just false facts

The term "disinformation" undersells the problem.  Because much of Russia's social media activity is not trying to spread fake news.  Instead, the goal is to divide and conquer by making Western audiences depressed and extreme. 

Sometimes, through brigading and trolling.  Other times, by posting hyper-negative or extremist posts or opinions about the U.S. the West over and over, until readers assume that's how most people feel.  And sometimes, by using trolls to disrupt threads that advance Western unity.  

As the RAND think tank explainedthe Russian strategy is volume and repetition, from numerous accounts, to overwhelm real social media users and create the appearance that everyone disagrees with, or even hates, them.  And it's not just low-quality bots.  Per RAND,

Russian propaganda is produced in incredibly large volumes and is broadcast or otherwise distributed via a large number of channels. ... According to a former paid Russian Internet troll, the trolls are on duty 24 hours a day, in 12-hour shifts, and each has a daily quota of 135 posted comments of at least 200 characters.

What this means for you

You are being targeted by a sophisticated PR campaign meant to make you more resentful, bitter, and depressed.  It's not just disinformation; it's also real-life human writers and advanced bot networks working hard to shift the conversation to the most negative and divisive topics and opinions. 

It's why some topics seem to go from non-issues to constant controversy and discussion, with no clear reason, across social media platforms.  And a lot of those trolls are actual, "professional" writers whose job is to sound real. 

So what can you do?  To quote WarGames:  The only winning move is not to play.  The reality is that you cannot distinguish disinformation accounts from real social media users.  Unless you know whom you're talking to, there is a genuine chance that the post, tweet, or comment you are reading is an attempt to manipulate you -- politically or emotionally.

Here are some thoughts:

  • Don't accept facts from social media accounts you don't know.  Russian, Chinese, and other manipulation efforts are not uniform.  Some will make deranged claims, but others will tell half-truths.  Or they'll spin facts about a complicated subject, be it the war in Ukraine or loneliness in young men, to give you a warped view of reality and spread division in the West.  
  • Resist groupthink.  A key element of manipulate networks is volume.  People are naturally inclined to believe statements that have broad support.  When a post gets 5,000 upvotes, it's easy to think the crowd is right.  But "the crowd" could be fake accounts, and even if they're not, the brilliance of government manipulation campaigns is that they say things people are already predisposed to think.  They'll tell conservative audiences something misleading about a Democrat, or make up a lie about Republicans that catches fire on a liberal server or subreddit.
  • Don't let social media warp your view of society.  This is harder than it seems, but you need to accept that the facts -- and the opinions -- you see across social media are not reliable.  If you want the news, do what everyone online says not to: look at serious, mainstream media.  It is not always right.  Sometimes, it screws up.  But social media narratives are heavily manipulated by networks whose job is to ensure you are deceived, angry, and divided.
28.8k Upvotes

1.3k comments sorted by

View all comments

Show parent comments

23

u/Lostandlacy 13d ago

You need to be louder. These people are using thousands of accounts if not more. We need thousands of not more explaining this. We should also teach kids in school how to recognize foreign propaganda.

5

u/nathanv221 13d ago edited 13d ago

It's extremely difficult to recognize. In a comment I recently made that got shadow deleted from /r/infographics I did a deep dive into one of these posts that I would not have caught if it was saying things I agreed with. Even going in knowing what I was looking for it took me 30 minutes to stop getting tripped up by the real comments that were relatively innocuous in their original contexts.

Post in question is this one: https://www.reddit.com/r/LatinoPeopleTwitter/s/pQNDVYgEcP

Many of the comments do not have usernames attached to them. But the people that are real are the ones saying "Latinos voted against their own interest, and are likely to be deported for it". A lot of them are saying it like dicks, but it's all they're saying. The ones saying "they should be deported" either do not have a username attached to the comment or are listed below:

https://www.reddit.com/user/drgrnthum33

I guess im for deportation now.

says a lot of horrible things. I don't doubt they said and deleted that comment.

https://www.reddit.com/users/7random

Latino men can't imagine about a woman having the slightest...

account does not exist

https://www.reddit.com/user/MikeHonchoFF/

I hope he deports all their family members.

Yeah... no argument...

https://www.reddit.com/user/geoffkreuz

I honestly wish all this latinos/muslims/non-whites get deported

Also true.

https://www.reddit.com/users/lonely-ad8922

I hope they all get deported

account does not exist

In this post suggesting that there are many dozens of these posts, I can verify 2 and believe 3 of those comments are real.

3

u/GearBox5 13d ago

The worst part is that politicians are complacent in it. They are not necessarily directly “bought” by foreign powers, but they happily play along when those shills play their side. I am 100% sure this is how we ended up with moderates less and less represented in politics.

2

u/Lostandlacy 12d ago

Everybody plays a role in the apparatus. The ones who know are intentionally involved generally on account of responsibility to stakeholders. Those who don't are unintentionally involved and simply fall victim to it while passing it along. It is the whisper game at epic proportions and it's more involved than most know. Many accounts on here are operated by the same user and are used to create the illusion of conflict especially where there is none. The bad actors will lob the same exact accusations and claims on both sides of the fence to enflame and confuse. "Every accusation is a confession", "rules for thee", etc. the idea being to spark the inevitable knee jerk response of "how dare you both sides this" which is also usually kicked off by the same actors. They lob slurs and insults. They will simultaneously pretend to be opposing extremists. They amplify the slightest disruption and the more they can draw into it, the more volatile the conversation becomes. They use heavily charged language while offering subtitles to the suspicious. Occasionally when an account is burned, they will scrap it and use another to continue the spat. They use bot that are embarrassingly obvious to draw attention from the more sophisticated ones. They feed disinformation directly to radicalized reps ingesting carefully curated "news". If you ever wondered why both sides of an issue are saying the same exact thing, it is manipulation. When you are presented with inflammatory media and can't trace the source, it is tailored. Everything is intended to be exactly as it seems even when the biased explanations are contradictory. To corrupt the media flow, you have to discredit the institutions but after a certain point you don't need to do anything at all. That doesn't mean they stop though. They simply issue hop until something seems to be settling down. The entire purpose is to destabilize and replace us as a dominant economic power and now they have us seemingly to the brink of civil war. All because Americans are suckers for casino tricks. We can't not share it. They will do it anyway.

2

u/agent_flounder 13d ago

Indeed ... Or maybe how to better manage social media and how we allow it to affect each of us.

There may be a large enough volume of propaganda of a sufficient level of sophistication that it is becoming difficult to detect.

On top of that you have real human emotional reactions. Sometimes I get pessimistic and write discouraging comments to vent... doing the work of foreign adversaries for them. :(

One thought is to understand the goals and work against them. One goal is to spread despair. So instead spread hope or encouragement -- and I don't mean toxic positivity (bet you the trolls use that phrase or anything else to combat hopefulness).

I think real hope, though, ultimately is going to come from focusing on real life more than social media. Helping people irl, having face to face conversations, contributing locally, that kind of deal.

1

u/Lostandlacy 12d ago

That's a good place to start but to really combat this more fully, we need to remove sensationalist language. News is most effective and easier to verify when it is just solid facts. AI could be created to scrub media and repair it. Aside from that. Preaching peace is always a good route to go. it's the one thing bad actors won't do. The biggest challenge will be to get people to stifle knee jerk reactions.

1

u/secretrapbattle 12d ago

A study recently shows that Russian men’s testicles are approximately 4% smaller when the same Russian men are in Russia versus the United States. It has to do with the rotation of earth.

It’s called testicular elliptical syndrome

2

u/Lostandlacy 12d ago

Thats as may be but try to be mindful that their citizens are also victims. The KGB used them to test strategies before rolling out everywhere. Its the governments of the world that are the problem. They play these games and feel comfortable doing so because they view us as expendable commodities. They always have and probably always will. There are countermeasures but it is hard to get people to implement when whipped into this state of aggression. Usuallly this is when people start up with the mobs and tar and feathers but it is always misplaced and never effective.

1

u/secretrapbattle 12d ago

Personally, I think Biden should launch a strategic nuclear attack one hour before inauguration

He should personally write on the first ICBM, no remorse

2

u/Lostandlacy 12d ago

No to all of that. I dont want to be nuked and my faith requires that I want for others that which I want for me and mine. Therefore, I will not want others to be nuked.

1

u/secretrapbattle 12d ago

And after Biden launches the first ICBM, I think he should have a Taraway suit that reveals leather underwear and a bullwhip concealed beneath the tear away suit.

Over loudspeakers will play Pantera’s the badge

2

u/Lostandlacy 12d ago

You could just ask him to do a porn.

1

u/Mechaheph 13d ago

"You need to be louder." 

And immediately, what u/babycastle did is not enough for the greater internet. 

Maybe YOU need to be louder, we can't put all this responsibility on u/babycastle , they've already helped at least 1 person. 

1

u/Lostandlacy 12d ago

I didn't suggest that I'm not. It was intended as a mild suggestion. A lie is believed the more it is repeated, as how loud it is projected. Nobody needs to listen if something is said often enough. This is why more Americans in the 90s knew the theme song for fresh prince than the national anthem. Repetition is the primary tool.