Information warfare abounds, and everyone online has been drafted whether they know it or not. Disinformation is deliberately generated, misleading content disseminated for selfish or malicious purposes.
Unlike misinformation, which may be shared unwittingly or with good intentions, disinformation aims to foment distrust, destabilize institutions, discredit good intentions, defame opponents, and delegitimize sources of knowledge such as science and journalism.
Many governments engage in disinformation campaigns.
For instance, Moscow has used images of celebrities to attract attention to anti-Ukraine propaganda. Meta, the parent company of Facebook and Instagram, warned last month that China has stepped up its disinformation operations.
Foreign governments, internet trolls, domestic and international extremists, opportunistic profiteers, and even paid disinformation agencies exploit the internet to spread questionable content.
Periods of civil unrest, natural disasters, health crises, and wars trigger anxiety and the hunt for information, which disinformation agents take advantage of. Certainly, it is worth watching for the warning signs of misinformation and dangerous speech.
But there are additional ploys used by these agents.
Hahaganda is a tactic in which disinformation agents use memes, political comedy from state-run outlets, or speeches to make light of serious matters, attack others, minimize violence or dehumanize, and deflect blame.
This approach provides an easy defense. If challenged, they say, “Can’t you take a joke?” often followed by accusations of being too politically correct.
Rumor-milling is a tactic in which disinformation agents claim to have exclusive access to secrets they allege are being purposefully concealed. They indicate that you will “only hear this here” and will imply that others are unwilling to share the alleged truth.
“The media won’t report this” or “The government doesn’t want you to know” and “I shouldn’t be telling you this …”
But they do not insist that the information be kept secret, and will instead include encouragement to share it, such as “most people won’t have the courage to share this.”
Again, it’s important to question how an author or speaker could have come by such “secret” information and what their motive is to prompt you to share it.
Often there is no real evidence, so instead disinformation agents will find or make up people to support their assertions. This impersonation can take multiple forms by using anecdotes as evidence, especially sympathetic stories from vulnerable groups such as women or children.
Similarly, they may disseminate the “concerned citizens” perspective.
These layperson experts present their social identity as providing the authority to speak on a matter; “As a mother …,” “As a veteran …,” “As a police officer ….” Convert communicators, or people who allegedly change from the “wrong” position to the “right” one, can be especially persuasive, such as the woman who got an abortion but regretted it. These people often don’t exist or may be coerced or paid.
If ordinary people don’t suffice, fake experts may be used. Some are fabricated, and you can watch out for “inauthentic user” behavior, for example, by checking X – formerly Twitter – accounts using the Botometer. But fake experts can come in different varieties.
- A faux expert is someone used for their title but doesn’t have actual relevant expertise.
- A pseudo-expert is someone who claims relevant expertise but has no actual training.
- A junk expert is a sellout. They may have had expertise once but now say whatever is profitable.
- An echo expert is when disinformation sources cite each other to provide credence for their claims. China and Russia routinely cite one another’s newspapers.
- A stolen expert is someone who exists, but they weren’t contacted and their research is misinterpreted.
- Likewise, disinformation agents also steal credibility from known news sources, such as by typosquatting, the practice of setting up a domain name that resembles a legitimate organization.
You can check whether accounts, anecdotal or scientific, have been verified by other reliable sources. Google the name. Check expertise status, source validity, and interpretation of research. Remember, one story or interpretation is not necessarily representative.
Conspiratorial narratives involve some malevolent force – for example, “the deep state,” – engaged in covert actions to cause harm to society. That certain conspiracies such as MK-Ultra and Watergate have been confirmed is often offered as evidence for the validity of new unfounded conspiracies.
Nonetheless, disinformation agents find that constructing a conspiracy is an effective means to remind people of past reasons to distrust governments, scientists, or other trustworthy sources.
But extraordinary claims require extraordinary evidence.
Remember, the conspiracies that were ultimately unveiled had evidence – often from sources like investigative journalists, scientists, and government investigations.
Be particularly wary of conspiracies that try to delegitimize knowledge-producing institutions like universities, research labs, government agencies, and news outlets by claiming that they are in on a cover-up.
Disinformation often serves the dual purpose of making the originator look good and their opponents look bad.
Disinformation takes this further by painting issues as a battle between good and evil, using accusations of evilness to legitimize violence. Russia is particularly fond of accusing others of being secret Nazis, pedophiles or Satanists.
Meanwhile, they often depict their soldiers as helping children and the elderly.
Be especially wary of accusations of atrocities like genocide, especially under the attention-grabbing “breaking news” headline. Accusations abound. Verify the facts and how the information was obtained.
A false narrative sets up the reader to believe that they have one of two mutually exclusive options – a good or a bad one, a red pill or a blue pill. You can accept their version of reality or be an idiot or “sheeple.”
There are always more options than those being presented, and issues are rarely so black and white. This is just one of the tactics in brigading, where disinformation agents seek to silence dissenting viewpoints by casting them as the wrong choice.
Whataboutism is a classic Russian disinformation technique they use to deflect attention from their wrongdoings. These allegations about the actions of others may be true or false but are nonetheless irrelevant to the matter at hand.
The potential past wrongs of one group do not mean you should ignore the current wrongs of another.
Disinformation agents also often cast their group as the wronged party.
They only engage in disinformation because their “enemy” engages in disinformation against them – they only attack to defend, and their reaction was appropriate, while that of others was an overreaction.
This type of competitive victimhood is particularly pervasive when groups have been embedded in a long-lasting conflict.
In all of these cases, the disinformation agent is aware that they are deflecting, misleading, trolling, or outright fabricating. If you don’t believe them, they at least want to make you question what, if anything, you can believe.
You often look into the things you buy rather than taking the advertising at face value before you hand over your money. This should also go for what information you buy into.
The views and opinions expressed in this article are those of the author and do not necessarily reflect the official policy of China Factor.