Close-up shot of a person using a smartphone. The individual is dressed in dark clothing, and only their hands and part of their phone are visible, set against a dark background.

The Misinformation Mitigation Toolbox: Dismantling the Digital Deceit

read time - icon

0 min read

Sep 04, 2024

It is no secret that the internet, a hub for innovation and connection, has also become a fertile ground for misinformation. From cleverly disguised clickbait to weaponized social media campaigns, untruths spread faster than a virtual wildfire. Every day, countless misleading posts flood our social feeds, challenging the fabric of democracy and directly affecting fields and industries, most notably public health.

Fighting misinformation is not an easy task. The ever-increasing polarization and greater sophistication of bad actors make it a formidable foe.1,2 These bad actors within the disinformation network can range from state-sponsored troll farms spreading propaganda to political operatives manipulating public opinion.3 Even well-meaning individuals can inadvertently share false or misleading information, contributing to the problem. It’s no surprise that the rise of generative AI has escalated the spread of disinformation and propaganda as never seen before.4  

But all hope is not lost. We must arm ourselves—not with pitchforks and torches, but with a far more potent weapon: evidence-based interventions.

A study by a global team of 30 misinformation researchers led by Dr. Anastasia Kozyreva offers a much-needed "toolbox" of strategies empowering individuals to cut through the noise.5 

Here's your guide to understanding and utilizing these tools effectively, categorized by their primary aim: influencing behaviors (nudges), boosting competencies (boosts), or directly targeting beliefs (refutation).

Some Notes Before We Dive In

The review evaluated 81 studies specifically focused on cognitive and behavioral changes at an individual level, excluding efforts that target misinformation at a systemic level like content moderation or algorithmic changes. The team also acknowledged that the evidence base leans heavily on research conducted in WEIRD (Western, Educated, Industrial, Rich, and Democratic) countries, but actively approached studies and experts in non-WEIRD countries as well. Aside from this, it’s important to note that while misinformation interventions aim to address universal challenges, they’re very culturally sensitive. In other words, the effectiveness of an intervention in one culture might not translate directly into another.

Additionally, the study highlights limitations in existing research. Few studies have explored the long-term impact of interventions, and the factors influencing their lasting effectiveness remain unclear. Furthermore, significant variations exist in how different studies measured intervention effectiveness. These variations, often stemming from differences in participant tasks and outcome measures, make direct comparisons challenging.

Here's where the research becomes truly exciting and applicable: moving from theory to practice. The insights gleaned from these studies aren't just academic; they offer real-world and evidence-based solutions to combat misinformation. Kozyreva et al.'s research goes a step further by identifying a practical "toolbox" of interventions consisting of three powerful strategies: nudges, boosts, and refutations. Each of these approaches offers unique ways to address the challenge of misinformation, providing individuals and organizations with concrete methods to improve information literacy and critical thinking.

Illustration titled 'Toolbox of Interventions Against Online Misinformation' with a Venn diagram of three overlapping circles. The left circle, 'Nudges or Prompts,' shows friction, social norms, and accuracy prompts. The right circle, 'Boosts or Educational Interventions,' lists lateral reading and media literacy tips. The bottom circle, 'Refutation Techniques,' includes debunking and rebuttals.

Toolbox of individual-level interventions against online misinformation (2024).

Let's delve into these tools and see how they can be used to combat misinformation.

Want to Alter Specific Behaviors? Consider Nudging

Manifestation of misinformation, like sharing fake news or vouching for untrue medical interventions, doesn’t always have to be mitigated with forceful interventions. Sometimes, the most effective approach is a gentle nudge in the right direction. Behavioral nudges are subtle cues that steer users away from sharing misinformation without explicitly restricting their choices.

Imagine scrolling through your social feed and encountering a juicy headline that seems too good (or bad) to be true. Before you hit that share button, a simple accuracy prompt pops up: “Before you share, take a moment to evaluate the accuracy of this headline.” This nudge plants a seed of doubt, encouraging you to pause and consider the information's credibility before potentially amplifying it by sharing.

Friction can also be your friend in the fight against misinformation. A seemingly innocuous message like “Want to read this before sharing?” can act as a speed bump, prompting users to take a moment and reflect on the content before blasting it out to their network.

A screenshot showing a message box with the text, "Want to read the article first?" alongside an icon of a face with glasses. Below the message, two options are visible: "Repost" and "Quote," each with corresponding icons. The background is dark, and the message is highlighted in blue.

In X, reposting a news article without clicking the link activates this prompt. This forces users to engage with the content directly, increasing the chances they will encounter credibility indicators or additional information before sharing.

Social norms are also powerful motivators of behavior. We humans crave social approval—and that extends online, too. A study by Andi & Akesson (2023) showed a simple message reminding users that “most responsible people think twice before sharing” significantly reduced the sharing of fake news by 5.1%, representing a 46.7% increase in participants choosing not to share because they identified the news as false or misleading.6 This highlights the power of descriptive norms in nudging behavior. By subtly reframing sharing as a mark of responsible citizenship in the digital world, the nudge prompts users to ditch the knee-jerk reaction from the clickbait and become more discerning information curators.

Looking to Improve Competencies? Consider Boosts

You can nudge nudge all you want, but what if you want to equip users with the tools to become self-sufficient truth-seekers? Here's where boosts and educational interventions come into play. These strategies focus on empowering individuals to gain competencies, which are skills and knowledge to critically evaluate, verify, and navigate the complex landscape of online information. This includes the ability to identify reliable sources, fact-check claims, and make informed judgments about the credibility of content they encounter.

One core strategy is lateral reading. This encourages users to consult multiple sources beyond the one initially encountered. Imagine a social media post with a shocking claim. A lateral reading boost might highlight related articles from established news outlets, prompting users to compare perspectives and gather a more comprehensive picture.

The image shows a Facebook post by Dr. Farrah Agustin-Bunch from February 2020, humorously sharing a conversation with her obstetrician about her age and health inventions. Below the post is an ultrasound image with a smiley face drawn inside a red circle, alongside technical scan details.

A controversial doctor shares an alleged conversation about her pregnancy and how she uses her “inventions.” Reading laterally means not taking this at face value and opening new tabs to confirm this claim.

Verification strategies are another key component of competency building. These interventions teach users how to identify reputable sources, check for factual inconsistencies within a text, and utilize fact-checking websites. Think of pop-up tutorials or short explainer videos embedded within social media platforms that demonstrate these techniques in action. Verification doesn't stop with text! Images can be misleading too. A simple reverse image search can also be a powerful verification tool.

When all else fails, there's still the power of media literacy tips. Engaging infographics or short quizzes on "How to Spot Fake News" can be a simple yet informative way to educate and boost critical thinking skills.

Challenging Beliefs? Utilize Refutation Techniques

Sometimes, a gentle nudge or a skill boost just won't cut it. We're talking about deeply ingrained beliefs or situations where users are actively consuming misinformation. Here's where refutation techniques come into play. These strategies directly address false claims and aim to replace them with factual information.

The foundation of this approach is debunking and rebuttals. When done effectively, these techniques can dismantle misinformation by clearly exposing its flaws and presenting accurate information in its place. However, a simple "this is wrong" approach might not always be enough. Studies show that simply repeating claims can inadvertently strengthen their perceived truth in people's minds (known as the illusory truth effect), even the ones that seem absolutely implausible.7

This is why effective refutations go beyond mere contradiction. Here's a four-step approach that leverages the power of debunking to make your message stick, whether you're designing formal interventions or engaging in everyday conversations about misinformation:8

  1. Lead with the correct information, especially if it's a clear counterpoint to the misinformation.
  2. Warn them about the misinformation exposure.
  3. Specify and explain the false claim's flaws, highlighting factual errors or logical fallacies.
  4. Reiterate the accurate information and provide links for further exploration. Reinforce the facts, solidifying their presence in the user's mind.

an infographic titled "Example of a Refutation" that explains how to effectively refute misinformation about climate change. It includes sections labeled Fact, Myth, Fallacy, and Fact, providing strategies for correcting false information, such as leading with clear facts, warning of myths, explaining fallacies, and reinforcing the facts repeatedly.

Example of a refutation by debunking. (The Debunking Handbook, 2020)

By following these principles, refutation techniques can be a powerful tool for dismantling false beliefs and promoting a more accurate understanding of the world.

What Happens When You Mix These Strategies?

Tackling misinformation is rarely a one-size-fits-all solution. Strategies can also combine all three tools in the misinformation toolbox. By integrating these methods, we leverage different aspects of information processing: nudges prime individuals to be on the lookout for misinformation, boosts develop critical thinking skills, and refutations provide factual alternatives. This multi-pronged approach capitalizes on the strengths of each technique, filling in gaps and creating a powerful defense against the spread of untruths. Here are a few examples of how these strategies can be combined for maximum impact.

Inoculation

Inoculation (also known as “prebunking”) borrows elements from both refutation techniques and boosts.  Imagine a media literacy campaign that functions like a mental vaccine. First, it educates the public on common tactics used in deepfake videos, such as manipulated facial expressions or mismatched audio (boosting their knowledge). Then, before a major election where deepfakes are anticipated to be a problem, the campaign "prebunks" potential falsehoods by showing examples of how a politician's words or actions might be falsified using these methods (refutation technique). This inoculation effect builds up users' resistance to future encounters with deepfakes, making them more likely to question and verify video content they see online.

Even games can be used as inoculation to understand misinformation better. For example, Go Viral!, a 5-minute game designed to protect users from COVID-19 misinformation, allows users to experience the creation and spread of misinformation firsthand. In the game, players navigate a social media feed filled with misinformation and outrage-evoking content, learning about common manipulation techniques like fearmongering, impersonation, and conspiracy theories. By actively engaging in these scenarios, players learn to recognize these tactics in real life. Research by Basol et al. (2021) showed this game to be particularly effective in enhancing resistance to COVID-19 misinformation. The game led to significant improvements in recognizing manipulative techniques and a reduction in willingness to share misinformation.9

The image shows a post by "Stephanie" at 07:18 a.m., claiming a "Big Pharma-owned company destroyed rainforest to fund a toilet paper monopoly" and profited from COVID-19. The post uses the hashtag "#ToiletPaperVsNature" and has 318 likes. Options to "Search for posts about #ToiletPaperVsNature" or "Completely ignore this because it's insane" are visible, along with a credibility meter and like counter at the top.

Snippet of the game, Go Viral!

Another inoculation game called Bad News had a similar premise but with a broader focus on various misinformation tactics. Players step into the shoes of a fake news maestro, manipulating audiences with six common tactics: impersonation, emotional exploitation, polarization, conspiracy theories, discrediting opponents, and trolling. The goal? To amass followers while maintaining a veneer of credibility. Studies about this game demonstrate that it is particularly effective in helping users identify strategies like impersonation and conspiracy theories. One experiment found a significant decrease in the perceived reliability of impersonation tactics, with average ratings dropping from 3.00 to 2.30 after playing the game.10

The image displays a message saying, "Hi there! Good to see you." followed by "My job is to guide you in your quest to becoming a disinformation and fake news tycoon." A small clickable text, "Oh boy!" is seen at the bottom. On the left, the user has "0 followers" and a "credibility" meter showing a low rating.

Snippet of the game Bad News.

By actively engaging in these scenarios, players as young as 14 learn to recognize these tactics in real life. These games simulate a social media environment, complete with likes and credibility points, mimicking the reinforcement mechanisms that often drive misinformation spread.

This gamified approach to inoculation against misinformation is effective because it combines interactive learning, immediate feedback, and real-world relevance. By "becoming the villain," players gain insight into the mechanics of misinformation, making them more resilient when encountering similar tactics online.

Source Credibility Labels

Source credibility labels combine the power of refutation with a nudge. These labels, often taking the form of website badges or social media indicators, highlight fact-checked content (refutation strategy). They essentially preempt the debunking of any potential misinformation on the site. At the same time, these labels nudge users towards reliable sources by visually highlighting their credibility. This one-two punch discourages users from engaging with dubious content and steers them towards trustworthy information.

The image shows three types of media content. On the left are search engine results for President Joe Biden, with images and links. The middle is a social media feed from Twitter/X showing a news post about China’s President Xi Jinping. On the right is a Facebook post from FiveThirtyEight about the Boston Red Sox's impressive start to the season.

NewsGuard has a browser extension that uses color-coded "shield" symbols to indicate source reliability. Green is for reliable, red is for unreliable, gray is for user-generated content, and gold is for satire. (Photo from this study.)

Warning and Fact-Checking Labels

Similar to source credibility labels, warning and fact-checking labels combine nudges with refutation. When encountering potentially misleading information, these labels might display warnings or summaries from fact-checking organizations (refutation strategy). This serves as a nudge, prompting users to pause and consider the information's veracity before sharing or accepting it as truth.

The image is a tweet from Michael Greger, M.D., stating that consuming as little as half an egg a day is associated with premature death, accompanied by a picture of a meal with eggs. Below, Twitter users have added context, clarifying that no correlation exists between eggs and cardiovascular risk, providing a link to a Harvard study and indicating that the physician's statement is misleading to promote his book.

Community notes in X can function as a form of warning by adding context to potentially misleading tweets.

Final words

The fight against misinformation is an ongoing battle, and the interventions here cover most of the up-to-date tools we can add to our arsenal. This equips you with the knowledge of different approaches to tackling misinformation. By strategically using nudges, boosts, and refutations, we can empower users and ourselves to become critical consumers of information, fostering a more informed and discerning online environment.

For behavioral scientists and practitioners applying this, the key lies in understanding the target audience and the specific type of misinformation being addressed. Tailoring interventions to the situation is crucial.  Sometimes, a gentle nudge in the right direction is all that's needed. Other times, an inoculation strategy might be necessary. Remember, the ultimate goal is to empower users, not control them. By equipping them with the knowledge and critical thinking skills they need, we can collectively dismantle digital deception and build a more informed online future.

If you want to learn more about Kozyreva, et al.'s research, the team created a website alongside their paper discussing the different interventions and the existing evidence related to them. If you’d like a further read, we at the Decision Lab also published a paper on the taxonomy of misinformation recently!

Are you passionate about combating misinformation through behavioral science? The Decision Lab is eager to collaborate with researchers and practitioners. Get in touch today to help us in our mission to create a more informed world!

References

  1. Geiger, A. (2021, April 9). Political polarization in the American public. Pew Research Center - U.S. Politics & Policy. https://www.pewresearch.org/politics/2014/06/12/political-polarization-in-the-american-public/
  2. Carothers, T., & O'Donohue, A. (2019, October 1). How to understand the global spread of political polarization. Carnegie Endowment for International Peace. https://carnegieendowment.org/posts/2019/10/how-to-understand-the-global-spread-of-political-polarization?lang=en 
  3. Ong, J. C., & Cabañes, J. V. (2018). Architects of Networked Disinformation. Newton Tech4Dev Network. https://newtontechfordev.com/wp-content/uploads/2018/02/ARCHITECTS-OF-NETWORKED-DISINFORMATION-FULL-REPORT.pdf 
  4. Ryan-Mosley, T. (2023, October 4). How generative AI is boosting the spread of disinformation and propaganda. MIT Technology Review. https://www.technologyreview.com/2023/10/04/1080801/generative-ai-boosting-disinformation-and-propaganda-freedom-house/
  5. Kozyreva, A., Lorenz-Spreen, P., Herzog, S.M. et al. Toolbox of individual-level interventions against online misinformation. Nat Hum Behav 8, 1044–1052 (2024). https://doi.org/10.1038/s41562-024-01881-0
  6. Andi, S., & Akesson, J. (2020). Nudging away false news: Evidence from a social norms experiment. Digital Journalism, 9(1), 106-125. https://doi.org/10.1080/21670811.2020.1847674 
  7. Fazio, L. K., Rand, D. G., & Pennycook, G. (2019). Repetition increases perceived truth equally for plausible and implausible statements. Psychonomic Bulletin & Review, 26(5), 1705-1710. https://doi.org/10.3758/s13423-019-01651-4 
  8. Lewandowsky, S. (2020). The debunking handbook 2020. https://skepticalscience.com/docs/DebunkingHandbook2020.pdf
  9. Basol, M., Roozenbeek, J., Berriche, M., Uenal, F., McClanahan, W. P., & Linden, S. V. (2021). Towards psychological herd immunity: Cross-cultural evidence for two prebunking interventions against COVID-19 misinformation. Big Data & Society, 8(1), 205395172110138. https://doi.org/10.1177/20539517211013868
  10. Roozenbeek, J., & Van der Linden, S. (2019). Fake news game confers psychological resistance against online misinformation. Palgrave Communications, 5(1). https://doi.org/10.1057/s41599-019-0279-9

About the Author

Woman smiling outdoors, wearing a mustard yellow shirt and flower earrings. She is surrounded by green trees and plants, under a clear sky.

Celestine Rosales

Celestine is a Junior Research Analyst at The Decision Lab. She is a researcher with a passion for understanding human behavior and using that knowledge to make a positive impact on the world. She is currently pursuing her Master's degree in Social Psychology, where she focuses on issues of social justice and morality. She also holds a Bachelor's degree in Psychology. Before joining TDL, Celestine worked as a UX Researcher at a conversion rate optimization company, where she collaborated with a variety of B2B and SaaS clients to help them improve their websites. She also participated in an all-women cohort of scholars trained to do data analytics. Outside of work, Celestine enjoys taking long walks, listening to podcasts, and trying new things.

About us

We are the leading applied research & innovation consultancy

Our insights are leveraged by the most ambitious organizations

Image

I was blown away with their application and translation of behavioral science into practice. They took a very complex ecosystem and created a series of interventions using an innovative mix of the latest research and creative client co-creation. I was so impressed at the final product they created, which was hugely comprehensive despite the large scope of the client being of the world's most far-reaching and best known consumer brands. I'm excited to see what we can create together in the future.

Heather McKee

BEHAVIORAL SCIENTIST

GLOBAL COFFEEHOUSE CHAIN PROJECT

OUR CLIENT SUCCESS

$0M

Annual Revenue Increase

By launching a behavioral science practice at the core of the organization, we helped one of the largest insurers in North America realize $30M increase in annual revenue.

0%

Increase in Monthly Users

By redesigning North America's first national digital platform for mental health, we achieved a 52% lift in monthly users and an 83% improvement on clinical assessment.

0%

Reduction In Design Time

By designing a new process and getting buy-in from the C-Suite team, we helped one of the largest smartphone manufacturers in the world reduce software design time by 75%.

0%

Reduction in Client Drop-Off

By implementing targeted nudges based on proactive interventions, we reduced drop-off rates for 450,000 clients belonging to USA's oldest debt consolidation organizations by 46%

Read Next

A group of people in a modern meeting room, some seated and working at a large table with laptops and cameras, while others stand and converse. The atmosphere is casual, with natural light filtering through large windows in the background.
Insight

Why Teams Make Bad Decisions

Sometimes, the best way to avoid group decision-making failures is not to make decisions as a group at all.

Notes illustration

Eager to learn about how behavioral science can help your organization?