Featured Product
This Week in Quality Digest Live
Quality Insider Features
Eryn Brown
Their prospects for surviving the pandemic may seem dim, but there are some encouraging signs, experts say
John Toon
Detailed 3D data could help improve reliability and performance
Michael Mallen
Understanding the impact of cost of quality generally leads to better bottom-line performance
Sharona Hoffman
Here’s how to promote algorithmic fairness
Matt Fieldman
A new Manufacturing Extension Partnership grant will fund a major push in manufacturing innovation

More Features

Quality Insider News
ProMation announces additional options for constructing motor-operated valves for industrial flow control
Simplifies workflows and enhances ease of use for manufacturing, assembly, and construction
Harnessing the forces that drive your organizations success
Use X-Rite spectrophotometers to render digital material file for 3D design, production, and quality control teams
IDS cameras help inspect up to 120 vials per minute for dimensional accuracy or surface condition
Manufacturing and supply-chain experts share best practices and predictions
Marposs can now offer complete range of solutions for functional testing and end-of-line testing of any type of electric motor
Enables professionals and small businesses to improve product development, shorten time to market, reduce development costs
Free education source for global medical device community

More News

Quality Insider

How Online Misinformation Spreads

Misinformation is running rampant. To slow this infodemic, researchers are tracking how it spreads on social media.

Published: Thursday, April 1, 2021 - 12:02

This story was originally published by Knowable Magazine.

You may have heard the outlandish claim: Bill Gates is using the Covid-19 vaccine to implant microchips in everyone so he can track where they go. It’s false, of course, debunked repeatedly by journalists and fact-checking organizations. Yet the falsehood persists online—in fact, it’s one of the most popular conspiracy theories making the rounds. In May 2020, a Yahoo/YouGov poll found that 44 percent of Republicans (and 19 percent of Democrats) said they believed it.

This particular example is just a small part of what the World Health Organization now calls an infodemic—an unprecedented glut of information that may be misleading or false. Misinformation—false or inaccurate information of all kinds, from honest mistakes to conspiracy theories—and its more intentional subset, disinformation, are both thriving, fueled by a once-in-a-generation pandemic, extreme political polarization, and a brave new world of social media.

“The scale of reach that you have with social media, and the speed at which it spreads, is really like nothing humanity has ever seen,” says Jevin West, a disinformation researcher at the University of Washington.

One reason for this reach is that so many people are participants on social media. “Social media is the first type of mass communication that allows regular users to produce and share information,” says Ekaterina Zhuravskaya, an economist at the Paris School of Economics who co-authored an article on the political effects of the Internet and social media in the 2020 Annual Review of Economics.

Trying to stamp out online misinformation is like chasing an elusive and ever-changing quarry, researchers are learning. False tales—often intertwined with elements of truth—spread like a contagion across the Internet. They also evolve over time, mutating into more infectious strains, fanning across social media networks via constantly changing pathways and hopping from one platform to the next.

Misinformation doesn’t simply diffuse like a drop of ink in water, says Neil Johnson, a physicist at George Washington University who studies misinformation. “It’s something different. It kind of has a life of its own.”


Bill Gates (left) meeting with Francis Collins (center), director of the National Institutes of Health, and Anthony Fauci (right), director of the National Institute of Allergy and Infectious Diseases, in 2017. Gates is one of many victims of online misinformation: In early 2020, he made a casual comment about better ways to keep electronic records of individuals’ vaccine history. Social-media users twisted this into a false conspiracy theory in which Gates and others would use the Covid-19 vaccine to implant microchips in everyone to remotely track their movements. Credit: National Institutes of Health

The Gates fiction is a case in point. On March 18, 2020, Gates mentioned in an online forum on Reddit that electronic records of individuals’ vaccine history could be a better way to keep track of who had received the Covid-19 vaccine than paper documents, which can be lost or damaged. The very next day, a website called Biohackerinfo.com posted an article claiming that Gates wanted to implant devices into people to record their vaccination history. Another day later, a YouTube video expanded that narrative, explicitly claiming that Gates wanted to track people’s movements. That video was viewed nearly two million times. In April, former Donald Trump advisor Roger Stone repeated the conspiracy on a radio program, which was then covered in the New York Post. Fox News host Laura Ingraham also referred to Gates’ intent to track people during an interview with then U.S. Attorney General William Barr.

But though it’s tempting to think from examples like this that websites like Biohackerinfo.com are the ultimate sources of most online misinformation, research suggests that’s not so. Even when such websites churn out misleading or false articles, they are often pushing what people have already been posting online, says Renee DiResta, a disinformation researcher at the Stanford Internet Observatory. Indeed, almost immediately after Gates wrote about digital certificates, Reddit users started commenting about implantable microchips, which Gates had never mentioned.

In fact, research suggests that malicious websites, bots, and trolls make up a relatively small portion of the misinformation ecosystem. Instead, most misinformation emerges from regular people, and the biggest purveyors and amplifiers of misinformation are a handful of human super-spreaders. For example, a study of Twitter during the 2016 election found that in a sample of more than 16,000 users, 6 percent of those who shared political news also shared misinformation. But the vast majority—80 percent—of the misinformation came from just 0.1 percent of users. Misinformation is amplified even more when those super-spreaders, such as media personalities and politicians like Donald Trump (until his banning by Twitter and other sites), have access to millions of people on social and traditional media.

Thanks to such super-spreaders, misinformation spreads in a way that resembles an epidemic. In a recent study, researchers analyzed the rise in the number of people engaging with Covid-19-related topics on Twitter, Reddit, YouTube, Instagram, and a right-leaning network called Gab. Fitting epidemiological models to the data, they calculated R-values that, in epidemiology, represent the average number of people a sick person would infect. In this case, the R-values describe the contagiousness of Covid-19-related topics in social media platforms, and though the R-value differed depending on the platform, it was always greater than one, indicating exponential growth and, possibly, an infodemic.

Differences in how information spreads depend on features of the particular platform and its users, not on the reliability of the information itself, says Walter Quattrociocchi, a data scientist at the University of Rome. He and his colleagues analyzed posts and reactions—such as likes and comments—about content from both reliable and unreliable websites, the latter being those that show extreme bias and promote conspiracies, as determined by independent fact-checking organizations. The number of posts and reactions regarding both types of content grew at the same rate, they found.

Complicating matters more, misinformation almost always contains kernels of truth. For example, the implantable microchips in the Gates conspiracy can be traced to a Gates Foundation-funded paper published in 2019 by MIT researchers, who designed technology to record someone’s vaccination history in the skin like a tattoo. The tattoo ink would consist of tiny semiconductor particles called quantum dots, whose glow could be read with a modified smartphone. There are no microchips, and the quantum dots can’t be tracked or read remotely. Yet the notion of implanting something to track vaccination status has been discussed. “It isn’t outlandish,” Johnson says. “It’s just outlandish to say it will then be used by Gates in some sinister way.”

What happens, Johnson explains, is that people pick nuggets of fact and stitch them together into a false or misleading narrative that fits their own worldview. These narratives then become reinforced in online communities that foster trust and thus lend credibility to misinformation.

Johnson and his colleagues track online discussion topics in social media posts. Using machine learning, their software automatically infers topics—say, vaccine side effects—from patterns in how words are used together. It’s similar to eavesdropping on multiple conversations by picking out particular words that signal what people are talking about, Johnson says.

And as in conversations, topics can evolve over time. Last year, for example, a discussion about the March 2020 lockdowns mutated to include the U.S. presidential election and QAnon conspiracy theories, according to Johnson. The researchers are trying to characterize such topic shifts, and what makes certain topics more evolutionarily fit and infectious.

Some broad narratives are especially tenacious. For example, Johnson says, the Gates microchip conspiracy contains enough truth to lend it credibility but also is often dismissed as absurd by mainstream voices, which feeds into believers’ distrust of the establishment. Throw in well-intentioned parents who are skeptical of vaccines, and you have a particularly persistent narrative. Details may differ, with some versions involving 5G wireless networks or radiofrequency ID tags, but the overall story—that powerful individuals want to track people with vaccines—endures.

And in online networks, these narratives can spread especially far. Johnson focuses on online groups, like public Facebook pages, some of which can include a million users. The researchers have mapped how these groups—within and across Facebook and five other platforms, Instagram, Telegram, Gab, 4Chan, and a predominantly Russian-language platform called VKontakte—connect to one another with weblinks, where a user in one online group links to a page on another platform. In this way, groups form clusters that also link to other clusters. The connections can break and relink elsewhere, creating complex and changing pathways through which information can flow and spread. For example, Johnson says, earlier forms of the Gates conspiracy were brewing on Gab only to jump over to Facebook and merge with more mainstream discussions about Covid-19 vaccinations.

These cross-platform links mean that the efforts of social media companies to take down election- or Covid-19-related misinformation are only partly effective. “Good for them for doing that,” Johnson says. “But it’s not going to get rid of the problem.” The stricter policies of some platforms—Facebook, for example—won’t stop misinformation from spilling over to a platform where regulations are more relaxed.


Researchers compared how often users of several social networks reacted to posts (e.g., liked, shared, or commented) from unreliable vs. reliable sources. Users of the predominantly right-wing network Gab reacted to unreliable information 3.9 times as often as reliable information, on average. In contrast, YouTube users mostly engaged with reliable information, while users of Reddit and Twitter fell in between. The way that misinformation spreads online seems to depend largely on the characteristics of a network and its users, the researchers say. (X axis shows the ratio of reactions to information from unreliable vs. reliable sources.)

And unless the entire social media landscape is somehow regulated, he adds, misinformation will simply congregate in more hospitable platforms. After companies like Facebook and Twitter started cracking down on election misinformation—even shutting down Trump’s accounts—many of his supporters migrated to platforms like Parler, which are more loosely policed.

To Johnson’s mind, the best way to contain misinformation may be by targeting these inter-platform links, instead of chasing down every article, meme, account, or even page that peddles in misinformation—which is ultimately a futile game of Whac-A-Mole. To show this, he and his colleagues calculated a different R-value. As before, their revised R-value describes the contagiousness of a topic, but it also incorporates the effects of dynamic connections in the underlying social networks. Their analysis isn’t yet peer-reviewed, but if it holds up, the formula can provide a mathematical way of understanding how a topic might spread—and, if that topic is rife with misinformation, how society can contain it.

For example, this new R-value suggests that by taking down cross-platform weblinks, social media companies or regulators can slow the transmission of misinformation so that it no longer spreads exponentially. Once regulators identify an online group brimming with misinformation, they can then remove links to other platforms. This needs to be the priority, Johnson says, even more than removing the group pages themselves.

Fact-checking may also help, as some studies suggest it can change minds and even discourage people from sharing misinformation. But the impact of a fact-check is limited because corrections usually don’t spread as far or as fast as the original misinformation, West says. “Once you get something rolling, it’s real hard to catch up.” And people may not even read a fact-check if it doesn’t conform to their worldview, Quattrociocchi says.

Other approaches, such as improving education and media literacy, and reforming the business model of journalism to prioritize quality over clicks, are all important for controlling misinformation—and, ideally, for preventing conspiracy theories from taking hold in the first place. But misinformation will always exist, and no single action will solve the problem, DiResta says. “It’s more of a problem to be managed like a chronic disease,” she says. “It’s not something you’re going to cure.”

This article is part of Reset: The Science of Crisis & Recovery, an ongoing series exploring how the world is navigating the coronavirus pandemic, its consequences, and the way forward. Reset is supported by a grant from the Alfred P. Sloan Foundation.

This article originally appeared Feb. 11, 2021, in Knowable Magazine, an independent journalistic endeavor from Annual Reviews. Sign up for the newsletter.


Discuss

About The Authors

Marcus Woo’s picture

Marcus Woo

Marcus Woo is a freelance science journalist based in the San Francisco Bay Area. He is not a bot, so follow him @sucramoow.

Knowable Magazine’s picture

Knowable Magazine

Knowable Magazine is an independent journalistic endeavor from Annual Reviews, a nonprofit publisher dedicated to synthesizing and integrating knowledge for the progress of science and the benefit of society. Sign up for Knowable Magazine’s newsletter.

Comments

Fake news is a real menace

This is an important article because a functioning democracy relies on an informed electorate, as opposed to people acting on rumors and disinformation. Fake news can spread very rapidly on social media but can and should also be challenged on social media; the latter was not possible when print newspapers had a monopoly on the propagation of news. The Hearst and Pulitzer papers published inflammatory anti-Spanish material that helped draw the U.S. into the Spanish-American War (1898) and one-sided coverage of the First World War got us into that one as well. The role of fake news in causing wars shows how dangerous it is. I think some of the conspiracy theorists are watching too many James Bond movies and believing they are real, e.g. the "space laser" was featured in Diamonds Are Forever, and the implantable tracking nanochips mentioned in this article were in Spectre.

The idea of using an R value to treat the propagation of fake news is very innovative and important, because it is an indicator of how rapidly rumors spread.

It's important to fact-check social media stories that seem too outrageous to be true before sharing them, and also to correct misinformation so the original poster can remove or correct the story.

No need for implanting chips

Bill Gates doesn't need to implant chips in anybody.  We all carry them around with us all the time - our phones.