On Twitter, bots spread conspiracy theories and QAnon talking points

4 mins read

Emilio Ferrara, USC Annenberg School for Communication and Journalism

Americans who seek political insight and information on Twitter should know how much of what they are seeing is the result of automated propaganda campaigns.

Nearly four years after my collaborators and I revealed how automated Twitter accounts were distorting online election discussions in 2016, the situation appears to be no better. That’s despite the efforts of policymakers, technology companies and even the public to root out disinformation campaigns on social media.

In our latest study, we collected 240 million election-related tweets mentioning presidential candidates and election-related keywords, posted between June 20 and Sept. 9, 2020. We looked for activity from automated (or bot) accounts, and the spread of distorted or conspiracy theory narratives.

We learned that on Twitter, many conspiracy theories, including QAnon, may not be quite as popular among real people as media reports indicate. But automation can significantly increase the distribution of these ideas, inflating their power by reaching unsuspecting users who may be drawn in not by posts from their fellow humans, but from bots programmed to spread the word.

Bots amplify conspiracy theories

Typically, bots are created by people or groups who want to amplify certain ideas or points of view. We found that bots are roughly equally active in online discussions of both right-wing and left-wing perspectives, making up about 5% of the Twitter accounts active in those threads.

Bots appear to thrive in political groups discussing conspiracy theories, making up nearly 13% of the accounts tweeting or retweeting posts with conspiracy theory-related hashtags and keywords.

Then we looked more closely at three major categories of conspiracies. One was a category of alleged scandals described using the suffix “-gate,” such as “Pizzagate” and “Obamagate.” The second was COVID-19-related political conspiracies, such as biased claims that the virus was deliberately spread by China or that it could be spread via products imported from China. The third was the QAnon movement, which has been called a “collective delusion” and a “virtual cult.”

These three categories overlap: Accounts tweeting about material in one of them were likely to also tweet about material in at least one of the others.

The link to right-wing media

We found that the accounts that are prone to share conspiratorial narratives are significantly more likely than nonconspirator accounts to tweet links to, or retweet posts from, right-leaning media such as One America News Network, Infowars and Breitbart.

[Deep knowledge, daily. Sign up for The Conversation’s newsletter.]

Bots play an important role as well: More than 20% of the accounts sharing content from those hyperpartisan platforms are bots. And most of those accounts also distribute conspiracy-related content.

Twitter has recently tried to limit the spread of QAnon and other conspiracy theories on its site. But that may not be enough to stem the tide. To contribute to the global effort against social media manipulation, we have publicly released the dataset used in our work to assist future studies.

Emilio Ferrara, Associate Professor of Computer Science; USC Viterbi School of Engineering; Associate Professor of Communication, USC Annenberg School for Communication and Journalism

This article is republished from The Conversation under a Creative Commons license. Read the original article.


Charlene is a Bay Area journalist who hails from the small community of Fresno. Drawing from her experience writing for her college paper, Charlene continues to advocate for free press and local journalism. She also volunteers in all the beach cleanups she can because she loves the water.