Monday, November 25, 2024

conspiracy resource

Conspiracy News & Views from all angles, up-to-the-minute and uncensored

QAnon

On Twitter, bots spread conspiracy theories and QAnon talking points

(The Conversation is an independent and nonprofit source of news, analysis and commentary from academic experts.)

Emilio Ferrara, USC Annenberg School for Communication and Journalism

(THE CONVERSATION) Americans who seek political insight and information on Twitter should know how much of what they are seeing is the result of automated propaganda campaigns.

Nearly four years after my collaborators and I revealed how automated Twitter accounts were distorting online election discussions in 2016, the situation appears to be no better. That’s despite the efforts of policymakers, technology companies and even the public to root out disinformation campaigns on social media.

In our latest study, we collected 240 million election-related tweets mentioning presidential candidates and election-related keywords, posted between June 20 and Sept. 9, 2020. We looked for activity from automated (or bot) accounts, and the spread of distorted or conspiracy theory narratives.

We learned that on Twitter, many conspiracy theories, including QAnon, may not be quite as popular among real people as media reports indicate. But automation can significantly increase the distribution of these ideas, inflating their power by reaching unsuspecting users who may be drawn in not by posts from their fellow humans, but from bots programmed to spread the word.

Bots amplify conspiracy theories

Typically, bots are created by people or groups who want to amplify certain ideas or points of view. We found that bots are roughly equally active in online discussions of both right-wing and left-wing perspectives, making up about 5% of the Twitter accounts active in those threads.

Bots appear to thrive in political groups discussing conspiracy theories, making up nearly 13% of the accounts tweeting or retweeting posts with conspiracy theory-related hashtags and keywords.

Then we looked more closely at three major categories of conspiracies. One was a category of alleged scandals described using the suffix “-gate,” such as “Pizzagate” and “Obamagate.” The second was COVID-19-related political conspiracies, such as biased claims that the virus was deliberately spread by China or that it could be spread via products imported from China. The third was the QAnon movement, which has been called a “collective delusion” and a “virtual cult.”

These three categories overlap: Accounts tweeting about material in one of them were likely to also tweet about material in at least one of the others.

The link to right-wing media

We found that the accounts that are prone to share conspiratorial narratives are significantly more likely than nonconspirator accounts to tweet links to, or retweet posts from, right-leaning media such as One America News Network, Infowars and Breitbart.

[Deep knowledge, daily. Sign up for The Conversation’s newsletter.]

Bots play an important role as well: More than 20% of the accounts sharing content from those hyperpartisan platforms are bots. And most of those accounts also distribute conspiracy-related content.

Twitter has recently tried to limit the spread of QAnon and other conspiracy theories on its site. But that may not be enough to stem the tide. To contribute to the global effort against social media manipulation, we have publicly released the dataset used in our work to assist future studies.

This article is republished from The Conversation under a Creative Commons license. Read the original article here: https://theconversation.com/on-twitter-bots-spread-conspiracy-theories-and-qanon-talking-points-149039.

*** This article has been archived for your research. The original version from The Baytown Sun can be found here ***