Sunday, November 24, 2024

conspiracy resource

Conspiracy News & Views from all angles, up-to-the-minute and uncensored

QAnon

It’s too late to stop QAnon with fact checks and account bans

The traditional understanding of QAnon was that its ideas are spread by a relatively small number of adherents who are extremely good at manipulating social media for maximum visibility. But the pandemic made that more complicated, as QAnon began merging more profoundly with health misinformation and rapidly increasing its presence on Facebook. 

At this point, QAnon has become an omniconspiracy theory, says DiResta—it’s no longer just about some message board posts, but instead a broad movement promoting many different, linked ideas. Researchers know that belief in one conspiracy theory can lead to acceptance of others, and powerful social-media recommendation algorithms have essentially turbocharged that process. For instance, DiResta says, research has shown that members of anti-vaccine Facebook groups were seeing recommendations for groups that promoted the Pizzagate conspiracy theory back in 2016. 

“The recommendation algorithm appears to have recognized a correlation between users who shared a conviction that the government was concealing a secret truth. The specifics of the secret truth varied,” she says. 

Researchers have known for years that different platforms play different roles in coordinated campaigns. People will coordinate in a chat app, message board, or private Facebook group, target their messages (including harassment and abuse) on Twitter, and host videos about the entire thing on YouTube.  

In this information ecosystem, Twitter functions more like a marketing campaign for QAnon: content is created to be seen and interacted with by outsiders. Meanwhile, Facebook is a powerhouse for coordination, especially in closed groups. 

“Q” has made many incorrect predictions and continues to advance the belief that there is a “deep state” plot against Donald Trump.

Reddit used to be a mainstream hub of QAnon activity, until the site started clamping down on it in 2018 for inciting violence and repeatedly violating its terms of service. But instead of losing its power, QAnon simply shifted to other mainstream social-media platforms where it was less likely to be banned. 

This all means that when a platform acts on its own to block or reduce the impact of QAnon, it only attacks one part of the problem. 

Friedberg said that to him, it feels as if social-media platforms were “waiting for an act of mass violence” in order to coordinate a more aggressive deplatforming effort. But the potential harm of QAnon is already obvious if you stop viewing it as a pro-Trump curiosity and instead see it for what it is: “a distribution mechanism for disinformation of every variety,” Friedberg said, and one that adherents are willing to openly promote and identify with, no matter the consequences.  

“Three years of almost unfettered access”

Steven Hassan, a mental health counselor and an expert on cults who escaped from Sun Myung Moon’s Unification Church, known as the “Moonies,” says that discussing groups like QAnon as solely a misinformation or algorithmic problem is not enough. 

“I look at QAnon as a cult,” Hassan says. “When you get recruited into a mind control cult, and get indoctrinated into a new belief system … a lot of it is motivated by fear.” 

They’ve had three years of almost unfettered access to develop and expand.

“People can be deprogrammed from this,” Hassan says. “But the people who are going to be most successful doing this are family members and friends.” People who are already close to a QAnon supporter could be trained to have “multiple interactions over time” in hopes of pulling the person out. 

If platforms wanted to seriously address ideologies like QAnon, they’d do much more than they are, he says.

First, Facebook would have to educate users not just on how to spot misinformation, but also on how to understand when they are being manipulated by coordinated campaigns. Coordinated pushes on social media are a major factor in QAnon’s growing reach on mainstream platforms, as recently documented by the Guardian. The group has explicitly embraced “information warfare” as a tactic for gaining influence. In May, Facebook removed a small collection of QAnon-affiliated accounts for inauthentic behavior

And second, Hassan recommends that platforms stop people from descending into algorithmic or recommendation tunnels related to QAnon, and instead feed them with content from people like him, who have survived and escaped from cults—especially from those who got sucked into and climbed out of QAnon. 

Friedberg, who has studied the movement deeply, says he believes it is “absolutely” too late for mainstream social-media platforms to stop QAnon, although there are some things they could do to, say, limit its adherents’ ability to evangelize on Twitter. 

“They’ve had three years of almost unfettered access outside of certain platforms to develop and expand,” Friedberg says. Plus, QAnon supporters have an active relationship with the source of the conspiracy theory, who constantly posts new content to decipher and mentions the social-media messages of Q supporters in his posts. Breaking QAnon’s influence would require breaking trust between “Q,” an anonymous figure with no defining characteristics, and their supporters. Given that Q’s long track record of inaccurate predictions has not broken that trust, that’s difficult, and critical media coverage or deplatforming have yet to really do much on that front. If anything, they only fuel QAnon believers to assume they’re on to something. 

The best ideas to limit QAnon would require drastic change and soul-searching from the people who run the companies on whose platforms it has thrived. But even this week’s announcements aren’t quite as dramatic as they might seem at first: Twitter clarified that it wouldn’t automatically apply its new policies against politicians who promote QAnon content, including several promoters who are running for office in the US.

And, Friedberg said, QAnon supporters were “poised to test these limitations, and already testing these limitations.” For instance, Twitter banned certain conspiracy-affiliated URLs from being shared, but people already have alternative ones to use. 

In the end, actually doing something about that would require “rethinking the entire information ecosystem,” says DiResta. “And I mean that in a far broader sense than just reacting to one conspiracy faction.”

*** This article has been archived for your research. The original version from MIT Technology Review can be found here ***