- 1Kresge Library, Oakland University, Rochester, MI, United States
- 2Fine and Performing Arts Library, University of New Mexico, Albuquerque, NM, United States
Introduction: This article explores two QAnon subgroups that were not active during the initial phase of the movement but now epitomize how QAnon has capitalized on social media to reach more people. We examine these smaller communities through the lens of information literacy and other literacies to identify opportunities for librarians and educators.
Results: The communities of conspiracy theorists explored here exhibit information behaviors distinct from the initial QAnon community, presenting opportunities for information professionals to employ new models of information literacy, metaliteracy, and other literacies to combat conspiracy ideation. Notable themes evidenced in both samples include an increasing religiosity affiliated with QAnon, affective states that promote conspiracy ideation, faulty hermeneutics and epistemologies, and specific literacy gaps.
Methods and discussion: We must update our understanding of QAnon and its adherents' shifting priorities and behaviors. Through investigating these smaller subgroups, researchers and educators can address the evolution of the QAnon movement by teaching to literacy gaps and logical fallacies, and acknowledging the troubling emotions that undergird broader belief systems.
1 Introduction
Librarian educators have spent the past two decades educating learners in information literacy; however, the current methods and frameworks falter when confronting the epistemologies that have given rise to QAnon and other conspiracy theories in American culture. Educators and librarians often feel overwhelmed by individuals or groups acting on their conspiracy interests, including the rise in book bannings and similar challenges nationwide (Beene and Greer, 2023; PEN America, 2023). These conspiracies threaten democratic norms, as illustrated by the January 6, 2021, breach of the U.S. Capitol building (McKay and Tenove, 2021; Fisher, 2022). Thus, information professionals and educators must understand the current challenges facing internet users who may be drawn into conspiracy theories to counteract them. As the superconspiracy QAnon has shifted to absorb other conspiracy theories and proliferated via social media, the movement has fractured into smaller groups, often led by influencers who have their own agendas and interests.
To better understand these smaller social media communities and the opportunities and challenges they present for librarians and educators, the authors explored two discrete examples: one influencer on Facebook with a small following of ~4,000 people, and another influencer on Telegram with a much larger following. Telegram has gained scholarly attention for its unique platform affordances in spreading conspiracy theories (Garry et al., 2021; Walther and McCoy, 2021; Peeters and Willaert, 2022), while Facebook has been explored as a social media platform ideally situated to wield outsized influence on the spread of hoaxes, conspiracy theories, misinformation and disinformation (Yasmin and Spencer, 2020; Innes and Innes, 2021; Oremus and Merrill, 2021; Marko, 2022; Mont'Alverne et al., 2023). Following these scholars, we sought to explore the evolution of the QAnon movement. Our previous research enabled us to more readily identify original QAnon themes (Beene and Greer, 2021, 2023) and trace emerging trends. In addition, we were interested in observing the information behaviors, affective states or emotional content, and cognitive patterns evident in these newer communities. As information professionals, we were especially interested in notable literacy gaps (e.g., information, media, digital, visual, and metaliteracy) and the lessons librarians and educators could glean.
In order to expand our praxis, information professionals must reckon with networked conspiracy theories, as well as with the role of emerging cognitive and parasocial authorities in creating and disseminating information. Investigating these communities will allow librarians and educators to account holistically for online information's cognitive, affective, and social dimensions. While research has been conducted on the original QAnon movement and its challenges for information professionals (Beene and Greer, 2021, 2023), this article provides insight into the evolving concerns and interests of splinter QAnon communities.
In addition to some original themes from the QAnon movement, we observed religiosity, white supremacy, and alternative health and wellness philosophies integrated with QAnon. We also noted a myriad of faulty epistemologies and hermeneutics that leveraged strong emotions to bypass critical literacies. This research fills a gap in examining information practices and discourse within the Q subcommunities. The themes that surfaced have implications for future pedagogical interventions or inoculations against conspiracism.
2 Literature review
2.1 The QAnon superconspiracy
In an earlier article (Beene and Greer, 2021), we examined the rise of QAnon and its implications for information professionals. In its most basic premise, QAnon posits that a global cabal of Satanic pedophiles traffic children worldwide. In the QAnon mythology, Donald Trump plays a pivotal role, working behind the scenes to fight and expose the cabal in an event known to followers as “the Storm” or the “Great Awakening.” The initial QAnon movement was constructed through members of anonymous image posting boards known as “chans,” where an anonymous user named “Q” would post daily cryptic messages, which were then scrupulously analyzed by followers. Soon after its inception, the movement jumped over to social media platforms like YouTube, where influencers interpreted Q drops, or posts, for the masses. Over the last seven years, QAnon has evolved from a single conspiracy theory into a “superconspiracy,” which Barkun (2003) defined as complex, evolving, and amorphous, “reflect[ing] conspiratorial constructs in which multiple conspiracies are believed to be linked together” and are “joined in complex ways so that conspiracies come to be nested within one another” (Barkun, 2003, p. 6). These nested conspiracies became the focus of online and social media communities.
This super conspiracy has infected popular social media, our political landscape, and our relationships. Exploring how, exactly, this happened is nearly impossible; in his analysis of how one falls down the rabbit hole, Pierre (2023) noted that, due to the amorphous and complex nature of the conspiracy worldview, “the process of becoming a QAnon follower can be understood as an expression of individual psychological vulnerability to conspiracy belief, cult affiliation, and ARG (alternate-reality game) engagement” (p. 21). To mitigate this conspiratorial mindset, researchers and educators must address “myriad subcategories of (the QAnon) identity and pathways of individuation” (p. 28). The multiple pathways to QAnon result in a spectrum of conspiracy ideation, as delineated by Pierre; there can be those who are “fence-sitters,” or more open to the ideas and questions promulgated by the Q community, “true believers” who have accepted the tenets and identify as committed followers, or “activists” who act as evangelists for the movement. Importantly, Pierre concludes his typology with “apostates” who have worked their way through the spectrum and come out on the other side, renouncing their belief in and affiliation with the community (p. 25–27).
2.2 Information behaviors and discourses in the QAnon community
Several scholars have analyzed discourses within the QAnon community, with most focusing on the nascent QAnon communities that evolved on the 4chan, 8chan, and 8kun message boards. For example, Packer and Stoneman (2021) noted QAnon's unique discursive affordances that “rely heavily on a communicative strategy of encoding and decoding that bears a strong resemblance to an esoteric hermeneutic, but one played out across social media” (p. 256). Q drops, or messages, required active participation from followers, which set up “an opportunity for social media celebrity and monetization” (Packer and Stoneman, 2021, p. 263).
Chandler's 2020 thematic analysis of original Q drops and its three significant subcategories has been used to frame our analyses: “(1) anti-elitism and discrediting the establishment; (2) claims of group marginalization and oppression, and (3) QAnon's role in unifying ‘patriotic' Americans, through covert and, ultimately, overt actions” (Chandler, 2020, p. 20). We also utilized Hodges' 2023 taxonomy of the QAnon worldview, which reflects “particular (if problematic) forms of engagement with authentic sources rather than the result of wholesale fabrication” (p. 129). His taxonomy of users' information practices included “close reading, exaggeration, accusation of ignorance, and deployment of esoteric yet unclear arguments” (p. 130). Hodges cautions against casually dismissing or deriding conspiracy theorists, concluding that
Understanding the relationship between QAnon and consensus reality more deeply and with greater nuance can assist in enabling more effective interactions and, potentially, even interventions between participants and third parties such as information, library, and media workers (2023, p. 140).
Hodges' conclusions turn on the nuance of interpretation as the driving force in the QAnon movement. Indeed, based upon the nascent QAnon community, some have suggested that the movement functions as a participatory game (Berkowitz, 2021; Fitzgerald, 2022; Gabriel Gatehouse, 2022) and that QAnon adherents demonstrate relatively sophisticated information literacy behaviors (Hartman-Caverly, 2019). We seek to update these earlier assessments to account for QAnon's current form on the social web.
2.3 Conspiracy adherents' epistemologies and hermeneutics
This debate between hermeneutics and epistemology is not new and has been explored by scholars in psychology, sociology, and philosophy, among other fields. For example, Ellis (2015) explored moral hermeneutics, coherence epistemology, and affect, while Westphal (2017) discussed the collapse of epistemologies into a hermeneutical framework. Both scholars complicated the notion that epistemology is opposing hermeneutics: that is, the way we construct knowledge is inherently tied up with how we interpret events – and both of these are inextricable from affective states. Building on this debate, Dentith (2018) and Butters (2022) examined these concepts in relation to conspiracy ideation. While Dentith examined the types of experts and other cognitive authorities deployed in conspiracy theory arguments, Butters interrogated how individuals, mainstream news media, and politicians amplified particular worldviews and ideologies. Both scholars found that QAnon followers use cherry-picked and misinterpreted evidence to prop up their worldviews. Thus, it could be assumed that helping users better understand the underlying modes of information production and discourse could assist in dismantling conspiracy interpretation.
QAnon is only one of a spectrum of conspiratorial worldviews, and recent research provides a greater understanding of conspiracy ideation in general. Conspiracy theorists, for example, find it easier to believe in additional conspiracies when they have already subscribed to one (Brotherton and French, 2015; Brashier, 2022). Additionally, those who fall prey to conspiracy theories are often motivated by reasoning and biases that lead them to search for information confirming their preconceptions, beliefs, and values; thus, conspiracy theories may seem more trustworthy because they confirm a particular worldview (Brotherton and French, 2015). As adherents fall deeper into the rabbit hole, they use motivated reasoning and cherry-picking of evidence, accepting only information that confirms a pre-existing philosophy and over-correlating innocuous and disparate events to confirm their beliefs (Sullivan et al., 2010). Finding like-minded users online essentially creates an echo chamber – a term that describes how users within an online community echo, confirm, and reiterate each other's beliefs and opinions. Douglas et al. (2019) argue that conspiracy beliefs are motivated by epistemic, existential, and social needs and not an innate tendency to simple credulity.
Moreover, there is evidence that the concept of authority operates differently on social media platforms, including within QAnon subgroups. Robertson and Amarasingam (2022) discovered that QAnon adherents often rely upon experiential knowledge rather than expertise acquired through research (2022, p. 2013). Their mixed-methods analysis of Telegram examined epistemic models of authority in right-leaning conspiratorial narratives, discovering an abundance of reliance on institutions such as the U.S. and its Constitution (p. 198–201). These narratives center on epistemic modes like scientific reasoning because they are “most prevalent in society and ones they need to most vociferously argue against and dismantle” (p. 198). Most importantly, they found reiterations of QAnon's original mantras, such as “The Storm,” “The Great Awakening,” and “Trust the Plan,” to proliferate in arguments, commenting that the ambiguity of these terms allows for broad interpretations, including theological and millennialist (p. 200). They concluded that evidence and authority predominate in these communities but only to prop up faulty epistemologies and hermeneutics.
These findings correlate with scholars' investigations into how social media users assess online information, where authority is conferred through parasocial and cognitive authorities (Rieh, 2010; Badke, 2015; Liebers and Holger, 2019), and second-hand knowledge (Wilson, 1983). When we do not possess enough knowledge on a topic, we decide whether or not to trust someone or something based on their perceived competence and expertise. For example, most of us have not empirically studied gravity, so we trust the expertise of those who have (Russo et al., 2019). Rather than characterizing our Information Age as “Post-Truth,” then, we might instead think of it as a plethora of competing truth claims. Because we cannot be experts on each claim, we increasingly trust others to connect the dots for us – increasingly in online spaces. The notion of authority is thus dynamic, shifting over time and space. It has shifted from the source itself to individual belief systems.
2.4 Information literacy and metaliteracy in an evolving information landscape
Academic librarians utilize as their primary guiding document the Framework for Information Literacy for Higher Education, codified after years of revising the concept of information literacy (Association of College and Research Libraries, 2016). The document posits an expanded understanding of information literacy as a “set of integrated abilities encompassing the reflective discovery of information,” with attention to how information is produced and disseminated through social, economic, and ethical lenses. The information landscape has evolved tremendously in the past decade, however. The way we think of and teach information literacy has shifted to encompass the new ways in which facts, truth, and trust are challenged by a heightened level of relativism, in which information is assessed using highly personal heuristics (Swanson, 2023).
Compounding the challenges accompanying the evaluation and use of new types of information, librarians have observed a surge in conspiracy theories. In a 2021 nationwide survey, we discovered that American library workers encountered conspiracy ideation in academic, public, special, and K-12 libraries (Beene and Greer, 2023). Importantly, affective states were evident in both patrons and library workers, often compounded by the political and social turmoil of another pandemic year just months after the January 6, 2021, capitol insurrection. Even though the existing literacy frameworks account for some affective components, we are among a growing body of researchers seeking to expand them to include a more holistic accounting of affective states and how they relate to information assessment, especially that which proliferates online.
Metaliteracy, as a separate framework from those authored by academic librarians, could serve as a valuable antidote to misinformation on the social web. Metaliteracy is a unified understanding of discrete literacies (e.g., information, visual, digital, media) and employs metacognitive reflective practices for information evaluation in collaborative online communities (Mackey and Jacobson, 2011). It includes explicit connections to the cognitive, behavioral, metacognitive, and affective domains of learning. As a framework, it addresses some of the shortcomings in discrete literacy models, seeking to form “an empowered and self-directed learner who actively creates meaningful content and effectively participates in a connected world” (Mackey and Jacobson, 2018, p. 1). This socially constructed online information ecosystem encourages users to actively create and produce knowledge.
In many ways, metaliteracy complements the work of other information literacy theorists who conceive of it as a socio-cultural practice (Lloyd, 2010). Lloyd reframes information literacy as a messy, social, participatory, and affective process through which meaning is made and shared (2010, p. 14–24). Her conception of an “information landscape,” which she defines as “the communicative space through which people develop identities and form relationships based on shared practices and ways of doing and saying things” (p. 9), is especially salient to this paper, providing a useful descriptor for how information and meaning are enacted and interacted within the social web.
Likewise, Swanson (2023) argues that the cognitive and affective dimensions inform and build upon each other. Indeed, the mind-gut connection is genuine regarding feelings of knowing. The profession needs an “affective turn,” as has taken place in other disciplines (Zembylas, 2021), which would provide a better understanding of how the mind parcels content, sifts it for essential bits, and connects information to feelings, emotions, and other forces that work upon an individual's experience of the world. This understanding would better prepare information professionals to teach library patrons and students to slow down and evaluate their feelings and thinking before assessing or using a source.
3 Materials and methods
The authors used opportunistic sampling to observe two current QAnon-focused social media communities. We settled on Pastor Chris1 and IET after following various conspiracy theory influencer trails. Pastor Chris provided engagement, a mid-size number of followers, and a salient example of the ongoing intersections of evangelicalism and QAnon via Facebook, which is one of the first mainstream platforms that Q followers congregated on after the chan boards (Kaplan, 2023). Despite banning overt QAnon groups in 2019, QAnon still has a very active presence there through communities like Pastor Chris'. We selected 112 of Pastor Chris' posts from the beginning of their account in May 2021, with additional posts selected from late spring of 2022 and 2023 to illustrate any continuity of themes through time. In addition, ad hoc posts that offered illustrative material for specific literacy gaps or identified themes were chosen.
IET provided a more sizable following with his public Telegram account, whose posts regularly reached over 50,000 views in 2022 and demonstrated regular interactions with his followers. The use of Telegram was important; the online social movement of QAnon migrated from platforms like Facebook to alternative social media platforms such as Telegram, Parler, and Truth Social once QAnon was banned on them (Mac and Lytvynenko, 2020). He has received more attention than Pastor Chris, as he has been profiled by Vice magazine and the Anti-Defamation League; however, he is still mostly unknown by the mainstream public (Gilbert, 2021b; Anti-Defamation League, n.d.). As a Christian wellness influencer and QAnon adherent, he provided a snapshot of the coalescence of communities, including health and wellness and anti-vaccination communities, QAnon and MAGA, and evangelical Christianity. Telegram does not allow screen capture or scraping and requires that one sign up via their phone number – because of these limitations, posts were transcribed verbatim into a spreadsheet for analysis. IET began posting on Telegram in 2021 but gathered momentum with more views and followers in 2022. For this study, we selected posts from August 1-5 in 2021 and 2022. 2023 proved more challenging because he decreased his activity on Telegram, with no posts for August 2023; therefore, 5 days in late April 2023 were selected for comparison to his earlier activity. Posts were coded deductively using the taxonomies described below, in addition to inductively for QAnon behaviors and themes, affiliated conspiracy theories, affective behaviors, and any literacy gaps.
Each of us took responsibility for one sample group and coded the posts individually, although both worked from the same initial codebook. Both deductive and inductive coding were used. Deductive coding was derived from the Chandler (2020) and Hodges (2023) QAnon taxonomies to determine if themes from the early movement were still applicable within these smaller splinter groups. Inductive coding explored themes that emerged in each sample:
• QAnon as a superconspiracy: What other conspiracy theories are being mentioned?
• Affective states: over-confidence, confirmation bias, encouragement, and emotions.
• Literacy gaps: How are news items or other content being interpreted (or misinterpreted)?
• Evangelicalism, exegesis, and Andre's (2023) five rhetorical strategies
Along with coding, discourse analysis was used to explore communication within the communities. Crossley (2021) explained that discourse “looks at the overall meanings conveyed by language in context,” with context indicating the “social, cultural, political, and historical background (which is) important to take into account to understand underlying meanings expressed through language” (Crossley, 2021, “What is discourse analysis?”). The number of reactions and views to each post were thus recorded, in addition to whether and how emojis were used in those reactions. The authors also recorded whether the influencers or users shared stories from mainstream or dubious news media outlets or other social media platforms, in addition to noting any cross-posting between groups, individuals, or other influencers. This practice helped the authors understand the discursive ecosystem of conspiracy theories on the social web.
3.1 The pilled pastor
Self-proclaimed Pastor Chris started a small evangelical church in the southeastern United States in 2009. In their early 60s, they are most active through a public, personal Facebook profile directly linked to a church's website. They currently have around 4,000 followers. A Facebook profile was chosen because Facebook, until the advent of right-wing focused social media such as Truth Social and Telegram, served as the main conduit for the early organization of QAnon followers (Argentino, 2020b). Facebook has also served as one of the primary news sources for Americans, with roughly half of Americans getting their news from social media (Pew Research Center, 2020). Numerous QAnon influencers have sprung up in this space, developing communities of highly engaged followers and connecting with other influencers.
Like many conspiracy theorists active on Facebook, Pastor Chris initially maintained two accounts in case one was put on a temporary hold due to content violations–this is a common practice to avoid disengagement during what are termed “Facebook jail” periods. Only one account remains active and has been since 2021. They also have accounts on other platforms, but those are not as active and do not have as many followers.
The discourse of Pastor Chris' community illustrates how a charismatic leader functions in a new religious movement. Although QAnon and Trump are seen as the ultimate authorities, Pastor Chris' posts exhort, encourage, and evoke the affective states of followers. A common refrain is that the benefits of events like the jubilee or the Storm will be greatest for followers who believe. This caveat shifts responsibility away from the influencer when things do not happen as promised. Many posts include long response threads consisting entirely of “hands praying” emojis, “I believe,” or the comment “Amen” in response to the pastor's exhortations about how God's plan (enacted by Trump and QAnon) is coming together. Skeptical posts are often met with statements that negate the skepticism because the pastor's belief that something is true overrules any fact-checking.
What is striking about Pastor Chris' Facebook activity is the repetition. The same themes, including that the long-awaited events of the Storm “could happen soon,” repeat throughout the years. New dates are added to the timeline and new secret communications are revealed with each repetition. Those who start to doubt are excoriated for questioning, despite the QAnon mantra to “Question Everything.” Pastor Chris once commented, “If you decide you want to walk away and quit, do not comment it on [sic] this post, they will be deleted! Go find a post that's negative doom and gloom and stick around for that [sic] but you are not gonna do it on my post!”
Unlike the original QAnon community, Pastor Chris does not perform close reading; instead, they share content from others doing the work. The influencer is a secondary source, a node on a vast misinformation network that proselytizes the gospel of Q farther and farther from its locus. They instead contribute personal prophecies and affirmations of belief to spur others on. Posts that describe sensing a “major shift” or dream visions that indicate the conspiracy is coming to light are frequent themes throughout the months and years.
3.2 The chiropractor conspiracy theorist
IET, a chiropractor and influencer in the health and wellness industry, is affiliated with the pseudoscience and CrossFit communities as well as QAnon. IET was “outed” by journalist David Gilbert of Vice Magazine (Gilbert, 2021a,b), who investigated other major QAnon influencers and their online presence (Gilbert, 2023). The story was quickly picked up by other outlets (Flomberg, 2021; LightWarrior, 2021), including Reddit communities (Cuicksilver, 2021), various activist sites (Unknown, 2021b; Garber, 2022), and the Anti-Defamation League ((Anti-Defamation League, n.d.).This spread across alternative and mainstream networks demonstrates the speed and power of the social web, features that can be exploited for legitimate or illegitimate purposes. IET is a good example of influencer culture and its role in spreading faulty (and harmful) information across platforms.
IET's use of Telegram illustrates the role that more niche social media spaces are playing in American information consumption. This fact correlates with findings from a 2022 Pew Research Institute survey on the role alternative social media plays in Americans' news and information environment; in it, they found that a very small percentage (only 6%) of Americans regularly get their news from such sites, but of that 6%, 65% feel they have found their community, which is often more pro-Trump, pro-America, and religious (Stocking et al., 2022). IET's rhetoric certainly aligns with a pro-Trump and religious worldview, albeit conspiratorial. These social media platforms have earned a reputation for their role in organizing followers for extremist events (Garry et al., 2021; Walther and McCoy, 2021; Myers and Frenkel, 2022). Users like IET exploit the secretive nature of Telegram, sometimes opting to use the disappearing posts feature, whereby posts disintegrate after an allotted period of time. In addition, Telegram challenges hackers to try to crack it, with no one succeeding yet. IET posts his own content in addition to other creators' content. In particular, he favors sharing TikTok videos, which the Pew Research Institute has recently shown to be a dominant information practice for about a third of Americans over 30 years old (Matsa, 2023). Like Telegram, Gab, and Parler, other scholars have also studied the rapid proliferation of misinformation via TikTok (Michel et al., 2022). This practice serves as a meditation on the power of multimodal information in spreading conspiracist ideologies. While IET shuns “fakebook,” he does have a public account on X (formerly Twitter), especially after Elon Musk acquired the platform and reinstated banned accounts (The Associated Press, 2022; Thorbecke and Catherine, 2022). In this way, he is complementary to Facebook influencers like Pastor Chris.
IET attracts those users already consuming alternative news and social media sources; in many cases, his readers follow the same demagogues. Indeed, his followers are enthusiastically engaged, and he regularly encourages interaction. The discourse within this Telegram community demonstrates the construction of knowledge between and among followers and their influencers, a practice found in other online communities that capitalize on user's affective states (Ziegler et al., 2014).
IET often posts as a wellness influencer, providing a gateway for followers to learn about Christianity, QAnon, and other worldviews. For example, IET was interviewed on Rumble, where he discussed alternative health solutions, including parasite cleanses, Ivermectin, and venoms, with another influencer who posts under the headline “The Kate Awakening” (a play on QAnon's mantra “The Great Awakening”) (The Kate Awakening - Parasite Cleanse Part 2 with IET and SSGQ, 2023).
Like Pastor Chris, IET began using Telegram in 2021 to delve into exegetical readings of the Bible to interpret current and future events, often linked to QAnon and white supremacist themes. This practice continued to a lesser extent into 2023. He seemed to grow his following through a bait-and-switch, where he acted as a layman's pastor, interpreting Biblical books for his followers. In almost every post, he implores his followers to pay attention to the groundwork he is laying for them and to abide by his interpretations. The use of capitalization, not unlike former president Trump's tweets, serves to highlight his main points; strategic use of emojis appears in later posts. These close biblical readings are commonly linked to white supremacist themes, such as when he exhorts his readers to “keep bloodlines pure.” As he posts, he becomes emboldened and begins to refer to Kekistan, Pepe, and anti-Semitic tropes. Through using these memes (Kek and Pepe), he signals his “insider knowledge” to his followers, as these memes originated on the chan boards and have been used to mobilize extremist movements globally (Niewert, 2017; McFall-Johnsen and Morgan, 2021; Topor, 2022; Pepe the Frog, 2023). For his influence in this area, he earned a listing in the Anti-Defamation League's glossary and other activist sites working to curb fascist ideologies (Unknown, 2021a).
After being a featured speaker at the Qanon conference, “For God And Country Patriots' Roundup,” in Dallas, Texas (Unknown, 2021a), IET skyrocketed in popularity. His posting about QAnon increased, especially in 2022, which coincides with the QAnon community's celebration of Q's return (Kaplan, 2022). Evidenced by interactions with and views of his Telegram posts, many of which garnered up to 50,000 views and hundreds of reactions and comments, he began to use more explicit conspiracist rhetoric, not only pertaining to QAnon, but also myriad political conspiracy theories, anti-vaccination themes, and COVID-19. He used his newfound popularity to proselytize and peddle various political conspiracy theories that have become associated with QAnon, including those surrounding “groomers” and saving the children, conspiracies involving the Bidens, Obama, and Trump, and those proliferating about the Ukraine/Russia conflict.
4 Findings
In exploring these two communities, numerous themes became apparent. Not only were the QAnon themes derived from the original community still overarching in the discourse, but additional concerns and information practices have developed. Partly due to the affordances of the particular social media platforms we chose, and partly due to the removal in time and digital space from the original chan boards, these themes highlight potential new areas for research and intervention.
4.1 Original QAnon themes
Some themes still predominated from original QAnon communities – for example, conspiracy theories surrounding the Illuminati and the political elite, the global cabal's exploitation of children, and Trump's role in fighting them. The authors observed original QAnon motifs such as “white hats” vs. “black hats” and the QAnon theme of prominent people being fake doubles (because the real person is locked up already or has been executed).
The taxonomies observed by Chandler (2020) and Hodges (2023) have been watered down in these spaces. A thread of anti-elitism and anti-establishmentarianism runs through many of Paster Chris' and IET's posts, but the theme of “QAnon unif[ying] ‘patriotic' Americans through action” has been supplanted in favor of passive watching and waiting. Similarly, although Hodges' categories of “exaggerated interpretation of correlation” and “esoteric but unclear meaning” occur in some posts, the close reading of the early QAnon communities is largely done by others and passed along by Pastor Chris and IET.
4.1.1 QAnon and religion
The QAnon movement has always had Christian overtones; as early as 2020, several who had been keeping an eye on the movement began describing it as a religion (LaFrance, 2020). Amarasingam et al. (2023) analyzed the QAnon movement in terms of the literature defining new, emerging religious movements, and “hyperreal religions,” movements that borrow heavily from popular culture to shape meaning and belief. They concluded that there is significant overlap between these movements and QAnon, and that viewing QAnonthrough these lenses allows for deeper insights into the movement. Andre (2023) noted with concern that “QAnon is reshaping the topography of many churches” (p. 191), analyzing five overarching rhetorical strategies used by Q in their posts. According to Andre, these rhetorical strategies align with Christianity: civil religion, spiritual language, spiritual warfare, explicitly Christian content, and Christian nationalism (p. 194–200).
Journalist Jeff Sharlet delved deeper into the conflation of Trump, prosperity gospel teachings, and QAnon (2023). He attended multiple Trump rallies and spoke with attendees, some of whom were ordained pastors and ministers. Because of his interactions with devoted Trump followers, many of them professed Christians, Sharlet began to investigate the overlaps between the so-called “prosperity gospel” and the Trump campaign:
The Christian Right that has so long dominated the political theology of the United States emphasizes a heavenly reward for righteousness in faith and behavior; the prosperity gospel is about…”amazing results” you can measure and count. The old political theology was about the salvation to come; the Trump religion was about deliverance, here and now (Sharlet, 2023, p. 49, 50).
For example, an attendee named Pastor Sean is quoted as saying, “‘Secret murders everywhere,' …‘Pedophiles, and evil;” in his conversations with Sharlet, he professed belief that “God had chosen Trump” (Sharlet, 2023, p. 116). Another attendee, Pastor Dave, told Sharlet that he had felt a “spiritual drawing” to devote his life to Trump (Sharlet, 2023, p. 117). Indeed, at the beginning of every Trump rally, a pastor would lead the crowd in prayer (Sharlet, 2023; Hassan and Johnson, 2024).
Lundsko and MacMillen (2023) explored the notion of QAnon as a religion or “replacement reality” extensively. Their assertion that “Psuedo-religion-seeking-certainty fuels conspiratorial thinking and is imbued in a certain type of propagandistic rendering of religion” (p. 59) reflects the rhetoric of Pastor Chris' and IET's communities. Pastor Chris, for example, exhorts followers to “innerstand,” a play on “understand” that emphasizes one's internal and affective epistemological constructs. Their posts frequently reference spiritual warfare and Christian nationalism, echoing the rhetorical strategies of Q. The battle between good and evil is Biblical; God is an American. Posts shared by Pastor Chris from cognitive authorities such as “Prophet Jeff Johnson” state things such as “The season of dancing in the streets is upon us as the TRUMP [sic] is sounding and the time of Presidential occupation is here.” Here, the Biblical trumpet of Revelation has been conflated with former president Trump, who, in other posts and comments, has been declared “a decedent [sic] of King David's line of Judah.” Frequently posted in these communities are generative-AI-produced images of Trump with Jesus, including iconography that communicates his divine appointment.
In this evolving iteration of QAnon, interpretive practices honed in Bible study are interwoven with a conspiracist lens. Most of IET's 2021 Telegram activity was devoted to an asynchronous, online Bible study, in which he would transcribe Biblical verses, interpreting them for his followers. After one such post, he bracketed his Bible reading with:
The same evil tactics then are the same evil tactics now - their playbook hasn't changed in millennia. These satanic POS's are so evil and disgusting!! Be grateful that God woke you up for such a HISTORIC time as this…Pray for those still asleep… It's going to be rough for them as all comes out.
While evangelizing, he alludes to QAnon by referring to Satanic actors, linking them inextricably with Biblical truth. In addition, his rhetoric around being “awakened” refers to being “red-pilled” (a reference to the Matrix movie in which choosing the red pill awakens characters to the reality of their world) as well as “The Great Awakening,” when undeniable truth will “be rough” on those “still asleep.” In this way, he exemplifies Argentino (2020a) finding that “QAnon conspiracy theories are reinterpreted through the Bible. In turn, QAnon conspiracy theories serve as a lens to interpret the Bible itself” (para. 4).
Both Paster Chris and IET function in these spaces as QAnon evangelists; they spread the word of QAnon authorities, provide their own literary glosses on the revelations of the day, and function as what Pierre (2023) termed “Activists” on his typology of conspiracy belief—they have adopted the conspiracy worldview in its entirety and “have decided that they must act on beliefs in order to defend them or to fight for the cause” (p. 26).
4.1.2 QAnon and white supremacy
Although the overlap between right-wing politics, conspiracism, and white supremacy may seem new, it is as old as the founding of America (Barkun, 2003; Walker, 2013; Uscinski and Parent, 2014; Kendi, 2016; Cooke, 2021; Hirschbein and Asfari, 2023). What is new is how the superconspiracy QAnon subsumes and manifests these long-standing conspiracy theories serving to justify white power structures and systemic oppression. Sharlet (2023) examined the emerging overlap between Trump followers and conspiracy theorists, including those who adhere to the “great replacement theory,” driven by white anxiety and consisting of an ideology of scarcity, as BIPOC peoples and immigrants outnumber whites. For example, in his chapter “Whole Bottle of Red Pills,” Sharlet visited with men's rights activists; he observed how their ideology dovetails with racism, Trumpism, and QAnon (Sharlet, 2023). The people he interviewed espoused components of the “great replacement theory,” expanding racist claims to include men being ruled by feminism and women. Their rhetoric is palpably violent; such conspiracy ideation is linked to extremist violence, xenophobia, misogynoir, and Neo-Naziism (Rousis et al., 2020; Bailey, 2021; Gonzalez, 2022; Lovelace, 2022; Topor, 2022; Hernandez Aguilar, 2023). Indeed, the overlap between QAnon conspiracism and fascism cannot be ignored; for instance, Connor ties QAnon to “fascist movements dating back to the Nazis,” which “have a long history of co-opting popular cultural movements in order to gain membership, draw widespread attention, and to normalize talking about ideas couched in racist ideology (i.e., questions and debates over who really controls society are little more than anti-Semitic dog whistling)” (Conner, 2023, para. 6). Although their posts are not overtly white supremacist or anti-Semitic, dog whistles about “bankers,” “the Rothschilds,” and Holocaust denialism all point to a worldview that tolerates, if not actively encourages, darker suppositions. For example, in a post from May 4, 2021, since deleted, Pastor Chris shared a post from “Mr. Patriot” referencing the “hiding of God” through denial of the Flat Earth and commented, “Wow what I am learning about God Creator, earth, holocaust [sic].”
In an assessment of how the Australian manifestation of QAnon coalesced with nationalism and conspiracy theories targeting marginalized communities, Jones (2023) asserted that QAnon promoted violent activism and support for right-wing politicians to bring about “The Storm” (a moral and legal reckoning) and “The Great Awakening” (unavoidable reality to the masses). Similarly, scholars who research how white supremacy has shifted through time have traced how the proliferation of “The Big Lie” is linked to past propaganda such as Mein Kampf , with devastating consequences (Neely and Montañez, 2022, p. 13, n. 6). Meanwhile, Indigenous scholar and activist Julian Brave NoiseCat has drawn parallels between QAnon rhetoric and white colonizers' historical and ongoing oppression of Indigenous peoples (Remski and Brave NoiseCat, 2023). Brave NoiseCat posited that the fear of ritualized child kidnapping and abuse in the QAnon worldview reflects the historical reality of Indigenous children being forcibly removed from their homes to be indoctrinated in colonial boarding schools. In yet another example, the sublimation of the term “groomer” in QAnon ideology is tied to the historical and contemporary threats to LGBTQIA+ affiliated communities (Mogul et al., 2011; Martiny and Lawrence, 2023). IET perpetuates such threats through his promotion of the “groomer” rhetoric and affiliated conspiracy theories, even challenging his followers to argue with him by posting an unattributed definition of “groomer” and stating, “If your official and legally documented actions match this description then being called a groomer isn't libel or slander: it's a fact.” Unsurprisingly, given QAnon's mantras and its use of terms like “The Great Awakening” and “The Storm,” it has attracted and absorbed adherents whose views align with white supremacists, extrajudicial violence, and policies that target marginalized communities.
In our research that surveyed library workers encountering conspiracy theorists over the fall of 2021, participants described potent examples of conspiracy theories, often tied to QAnon, targeting ethnic, religious, and racial categories of people (Beene and Greer, 2023). Most common in the data was conspiracism surrounding Holocaust denial or other anti-Semitic conspiracies, a finding tied to scholarly discussions of conspiracy theories targeting Jews (Barkun, 2003; Hirschbein and Asfari, 2023). There were also instances of patrons adhering to conspiratorial plots perpetrated by the Chinese and Chinese-Americans, usually involving the COVID-19 pandemic. Participants described conspiracy ideation that targeted Black people; for instance, those conspiracy theories surrounding the murders of George Floyd and other unarmed Black men, the Black Lives Matter movement, and the NAACP. A subset of respondents noted conspiracy theories involving immigrants or Indigenous peoples linked rhetorically to the “great replacement theory.” These and affiliated conspiracies are now more enmeshed with QAnon and mainstream right-wing rhetoric (McFall-Johnsen and Morgan, 2021; Zihiri et al., 2022).
Of the two case studies, IET exemplifies the intermingling of anti-Semitism and white supremacy. For example, he references “Kek” while disseminating a conspiracy theory about former President Obama and his finances:
These people are evil.
They're also stupid.
We see everything
When you see it, post it in the comments
Also, what [sic] are they broke? Kek... Where did their money go?
But who, or what, is Kek? As the Roose (2020) explains, Kek “is bandied about on social media… where alt-right activists and avid Donald Trump supporters lurk” and is “a ‘deity' of a semi-ironic ‘religion' the white nationalist movement has created for itself online” (para. 1-2). Originating in gaming culture and spreading to the chan boards, Kek came to symbolize chaos and darkness and is invoked when signaling to others online that one is “awakened” and that destroying the world order is necessary to bring about an enlightened race. Kekistan is thus the mythical country that asserts a satirical ethnicity, most often affiliated with white supremacy; indeed, the flag of Kekistan is based on the Nazi war flag (Pepe the Frog, 2023). When IET reiterates to his followers to “keep bloodlines pure,” or when he refers to “Kek” or “Kekistan” and employs Pepe memes, he is signaling his bona fides in an explicitly racist, destructionist ideology.
4.1.3 QAnon, wellness culture, and COVID-19
The overlap between yoga and fitness industries, alternative health philosophies, and conspiracy theories has been well-documented in both news media and scholarly literature. For example, Conner (2023) analyzed the overlap between wellness, fitness and yoga instructors, finding that a “‘soft[er]' version of QAnon helped draw in those…wanting to get more physically fit, those seeking alternative forms of therapy when modern medicine has failed them, and still for others a search for both a community of likeminded individuals and a spirituality that resonated with them” (para. 34). Monroe (2017) discussed how essential oils became a cure-all for our “Age of Anxiety,” while Asprem and Dyrendal (2015) chronicle the confluence between conspirituality and conspiracy ideation, arguing that this confluence can be traced “deep into the history of Western esotericism” (abstract). So, while this coalescence may seem new, it is more likely that QAnon infiltrated an area already ripe with conspiracism (CBS News, 2021).
Social media has provided the gateway for QAnon ideology to seep into otherwise domestic spaces on platforms such as Instagram. Influencers promoting yoga, CrossFit, and alternative health have interwoven posts about child trafficking, for example, or used a hashtag associated with QAnon (e.g., #greatawakening). Indeed, the idea of an “awakening” can be deployed for New Age spirituality just as readily as it can for QAnon or Christianity (Meltzer, 2021; Conner, 2023). A casual user of social media, dabbling in New Ageism, yoga, or alternative health may receive targeted advertising or see pretty Instagram posts promoting QAnon and other conspiracy theories, thereby drawing them in (Chang, 2021; Baker, 2022; Hermanová, 2022). Such infiltration has not gone unnoticed in fitness circles. For instance, CrossFit distanced itself from the avid practitioner and congressional representative Marjorie Taylor Greene, who is also a QAnon evangelist (Brooks and Jamieson, 2021). Meanwhile, yoga instructors began speaking out about QAnon and other conspiracy theories infiltrating their communities (Roose, 2020; CBS News, 2021).
While anti-vaccination sentiment is not new, QAnon influencers capitalized on the outbreak of the COVID-19 pandemic, with online misinformation spreading to offline decisions and actions. Pastor Chris has been particularly vituperative about the COVID-19 vaccine, linking it with other conspiracies relevant to Big Pharma. Commenting on their own post from May 3, 2021, they appear to have copied content from “TheDonaldUS” Telegram account entitled “10 Reasons why President Trump had to approve these deaths” (excerpted below) that served to position pandemic deaths and COVID-19 vaccines (“the shot”) as part of a much larger plan:
With the approval of the shots (but placing realistic information in the sights of alternative news), the deep state shot was to remain Experimental and NOT MANDATORY (which was NOT part of their plan). In war, you sometimes have to choose between two bad options.
Considering the strategy of the globalists, the DJT and the military have chosen the path with the fewest casualties, namely to continue the facade of public approval of the toxic (but NOT OBLIGATORY) shot.
The alternative FAR would have been more deadly to all of us in many ways. If the Deep State had created an approved BINDING shot, it would have killed millions more.
Is this virus fictitious? If it cannot be separated and there is no scientific trace, does it exist? Only in the minds of the masses!!!
As evidenced by these two small communities spearheaded by IET and Pastor Chris, much of the recent scholarly attention has focused on how the QAnon worldview has unified suspicions against vaccinations or the capitalist inclinations of the pharmaceutical industry with a broader conspiracist mindset.
For IET, a certified chiropractor and CrossFit enthusiast, blending alternative health philosophies, New Ageism, Christianity, QAnon, and anti-vaccination rhetoric (especially as it pertains to the COVID-19 pandemic) seems to come easily. Whether this practice serves to promote his own businesses or beliefs or to share posts from other social media influencers and platforms, IET peddles conspiracist worldviews. Throughout his Telegram, one can find references to a global cabal pushing Big Pharma policies, valorization of resisting vaccinations, and promotion of alternative health solutions. He interacts and encourages his followers through comments and posts, reinforcing the concept of a community fighting the mainstream. Threaded between various political and scientific conspiracy theories are posts directly linking followers to QAnon. For instance, earlier this year, he shared an interview he conducted in July 2019 on Conspyre.tv, dubiously titled “The 2nd American Revolution,” in which he outlines “WHY [he] supports Donald Trump and WHY [he] believes in the Q Operation,” encouraging followers to watch and comment. This post received over 42,000 views and 600 interactions and comments. In this way, he provides a gateway for those followers who may not have heard of QAnon, or those who harbor doubts, while reinforcing a particular conspiracist worldview. Indeed, the comments from this post praise his research, referring to him and other community members as “true Patriots.”
Public health scholarship has long devoted research to combatting vaccine hesitancy, including how misinformation and conspiracy theories can hinder public health outcomes (Larson, 2020; Sobo, 2021; Enders et al., 2022; Farhart et al., 2022; Seddig et al., 2022). A few sources have explicitly discussed the commingling of the QAnon and anti-vaccination movements, highlighting that the ideologies behind both movements emphasize fighting big institutions and mainstream experts by doing one's own research to inform personal decisions (Dickinson, 2021; Inform, 2023). Such anti-science belief can “act as an affective epistemological hinge for other forms of conspiratorial thought” (Fitzgerald, 2022, p.8); once one accepts that a public health campaign is hiding something and working to subvert the truth, it is easier to then accept other scientific conspiracies (e.g., flat earth). Lewandowsky has devoted his research to this overlap, finding that conspiracist ideation and worldviews predict a broader rejection of science (Lewandowsky et al., 2013a).
4.1.4 Observed information practices, hermeneutics, and epistemologies
As creators on new platforms such as TikTok spin ever-more-fantastical suppositions, they are shared with excitement by both IET and Pastor Chris, with their followers accepting these new conspiracies as gospel. The phenomena of parasociality and cognitive authority function in these spaces to propel and reaffirm conspiracy beliefs. Logical fallacies abound, but specific examples stand out: the conjunction fallacy, the non sequitur fallacy, the divine fallacy, and false attribution. The conjunction fallacy presupposes that two or more conclusions are likelier than any of them on their own, defying the laws of probability, while the non sequitur fallacy is an illogical conclusion based on preceding statements or tangents in conversations. The divine fallacy was apparent whenever followers assumed causality from divine intervention or supernatural force, either because they did not know how to explain the phenomenon or because they were driven by faith to assume so. False attributions or falsely exaggerated correlations were constructed from the close reading of signs and communications, leading followers to see collusion where there was none and results that did not follow a straight causal path.
The information practices observed in these communities reaffirm the religiosity of these spaces, most notably, complete credulity and deference to prophecy or authority. Indeed, scholars have noted how hard-right Christian groups expect their followers to accept everything, with no questions asked (e.g., the literal truth of the Bible) – this same ideology has been applied to the Trumpist MAGA movement (D'Angelo, 2024; Hassan and Johnson, 2024). In Pastor Chris' feed, the most outlandish assertions are accepted without hesitation, such as planes running on air, cars running on water, and MedBeds that will make us all young and perfect. Followers are bullied for any sign of skepticism; they must accept everything Pastor Chris posts or they are removed and blocked from participating in the community. In one post from May 2023, Pastor Chris summarizes “intel” received from “Bruce,” one of the cognitive authorities they often reference, about the coming Nesara/Gesara wealth transfers, a financial conspiracy theory that has been integrated into QAnon. After reporting on Bruce's call, Pastor Chris exhorts followers to continue praying and believing to enact the coming “Heavenly blessing… It might also be at the time of… those days of disclosure and lockdowns,” referring to the original QAnon Storm. In this and other posts, logical fallacies abound alongside a cognitive authority (Bruce) acting as a substitute for truth.
In addition to logical fallacies, these communities exemplify echo chambers of emotions and confirmation biases. A post from IET in April 2023 shared one of his tweets that coalesces COVID-19 conspiracy theories, political conspiracy theories, and QAnon. In this post, he exhibited an exaggerated correlation of actual and imagined events, drumming up fervor among his readers:
Look at the list of current events now front and center habbening [sic]: 1. COVID jabs being removed from market 2. Hunter Biden investigation 3. Bio weapons [sic] in Ukraine 4. Biden financial dealings busted open 5. J6 fedsurrection [sic] 6. Fauci files 7. [clown emoji] IC agents being exposed. All at once.” Underneath this screenshot, he shared a meme of Pepe the Frog with glasses, reading a book entitled “1001 I Told You So's.”
This post was viewed 45,000 times, with over 750 reactions and 80 comments from followers. Comments ranged from disheartened resignation over the fact that “normies” are not “awake” and following the signs to those who added conspiracy theories to the list (too much acclaim from other followers). In a seamless blending of fact (Hunter Biden and January 6th investigations) with conspiracy theories (Biden financial dealings, Fauci files, COVID-19 vaccines, Ukraine bioweapons, and IC agents), IET exemplifies a faulty hermeneutics, mingling established record with supposition to weave a conspiratorial narrative.
In both communication groups, belief is prioritized over empirical evidence – it surpasses reality. Truth is relational, and belief is the only metric of import. Teleologism is evident as disparate theories are linked in a web of conspiracy. Followers prop up each other's comments and reactions, reinforcing the notion of a community fighting against a normative, mainstream, and false understanding of events.
4.1.5 Emotions in the echo chamber
These communities define authority differently than mainstream societal definitions. Those who follow QAnon and its more outlandish offshoots strongly reject “elitist” knowledge (i.e., academic information) as part of their worldview, so traditional epistemologies and pedagogies cannot function alone to dislodge the conspiracy worldview. In-group trust is central to QAnon's mantra, encouraging those who doubt or falter to hold steady: “Trust the plan.” For example, whenever a follower questioned a comment by a fellow follower, or worse, a post by the influencer, they were vehemently attacked or removed from the community. Those allowed to remain were treated with disdain, accused of being outsider trolls, or told they did not believe enough. Tribalism is thus reinforced: Are you with us or against us? Such shaming reinforces the behavior and belief in conspiracy ideation, as those who are there to fulfill psychological and affective needs will ensure that their behavior submits to group norms. Thus, various emotions promote conspiracy theorizing within these online communities, including the subgroups of QAnon that have splintered and spread across social media platforms.
Moreover, followers of conspiracy influencers like Pastor Chris and IET recognize them as cognitive authorities with whom they have developed a parasocial relationship, or psychological relationships between influencers and their audience; often driven by strong emotions, followers come to consider these personalities as friends. Emotions are powerful motivators for engagement with conspiracy theories, including anxiety, hopelessness, isolation, anger, and shock (Douglas et al., 2019; Green et al., 2023). In addition to negative emotions, positive affect has also been shown to subvert critical cognitive processing and influence credulity; when we are calm and happy, we are more trusting (Forgas, 2019). Such affective states drive users' information-seeking behaviors and their ability to effectively evaluate faulty information, providing epistemic closure. Since followers consider IET and Pastor Chris to be parasocial cognitive authorities, these influencers appear as savvy manipulators of their followers' emotions, using emojis and capital letters to call attention to potent ideas and themes. They encourage their followers and prolong their engagement, supporting positive affective states in their followers, and increasing followers' trust and confidence in the information IET and Pastor Chris provided. Some of this work is the work of a competent influencer. IET, for example, could also seek to prolong engagement with his followers in order to promote his businesses and sell products; while it is unclear how much he believes of what he posts, his work as an influencer includes a careful calculation for monetizating his worldview. Influencers such as IET have been researched by others who have surveyed QAnon, alternative health, and celebrity influencers (Bloom and Moskalenko, 2021; Yallop, 2021; Petersen, 2023).
Prolonging engagement not only serves the influencer, but the community as well, reinforcing a sense of solidarity. Meanwhile, arousing rhetoric or inflammatory posts keeps their followers engaged and enraged. This is typified in a post from May 2, 2023:
The real crash is the Kilbal system, they lose everything, but WE GAIN EVERYTHING! Never underestimate ur GOD!
This post from Pastor Chris received 31 comments, with most posting affirmative GIFs or “Amen/Hallelujah.” To one who expressed a loss of hope due to the extended timeline, Pastor Chris exhorted them to
think how tired the children who [sic] been trafficking and those who have been in hiding to help bring freedom to those children and to humanity! When u realize this then our wait is nothing compared to that! This is the real tribulation that is happening and we are about to come out into real Freedom from God the Source God the Creator!
This comment illustrated the closed-off trust system developed within these communities, which can make it extremely hard to break through. Any questioning or nudges from skeptics are met with disdain, counterpoints from the in-group, or, as in the comment above, appeals to emotion.
As additional conspiracies are subsumed into the QAnon movement and promulgated via the social web, QAnon could be considered an overarching superconspiracy (Beene and Greer, 2023). For instance, one of Pastor Chris's followers commented on the absolute trust that the community has in former president Trump to enact the financial blessings wrought by the Nesara/Gesara conspiracy:
This is the way I look at it, [sic] I don't think for 1 min that DJT would let us all lose our money while he rides around on his golf cart living large, [sic] there must be a plan. Do you honestly think that anyone would support him or the plan, if he let us all lose our money in the process of doing so? NO, all he has gained from the MAGA movement would be gone, poof, if we all went broke and lost our life savings.
This comment illustrates the vulnerability of the community members and their absolute trust in Trump, an affective state that the influencers exploit. Through actively encouraging the continued belief in things like all debt being wiped out, Pastor Chris and their ilk engender a hope that will only lead to greater disappointment or financial ruin as some choose to forego payments because their loans are seen as illegal, connected again to the global cabal, under the gospel of the conspiracy.
The conspiracy phenomenon of “MedBeds” serves as perhaps the most heartbreaking example of this exploitation. Often included as part of the supposed Nesara/Gesara reveals, MedBeds present a combination of conspiracy (hidden technologies), scientific illiteracy, and magical thinking. These beds, sometimes attributed to Elon Musk and Tesla, sometimes rumored to be military technology, have numerous benefits, as elucidated in a lengthy post widely shared among conspiracy communities throughout 2023. Illustrative of the magical thinking and misplaced hope of this particular conspiracy, enumerated among the purported abilities of the Medbeds was the regrowth of organs and teeth, the cure of all physical and mental ailments, changing hair and eye color at will, and “downloading languages” and other knowledge to one's brain.
Through constant encouragement and repeated appeals of their soon-to-be-announced availability, the conspiracy influencers prime the vulnerable for exploitation. Followers echo the influencers' overconfidence, enthusiasm, and excitement. Comments abound with people announcing they have stopped their medications, quit going to doctors, or lamenting that the MedBeds did not come soon enough for their loved ones.
4.1.6 (il)literacies
The following examples illustrate information literacy and other literacy gaps linked to information behaviors relevant to these groups. Educators and librarians can close these gaps by first recognizing them. Through an exploration of these communities' behaviors, we believe there are many opportunities for better education.
4.1.6.1 Visual and media literacy
Furthering visual and media literacies is a crucial endeavor for these communities, since much of the content community members share consists of unattributed and decontextualized images, viral memes, and questionable videos. Hodges (2023) summarized recent conspiracy theory literature findings that “present-day conspiracy narratives spread in large part through the networked circulation of decontextualized digital images, which accrue new narrative significance as they enter new contexts and arrangements, traveling further from their authentic sources and meanings over time” (p. 132). This networked circulation of decontextualized imagery does not just happen in conspiracy communities. Palmer (2019) analyzed the common practice in online media of reusing images, videos, or screenshots outside of their original contexts to reshape a narrative, a phenomenon he termed the intersemiotic contextual misrepresentation of photojournalism (p. 104). Sharing bits from mainstream media sources or dubious news outlets helps spread misinformation. This practice was evident in both samples. In 2022, IET posted a YouTube video from “We The Media (William Scott)” - “The Unvaccinated will be Vindicated,” From The Highwire, a dubious news outlet profiting from misinformation. This post, which paints the unvaccinated as heroes working to expose and vanquish the Deep State, was viewed by over 46,000 and garnered over 1,000 reactions and 100 comments. One commenter said, “Very powerful. And some people says [sic] Trump pushed this vaccine. Trump tricked the Deep state epic move,” while another commented that posting this video “got me 30 days in fakebook jail.” For that commenter, the “time out” forced upon them did not result in a critical analysis of the video and why it may have sparked the ban but rather reinforced the idea that the Deep State is out to get the truth-tellers. With each new edit and share of news media clips, the visual material takes on additional meanings and wilder interpretations.
The human mind processes and creates meaning from images far quicker than text (Palmer, 2019, p. 107); underscoring this reality, Palmer describes the image as “the juggernaut in the perceptual process” (p. 123). This cognitive reality shaped the Framework for Visual Literacy for Higher Education's (Association of College and Research Libraries, 2016) knowledge practices and dispositions, with knowledge practices including:
• Interpret visuals within their disseminated context by considering related information such as captions, credits, and other types of metadata.
• Question whether a visual could be considered authoritative or credible in a particular context, which can include comparing it to similar visuals, tracking it to its original source, analyzing its embedded metadata, and engaging in similar evaluative methods. (p. 8)
Multimodal media can present an even greater challenge for educators looking to debunk conspiracy theories, as the multiple channels of information can lend credibility to what otherwise could be doubted if the information was shared purely through one mode (e.g., image, text, audio, etc.) (Serafini, 2013, 2022; Campbell and Olteanu, 2023; Neylan et al., 2023). Neylan et al. (2023), writing for PubMed, examined techniques to inoculate online users against multimodal misinformation, including games and gamification; however, they found that multimodal misinformation is uniquely resilient to inoculation.
Both Pastor Chris and IET populate their feeds with multimodal imagery. For IET, sharing YouTube and TikTok videos is a powerful rhetorical tool that instantly communicates to followers. For example, in a Telegram post-dated 2 August 2022, IET shared a TikTok video, which quickly explains that Satan is driving illnesses and the pharmaceutical companies that profit from them, including through vaccines; near the end of the video, hospital logos are visually compared to Satanic symbols. This post garnered 163 comments, ranging from “I knew it!” to “I never noticed this, but it makes sense.” By sharing a popular form of online video, IET only needs to bracket it with his own rhetoric, which he does through a headline that says, “This world isn't what it seems, Let me explain,” and his one comment underneath: “Preach!” For Pastor Chris, images are often posted from the White House webcam, with additional textual glosses praising any movement or change in scenery as a clear sign that conspiracies are being revealed.
Thus, the power of the shared videos and images in these communities is enhanced through the influencers' commentary. Because of the challenges that occur with reading and evaluating multimodal imagery, Campbell and Olteanu (2023) advocate for a new post-digital literacy to combat mis- and disinformation spread through multimodal means; they argue that a new, embodied form of meaning-making is deeply embedded within the digital information ecosystem. A new framework for literacy is both necessary and compelling.
4.1.6.2 Digital literacy
Another area ripe for further investigation is the lack of digital literacy evident within these communities. While there is not one definition of digital literacy, strides are being made by educators and researchers (Tinmaz et al., 2022; Marín and Castaneda, 2023; Palmer, 2023). At its most basic, digital literacy entails understanding the online information environment – how information is created, shared, and repurposed through systems and platforms. In conspiracist communities, there is often a lack of understanding of how information travels and changes online. This is another departure from the early QAnon community, which was comprised of coders and gamers; newer conspiracist communities, such as the groups explored here, display more misunderstanding of how digital information is created and changed online.
A great example of this misunderstanding is Paster Chris' post from November, 2023, where they posted excitedly that the White House was up for sale. They followed up this post with another that included a screenshot from the real estate website Zillow that purportedly showed the White House up for sale (Figure 1).
Although clearly labeled “Off-market,” a standard stub placeholder listing on Zillow was enough proof for Pastor Chris and many of their followers. In comments, Pastor Chris elaborated: “I sense the white hat will sell it, get the money, put it into the treasury to get back to the people [sic] and then the rods of God will come along and destroy it. Not sure how it all will play out [sic] but you did see a Q drop about the white house for sale.” Commenters pointing out that it was off-market were met with rumors of it already being sold or teleological reasoning such as one follower's note that “yes, I seen [sic] it was off market. However, which third party placed it there? There appears to always be a reason for such events that reveal something… just what... Who knows?”
This example illustrates an all-too-common problem within conspiracy communities: users must understand the ever-growing and ever-sophisticated systems built to populate the internet with content. The conspiracy theory surrounding Wayfair, an online furniture company, provides another example: out-of-stock products like storage cabinets were assigned random names by the company, sometimes resembling human names. The QAnon community circulated rumors that the products were actually codes for children being trafficked and that the products were being used to store the children (Spring, 2020); the theory spread across social media, leading to digital warriors interrupting with and subverting actual child trafficking work (Contrera, 2021). Literacy training that includes basic knowledge about the purposes for online information and how it flows through systems would be beneficial.
4.1.6.3 Science literacy
Scientific illiteracy and conspiracism have been well-studied, especially among climate change scholars (Lewandowsky et al., 2013a,b; Scheufele and Krause, 2019; McIntyre, 2021). Scientific disinformation is rampant in these communities and reflects a lack of knowledge and a willingness to set aside one's knowledge for the more desirable “hidden truths.” For example, Pastor Chris shared a post in May 2023 from the group “PLANET X FOR REAL”:
I can't believe people are so blind they don't even see their freaking moon is doing a cartwheel all over the sky and rising and setting all over the place….Yet the dumb trolls/ NCPs in here will say yeah [sic] that's normal. Don't listen to people who don't even look up while they are too busy looking down [sic] sitting on their couch [sic] buying into all other distractions. Those are the same people who took the [emoji of a syringe]
The group's reference to “Planet X” refers to an actual hypothesized additional planet in the solar system (Kauffman, 2016); the group's posts, however, promote the conspiracy theory that Planet X and other celestial bodies have been hidden from the public. The post shared by Pastor Chris seems to reference a flat earth theory that states that the moon is hung in the sky above the earth in a sealed dome (Modern Flat Earth Beliefs, n.d.). At the end of the post, a jab at those vaccinated against COVID-19 illustrates how adherence to one scientific conspiracy theory quickly begets adherence to another.
The motivations for such beliefs are complicated, but like all others, they stem from affective states and psychological needs. Centralized, government-backed scientific institutions, researchers, and the academic elite are prime targets for mistrust. The scientific method is misunderstood, with changing scientific recommendations portrayed in these communities not as evidence of a growing body of knowledge but as evidence of malfeasance, fraud, or stupidity: “These idiots have no idea what they are doing.” This mindset was common during the COVID-19 pandemic. These communities promote their own scientific experts and cognitive authorities to counteract those that mainstream “normies” accept. For example, IET has promoted creators on alternative news podcasts, others on Telegram, and even creators on TikTok to draw attention to “what's really happening” and “the truth,” each instance of which is applauded and cheered by his followers, who sometimes comment that they, too, have been following these influencers for some time – alongside IET, of course.
Evident in IET's Telegram posts and Rumble interviews are many of the anti-pharmaceutical, anti-vaccation, and anti-science stances. Alongside a libertarian, individualist stance in making and facilitating one's own medical decisions is the promotion of debunked cures such as “parasite cleanses,” Ivermectin and hydroxychloroquine use, and assorted alternative health theories. Concurrent with these views is a hardened stance against scientific expertise, especially those who mandate or recommend COVID-19 vaccinations. These stances have been well-documented in the scholarly literature (Larson, 2020; Voinea et al., 2023). And while many of the fears and associated beliefs around vaccination are warranted, especially in marginalized communities with a distrust of authorities (Stein et al., 2021; Bilewicz, 2022; Cortina and Rottinghaus, 2022; Fredericks et al., 2022), conspiracy ideation also plays an outsized role (Mylan and Hardman, 2021; Enders et al., 2022; Farhart et al., 2022; Seddig et al., 2022). The consequences of distrust have many detrimental effects, notably on public health. For example, vaccine hesitancy has led to a resurgence of measles, which had been nearly eradicated in the United States (Orenstein et al., 2004). A 2020 report on a recent outbreak noted that 9% of the infected required hospitalization, with one death (Feemster and Claire, 2020). Fear of the COVID-19 vaccines, due to either previous vaccine hesitancy or fear of the new mRNA technology used in the vaccines, led to a number of conspiracies and lower vaccine uptakes, with one report estimating that the Delta variant was 11% more fatal for those who were unvaccinated (Dyer, 2021).
Therefore, we must strengthen a basic understanding of science and work toward a holistic scientific literacy – one that begins with understanding the scientific method and how scientific discourse evolves and ends with a fundamental insight into (and respect for) scientific authorities and authoritative bodies. As educators, the health of future generations depends on it.
5 Discussion: lessons for educators
Conspiracy adherents most likely lack the skills and knowledge to evaluate conspiracy claims critically, and many suffer from logical fallacies (e.g., correlation, not causation), teleogism, and other types of misinterpretation. As the well-known “Dunning-Kruger effect” (Kruger and Dunning, 1999) stipulates, those who do not have sufficient knowledge in a subject area will tend to overestimate their abilities and cannot recognize their cognitive and epistemic deficits until they have further training. The deficit model is not the only answer, although it does play a part in these communities. What seems to be happening in these influencer groups is a combination of literacy deficits, maladaptive heuristics, and, as Hodges (2023) phrased it, an “exaggerated interpretation of correlation.” Followers ignore facts that counter their belief systems, filling in the gaps in information with faulty interpretations to fit the desired outcome. The behavior observed in these two communities illustrates how a lack of information and affiliated literacies (e.g., visual, media, digital, etc.) leads to greater gullibility when faced with bad information (Mercier, 2021). An ignorance regarding the various systems of information production and dissemination has led to naïveté when faced with information that has been manipulated. Information professionals can play a role in helping people assess and evaluate information, leveraging the short snippets of time they have with them and focusing on engendering transferable critical thinking dispositions (Weiner, 2013).
With its impetus in the social, participative environment, metaliteracy provides necessary metacognitive prompts that can be leveraged to address many of the logical and literacy gaps seen in these communities. While those entrenched in the groups may be harder to reach, educators can better prepare those in classrooms or other learning environments, inoculating them against information-disordered environments (Compton, 2019). For example, Brungard and Klucevsek (2019) argue that the metaliteracy domains of affect and metacognition provide essential skills that promote critical evaluation of evidence and whether to use that information to produce original content.
Educators also need to prioritize visual and media literacy. Digital media scholar Palmer (2019) noted a “continuing bias oriented toward the written word over the image” (p. 132). Although he specializes in photojournalism and digital media, this idea holds in most educational settings. The current information environment is overwhelmingly visual, with users on primarily visual social media platforms such as TikTok, Snapchat, Instagram, and YouTube posting millions of images and hours of content every minute of the day (DOMO, 2022). Inculcating slow-looking practices (Thompson, 2019; Beene and Statton Thompson, 2022; Thompson and Beene, 2023), reflective practices, those that encourage a careful reading of an image, its original context, and associated content, will much better prepare learners for the daily glut of images that we all consume (Thompson et al., 2022).
Now that generative AI tools can author texts and images, slowing down to assess information critically is more important than ever. ChatGPT, for example, has been known to fabricate sources (Frosolini et al., 2023), while DALL-E and other image generators continue to train on and amalgamate copyrighted work from creators (Chayka, 2023; Chesterman, 2023). This election cycle of 2024 has seen an uptick in what are termed “cheap fakes” and “deep fakes” (Benson, 2023; Concha, 2023; Rascoe, 2023). It is unlikely that misinformation or AI will disappear anytime soon. Therefore, librarians and educators must work to engender a critical stance in their patrons and students.
In today's misinformation and disinformation environment, information professionals are quick to recommend fact-checking techniques and tools (Caulfield, 2017; Wineburg and McGrew, 2017; Baer and Kipnis, 2023; Blechinger, 2023). As appealing as fact-checking is, these techniques may not always work, especially with conspiracy ideation. Two psychological reactions have been suggested for why conspiracy theorists so heartily reject fact-checking. Literature has referred to the “backfire effect,” which describes a person who becomes even more entrenched in their original position when confronted with facts, especially when emotionally or ideologically invested (Brotherton, 2015, p. 233; Swire-Thompson et al., 2022). Lewandowsky et al. (2020) noted, however, that concerns about the prevalence of the backfire effect do not seem to be supported by research (p. 9). What may actually be occurring is the “sunk-cost fallacy:” As people are confronted with information that contradicts their worldview, they experience cognitive dissonance and instead”choose to throw themselves deeper into the ideology and ignore anything that contradicts it” (McCaffee, 2020, p. 30). Another is the “continued influence effect,” which describes a continued reliance on inaccurate information and reasoning, even after a credible correction has been offered (Lewandowsky et al., 2012, 2020, p. 5). Thus, strict fact-checking is usually not a strong enough motivator to abandon conspiracy beliefs.
When attempting to correct wrong information, it is vital to assess where a person might be on the spectrum of ideation, from those just beginning their exploration into a conspiracy theory to the devoted conspiracy adherent who rejects all authoritative and mainstream accounts (Pierre, 2020, para. 10). While healthy skepticism encourages us to evaluate critically, global skepticism leads to a suspicious mindset. Instead, we should aim for cognitive flexibility, which embraces open-mindedness and acknowledges where we might be wrong (Pierre, 2020, para. 11).
Importantly, many of these methods described here only work when some sort of trusting relationship has been established. Known as “nudging” someone toward critical thinking (Thaler and Sunstein, 2009), librarians and educators can practice active listening, prioritizing the relationship and building trust before questioning someone's stance and inserting opposing viewpoints (Beene and Greer, 2021, 2023). Research on the “stickiness” of other conspiracy theories, such as climate change or the various theories associated with the COVID-19 pandemic, provides some strategies for how we might nudge people away from deeply held beliefs. This shift will be a slow process; as Swanson (2006) noted, “One of the inescapable facts of personal epistemology is that information, knowledge, and learning are deeply intertwined with belief” (p. 98).
We can create meaningful connections with patrons and learners by centering our own vulnerability. Thoughtful, conscious reflection allows us to slow down and respond to the content of a claim and the emotions that may be fueling it. “Debunking” and “prebunking” are methods that model metacognition for students and patrons. The Debunking Handbook by Lewandowsky et al. (2020) proposed a more robust formula than just fact-checking to debunk climate change myths. It includes:
1. Start with a fact,
2. Warn that a myth is coming
3. Explain how the myth misleads
4. Repeat the fact (pp. 12–15)
The debunking formula can be adapted to fit any mis- or disinformation. The authors cautioned that this will not be a one-time process; establishing a relationship with the conspiracy theorist will be most helpful. The second step outlined in the debunking process, “warn that a myth is coming,” stems from the psychological inoculation theory (Banas and Rains, 2010; Barbati et al., 2021; Maertens et al., 2021). Inoculation against misinformation, similar to medical inoculation, creates immunity before exposure. A misinformation inoculation typically includes two components: a threat or a warning, and prebunking, or valid information that counteracts the threat (Compton, 2013; van Der Linden et al., 2020). The COVID-19 pandemic provided numerous tests for this theory (van Bavel et al., 2020; van der Linden et al., 2021), with research indicating that utilizing inoculation theory to prebunk misinformation can “not only help people to identify fake news easily, but it could also change their attitudes and behavior to rely on more trusted information in the public health area” (Kassen, 2023, p. 5). A recent study extended these results into a study of how inoculation worked against different misinformation types on social media, with similarly positive findings (McPhedran et al., 2023). Prebunking in practice can be explored through online resources like the Bad News Game, Harmony Square, or Cranky Uncle, but should not be relied upon entirely to build resistance. Although research has shown these gamified interventions do reduce susceptibility to misinformation, recent studies suggest that the inoculations also reduce belief in valid information, so they should be used with caution (Basol et al., 2020; Graham et al., 2023; Kassen, 2023; Modirrousta-Galian and Higham, 2023).
One solution to countering conspiracy ideation lies in shifting our thinking about conspiracy ideation from it being a problem of faulty thinking to a problem of faulty interpretation. Hartman-Caverly (2019) has argued that what some see as an epistemic crisis is a legitimate response to actual historical conspiracies, failures of traditional media, and information gatekeeping that “insulate[s] established institutions from legitimate critiques” (Hartman-Caverly, 2022, p. 456). Instead of the dichotomy of good vs. bad information, Hartman-Caverly encourages readers to acknowledge their own epistemic fallibility and endorse a more complex approach to information use and evaluation that relies on epistemic virtues:
Responsibilist epistemic virtues include practices like appropriate skepticism, attentiveness to evidence, awareness of fallibility, conscientiousness, curiosity, disinterestedness, fair-mindedness, impartiality, knowledgeableness, objectivity, open-mindedness, patience, and rigor, which manifest in behaviors like conscious reflection, deliberation, and justified belief endorsement within a community (2019a, p. 457).
We must do more to acknowledge the affective states that drive participation in these communities. Often, community members have only each other; comments from posts such as “my family thinks I am kukoo [sic]” or discussing how family members have stopped talking to them illustrate the growing isolation and reliance on the community that affirms their beliefs. As a result, relationship-building and empathy are critical to any discussion of disinformation subversion or prevention. Weil (2022) described this in her book on flat earth theorists, writing that “(conspiracy theories) thrive when we mistrust others and wall ourselves behind our personal paranoias. Fittingly, one of the best ways out of conspiratorial thinking can be to place our faith in others” (p. 214).
6 Conclusion
The examples presented in this article showcase the evolution of QAnon communities, which parallels the affordances of social media platforms. Whether through Facebook, Telegram, or TikTok, these platforms encourage users to actively participate in constructing a narrative, aiding in the spread of mis- and disinformation. In 2020, Americans' awareness of QAnon doubled, from 23% to 47% (Pew Research Center, 2020). By 2023, roughly 25% believed in one or more of the core tenets of QAnon: that the coming “Storm” would restore leaders such as former President Trump to power; “violence may be necessary to save the country”; or that “the government, media, and financial worlds are controlled by Satan-worshipping pedophiles” [Public Religion Research Institute (PRRI), 2023, “QAnon Beliefs”]. With more Americans turning to alternative news (Stocking et al., 2022) and social media platforms like TikTok (Matsa, 2023), it is imperative that educators and information professionals update their understanding of the types and range of literacies necessary to navigate the current and evolving online information ecosystem.
Data availability statement
The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.
Ethics statement
Written informed consent was not obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article because the data is from a publicly available social media profile.
Author contributions
KG: Conceptualization, Formal analysis, Investigation, Methodology, Writing – original draft, Writing – review & editing. SB: Conceptualization, Investigation, Methodology, Writing – original draft, Writing – review & editing.
Funding
The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This research was funded by a grant from the Institute for Information Literacy at Purdue University.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Footnotes
1. ^Name has been changed
References
Amarasingam, A., Argentino, M. A., Johnston, D., and Mosurinjohn, S. (2023). “Categorizing QAnon: is this a new religious movement?,” in The Social Science of QAnon: A New Social and Political Phenomenon, ed M. K. Miller (Cambridge: Cambridge University Press), 271–290.
Andre, C. (2023). “The religious rhetoric of QAnon,” in QAnon, Chaos, and the Cross: Christianity and Conspiracy Theories, eds M. W. Austin and G. L. Bock (Grand Rapids, MI: Wm. B. Eerdmans Publishing Co), 189–204.
Anti-Defamation League (n.d.). Inevitable ET or IET | Center on Extremism. Anti-Defamation League. Available online at: https://extremismterms.adl.org/glossary/inevitable-et-or-iet (accessed February 15 2024).
Argentino, M. A. (2020a). The Church of QAnon: Will Conspiracy Theories Form the Basis of a New Religious Movement? The Conversation. Available online at: https://theconversation.com/the-church-of-qanon-will-conspiracy-theories-form-te-basis-of-a-new-religious-movement-137859 (accessed May 18, 2020).
Argentino, M. A. (2020b). QAnon Conspiracy Theory Followers Step out of the Shadows and May be Headed to Congress. The Conversation. Available online at: https://theconversation.com/qanon-conspiracy-theory-followers-step-out-of-the-shadows-and-may-be-headed-to-congress-141581(accessed July 8, 2020).
Asprem, E., and Dyrendal, A. (2015). Conspirituality reconsidered: How surprising and how new is the confluence of spirituality and conspiracy theory? J. Contemp. Relig. 30, 367–382. doi: 10.1080/13537903.2015.1081339
Association of College and Research Libraries. (2016). Framework for Information Literacy for Higher Education. Available online at: http://www.ala.org/acrl/standards/ilframework (accessed February 15, 2024).
Badke, W. B. (2015). “Expertise and authority in an age of crowdsourcing,” in Not Just Where to Click : Teaching Students How to Think about Information, ed T. Swanson (Chicago: Association of College and Research Libraries), 191–215.
Baer, A., and Kipnis, D. (2023). Navigating Online Information Spaces With Lateral Reading: Lessons Learned From Two Librarians Working With Students and Educators, Libraries Scholarship 47. Avaialable online at: https://rdw.rowan.edu/lib_scholarship/47 (accessed February 15, 2024).
Bailey, M. (2021). Misogynoir Transformed: Black Womens Digital Resistance. New York, NY: NYU Press.
Baker, S. A. (2022). Alt. health influencers: how wellness culture and web culture have been weaponised to promote conspiracy theories and far-right extremism during the COVID-19 pandemic. Eur. J. Cult. Stu. 25, 3–24. doi: 10.1177/13675494211062623
Banas, J. A., and Rains, S. A. (2010). A meta-analysis of research on inoculation theory. Commun. Monographs 77, 281–311. doi: 10.1080/03637751003758193
Barbati, J. L., Rains, S. A., Ivanov, B., and Banas, J. A. (2021). Evaluating classic and contemporary ideas about persuasion resistance in inoculation theory: argument strength, refutation strength, and forewarning. Commun. Res. Rep. 38, 272–281. doi: 10.1080/08824096.2021.1956450
Barkun, M. (2003). A Culture of Conspiracy: Apocalyptic Visions in Contemporary America. Berkeley, CA: University of California Press.
Basol, M., Roozenbeek, J., and Van der Linden, S. (2020). Good news about bad news: Gamified inoculation boosts confidence and cognitive immunity against fake news. J. Cognit. 3, 91. doi: 10.5334/joc.91
Beene, S., and Greer, K. (2021). A call to action for librarians: countering conspiracy theories in the age of QAnon. The J. Acad. Libr. 47, 102292. doi: 10.1016/j.acalib.2020.102292
Beene, S., and Greer, K. (2023). Library workers on the front lines of conspiracy theories in the US: one nationwide survey. Ref. Serv. Rev. 51, 251–272. doi: 10.1108/RSR-11-2022-0056
Beene, S., and Statton Thompson, D. (2022). Focusing on slow looking: an exploration of techniques to develop critical observation habits. Art Docum. 41, 1–18. doi: 10.1086/719405
Benson, T. (2023). Brace Yourself for the 2024 Deepfake Election. The Wired. Available online at: https://www.wired.com/story/chatgpt-generative-ai-deepfake-2024-us-presidential-election/ (accessed February 16, 2024).
Berkowitz, R. (2021). QAnon Resembles the Games I Design. But for Believers, There is no Winning. Washington Post. Available online at: https://www.washingtonpost.com/outlook/qanon-game-plays-believers/2021/05/10/31d8ea46-928b-11eb-a74e-1f4cf89fd948_story.html (accessed May 11, 2021).
Bilewicz, M. (2022). Conspiracy beliefs as an adaptation to historical trauma. Curr. Opin. Psychol. 12, 101359. doi: 10.1016/j.copsyc.2022.101359
Blechinger, J. (2023). Reflections on information literacy in the ChatGPT era. Pathfinder 4, 163–172. doi: 10.29173/pathfinder78
Bloom, M., and Moskalenko, S. (2021). Pastels and Pedophiles: Inside the Mind of QAnon. Redwood City, CA: Redwood Press.
Brashier, N. M. (2022). Do conspiracy theorists think too much or too little? Curr. Opin. Psychol. 49, 101504. doi: 10.1016/j.copsyc.2022.101504
Brooks, R., and Jamieson, A. (2021). CrossFit is Finally Fed up with Marjorie Taylor Greene. BuzzFeed News. Available online at https://www.buzzfeednews.com/article/ryancbrooks/marjorie-taylor-greene-qanon-crossfit (accessed March 11, 2021).
Brotherton, R. (2015). Suspicious Minds: Why We Believe Conspiracy Theories. New York, NY: Bloomsbury Publishing.
Brotherton, R., and French, C. C. (2015). Intention seekers: Conspiracist ideation and biased attributions of intentionality. PLoS ONE 10, e0124125. doi: 10.1371/journal.pone.0124125
Brungard, A. B., and Klucevsek, K. (2019). “Constructing scientific literacy through metaliteracy,” in Metaliterate Learning for the Post-Truth World, eds T. P. Mackey and T. E. Jacobson (Chicago, IL: ALA Neal-Schuman), 103–142.
Butters, A. M. (2022). (Dis)Belief in QAnon: Competing Hermeneutics in the 2020 U.S. Presidential Election. Available online at: https://www.utupub.fi/handle/10024/173603 (accessed February 15, 2024).
Campbell, C., and Olteanu, A. (2023). The Challenge of postdigital literacy: extending multimodality and social semiotics for a new age. Postdigital Sci. Educ. 22, 1–23. doi: 10.1007/s42438-023-00414-8
Caulfield, M. (2017). Web Literacy for Student Fact-Checkers. Vancouver, WI: Washington State University Vancouver.
CBS News (2021). How Conspiracy Theories Infiltrated the Wellness Community. Available online at: https://www.cbsnews.com/news/covid-19-vaccine-conspiracy-theories-online-wellness-communities/ (accessed September 23, 2021.).
Chandler, K. J. (2020). Where we go 1 we go all: a public discourse analysis of QAnon. McNair Scholars Res. J. 13, 4.
Chang, C. (2021). The Unlikely Connection Between Wellness Influencers and the Pro-Trump Rioters. Cosmopolitan. Available online at: https://www.cosmopolitan.com/health-fitness/a35056548/wellness-fitness-influencers-qanon-conspiracy-theories/ (accessed January 12 2021.).
Chayka, K. (2023). Is A.I. Art Stealing From Artists? The New Yorker. Available online at: https://www.newyorker.com/culture/infinite-scroll/is-ai-art-stealing-from-artists (accessed February 10, 2023).
Chesterman, S. (2023). Good Models Borrow, Great Models Steal: Intellectual Property Rights and Generative AI. NUS Law Working Paper No. 2023/025. doi: 10.2139/ssrn.4590006
Compton, J. (2013). “Inoculation theory,” in The SAGE Handbook of Persuasion: Developments in Theory and Practice, eds J. P. Dillard and L. Shen (London: Sage Publications, Inc), 220–236.
Compton, J. (2019). “Inoculation theory and the metaliterate learner,” in Metaliterate Learning for The Post-Truth World, eds T. P. Mackey and T. E. Jacobson (Chicago, IL: ALA Neal-Schuman), 63–80.
Concha, J. (2023). The Impending Nightmare That AI Poses for Media, Elections. The Hill. Available online at: https://thehill.com/opinion/technology/3964141-the-impending-nightmare-that-ai-poses-for-media-elections/ (accessed April 23 2023).
Conner, C. T. (2023). QAnon, authoritarianism, and conspiracy within American alternative spiritual spaces. Front. Sociol. 8, 1136333. doi: 10.3389/fsoc.2023.1136333
Contrera, J. (2021). A QAnon con: How the Viral Wayfair Sex Trafficking Lie Hurt Real Kids. Washington Post. Available online at: https://www.washingtonpost.com/dc-md-va/interactive/2021/wayfair-qanon-sex-trafficking-conspiracy/ (accessed December 16, 2021).
Cooke, N. A. (2021). Tell me Sweet Little Lies: Racism as a Form of Persistent Malinformation, Project Information Literacy. Available online at: https://projectinfolit.org/pubs/provocation-series/essays/tell-me-sweet-little-lies.html (accessed August 11, 2021).
Cortina, J., and Rottinghaus, B. (2022). Conspiratorial thinking in the Latino community on the 2020 election. Res. Polit. 9, 20531680221083535. doi: 10.1177/20531680221083535
Crossley, J. (2021). What (exactly) is Discourse Analysis? GradCouch. Available online at: https://gradcoach.com/discourse-analysis-101/ (accessed June 21, 2021).
Cuicksilver (2021). R/Qult_Headquarters: VICE Reveals Anti-Semitic Q Promoter IETs Identity as Denver Chiropractor/Physical Therapist Craig Longley. Reddit. Available online at: https://www.reddit.com/r/Qult_Headquarters/comments/n0mnch/vice_reveals_antisemitic_q_promoter_iets_identity/ (accessed February 15, 2024).
D'Angelo, C. (2024). Extremism Is Rampant In This Part Of The U.S. —And Theres A Toxic Reason For It. HuffPost. Available online at: https://www.huffpost.com/entry/true-west-book-interview-betsy-gaines-quammen_n65a7e37be4b041f1ce644683 (accessed January 18, 2024).
Dentith, M. R. X. (2018). Expertise and conspiracy theories. Soc. Epistemol. 32, 196–208. doi: 10.1080/02691728.2018.1440021
Dickinson, T. (2021). How the Anti-Vaxxers Got Red-Pilled. Rolling Stone. Available online at: https://www.rollingstone.com/culture/culture-features/qanon-anti-vax-covid-vaccine-conspiracy-theory-1125197/ (accessed February 10, 2021).
DOMO (2022). Data Never Sleeps 10, 0. Available online at: https://www.domo.com/data-never-sleeps# (accessed February 15, 2024).
Douglas, K. M., Sutton, R. M., and Cichocka, A. (2019). “Belief in conspiracy theories: Looking beyond gullibility,” in The Social Psychology of Gullibility: Fake News, Conspiracy Theories, and Irrational Beliefs, eds J. P. Forgas and R. F. Baumeister (London: Routledge), 61–76.
Dyer, O. (2021). COVID-19: Unvaccinated face 11 times risk of death from delta variant, CDC data show. BMJ 374:2282. doi: 10.1136/bmj.n2282
Ellis, R. D. (2015). Moral hermeneutics, coherence epistemology, and the role of emotion. Kilikya 2, 24–36. doi: 10.5840/kilikya20152315
Enders, A., Farhart, C., Miller, J., Uscinski, J., Saunders, K., Drochon, H., et al. (2022). Are republicans and conservatives more likely to believe conspiracy theories? Political Behav. doi: 10.1007/s11109-022-09812-3
Farhart, C. E., Douglas-Durham, E., Trujillo, K. L., and Vitriol, J. A. (2022). Vax attacks: how conspiracy theory belief undermines vaccine support. Prog. Mol. Biol. Transl. Sci. 188, 135–169. doi: 10.1016/bs.pmbts.2021.11.001
Feemster, K.risten A., and Claire, S. (2020). Resurgence of measles in the United States: How did we get here? Curr. Opin. Pediatrics 32, 139–144. doi: 10.1097/MOP.0000000000000845
Fisher, M. (2022). How Democracy Is Under Threat Across the Globe. The New York Times. Available online at: https://www.nytimes.com/2022/08/19/world/democracy-threat.html (accessed August 19, 2022).
Fitzgerald, J. (2022). Conspiracy, anxiety, ontology: theorising QAnon. First Monday 27, 12618. doi: 10.5210/fm.v27i5.12618
Flomberg, D. (2021). The Shomer: Denver Chiropractor Exposed As Central QAnon Conspirator. Colorado Times Recorder. Available online at: https://coloradotimesrecorder.com/2021/05/the-shomer-denver-chiropractor-exposed-as-central-qanon-conspirator/36141/ (accessed May 3, 2021.)
Forgas, J. P. (2019). “On the role of affect in gullibility: Can positive mood increase, and negative mood reduce credulity?,” in The Social Psychology of Gullibility: Fake News, Conspiracy Theories, and Irrational Beliefs, eds J. P. Forgas and R. F. Baumeister (London: Routledge), 179–197.
Fredericks, B., Bradfield, A., McAvoy, S., Ward, J., Spierings, S., Combo, T., et al. (2022). The burden of the beast: countering conspiracies and misinformation within indigenous communities in Australia. M/C J. 25, 2862. doi: 10.5204/mcj.2862
Frosolini, A., Gennaro, P., Cascino, F., and Gabriele, G. (2023). In reference to role of chat GPT in public health, to highlight the AIs incorrect reference generation. Annal. Biomed. Eng. 51, 2120–2122. doi: 10.1007/s10439-023-03248-4
Gabriel Gatehouse (February 22 2022). Epilogue, episode 8. “The Coming Storm on Apple Podcasts”. BBC Radio 4. Available online at: https://podcasts.apple.com/gb/podcast/the-coming-storm/id1601195264 (accessed January 11, 2024).
Garber, P. (2022). GhostEzra and the Tip of the Iceberg of Anti-Semitism. Medium. Available online at: https://philgarber.medium.com/ghostezra-and-the-tip-of-the-iceberg-of-anti-semitism-3c69c7bfb1a7 (accessed April 30, 2022).
Garry, A., Walther, S., Rukaya, R., and Mohammed, A. (2021). QAnon conspiracy theory: examining its evolution and mechanisms of radicalization. J. Deradic. 26, 26.
Gilbert, D. (2021a). Ive Been Investigating IET, One of QAnons Biggest — and Most Anti-Semitic —Influencers for the Last Couple of Months. X. https://twitter.com/daithaigilbert/status/1387477863352475649?lang=en (accessed April 28, 2021).
Gilbert, D. (2021b). This CrossFit Chiropractor Is Secretly One of QAnons Leading Influencers. Vice. Availalble online at: https://www.vice.com/en/article/7kvzmq/this-crossfit-chiropractor-is-secretly-one-of-qanons-leading-influencers (accessed April 28, 2021).
Gilbert, D. (2023). The QAnon Shaman Is Out of Prison and Selling Yoga Leggings Now. Vice. Availalble online at: https://www.vice.com/en/article/4a3xm9/qanon-shaman-out-of-jail-selling-leggings (accessed May 30, 2021).
Gonzalez, N. (2022). Reproducing for the Race: Eugenics, Race Suicide, and the Origins of White Replacement Conspiracy Theories, 1882-1924 [Ph.D. Thesis]. Davis, CA: University of California, Davis.
Graham, M. E., Skov, B., Gilson, Z., Heise, C., Fallow, K. M., Mah, E. Y., and Lindsay, D. S. (2023). Mixed News about the Bad News Game. J. Cognit. 6, 324. doi: 10.5334/joc.324
Green, R., Trella, C., Biddlestone, M., Douglas, K. M., and Sutton, R. M. (2023). “Psychological motives of QAnon followers,” in The Social Science of QAnon: A New Social and Political Phenomenon, ed M. K. Miller (Cambridge: Cambridge University Press), 33–48.
Hartman-Caverly, S. (2019). “TRUTH always wins: dispatches from the information war,” in We Have a Problem with Political Diversity: Discourse on the Margins of Scholarly Communities London: Association of College and Research Libraries, 187–233.
Hartman-Caverly, S. (2022). Long tail metaphysics: the epistemic crisis and intellectual freedom. IFLA J. 48, 449–465. doi: 10.1177/03400352211057146
Hassan, S., and Johnson, A. (2024). ?Breaking the Silence: Alva Johnsons Journey in the Cult of Trump as the 2016 Campaigns Director of Outreach and Coalitions. On The Influence Continuum with Dr. Steven Hassan Podcast. Available online at: https://podcasts.apple.com/us/podcast/the-influence-continuum-with-dr-steven-hassan/id1603773245 (accessed January 8, 2024).
Hermanová, M. (2022). Sisterhood in 5D: conspirituality and instagram aesthetics. M/C J. 25, 2875. doi: 10.5204/mcj.2875
Hernandez Aguilar, L. M. (2023). Memeing a conspiracy theory: on the biopolitical compression of the great replacement conspiracy theories. Ethnography. doi: 10.1177/14661381221146983. [Epub ahead of print].
Hirschbein, R., and Asfari, A. (2023). Jews and Muslims in the White Supremacist Conspiratorial Imagination. London: Routledge.
Hodges, J. A. (2023). Taxonomizing information practices in a large conspiracy movement: Using early QAnon as a case study. Inf. Cult. 58, 129–144. doi: 10.7560/IC58201
Inform (2023). Online Cultic Milieus: QAnon and Anti-Vax Conspiracy Movements. Global Network on Extremism and Technology. Available online at: https://gnet-research.org/2023/08/07/online-cultic-milieus-qanon-and-anti-vax-conspiracy-movements/ (accessed August 7, 2023).
Innes, H., and Innes, M. (2021). De-platforming disinformation: conspiracy theories and their control. Inf. Commun. Soc. 26, 1262–1280. doi: 10.1080/1369118X.2021.1994631
Jones, C. (2023). We the people, not the sheeple: QAnon and the transnational mobilisation of millennialist far-right conspiracy theories. First Monday 28, 1285. doi: 10.5210/fm.v28i3.12854
Kaplan, A. (2022). QAnon Community Celebrates the Return of Q After an 18-month Absence. Media Matters for America. Available online at: https://www.mediamatters.org/qanon-conspiracy-theory/qanon-community-celebrates-return-q-after-18-month-absence (accessed June 28, 2022).
Kaplan, T. (2023). “QAnon and social media,” in The Social Science of QAnon: A New Social and Political Phenomenon, ed M. K. Miller (Cambridge: Cambridge University Press), 271–290.
Kassen, M. (2023). Curbing the COVID-19 digital infodemic: strategies and tools. J. Pub. Health Policy 12, 1–15. doi: 10.1057/s41271-023-00437-2
Kauffman, M. (2016). The Case Strengthens for Planet 9 | News | Astrobiology. Available online at: nasa.gov (accessed September 2, 2016).
Kendi, I. X. (2016). Stamped from the Beginning: The Definitive History of Racist Ideas in America. New York, NY: Bold Type Books.
Kruger, J., and Dunning, D. (1999). Unskilled and unaware of it: how difficulties in recognizing ones own incompetence lead to inflated self-assessments. J. Pers. Soc. Psychol. 77, 1121–1134. doi: 10.1037//0022-3514.77.6.1121
LaFrance, A. (2020). The Prophecies of Q: American Conspiracy Theories are eNTering a Dangerous New Phase. The Atlantic. Available online at: https://www.theatlantic.com/magazine/archive/2020/06/qanon-nothing-can-stop-what-is-coming/610567/ (accessed June 23, 2016).
Larson, H. J. (2020). Stuck: How Vaccine Rumors Start – and Why They Dont Go Away. Oxford: Oxford University Press.
Lewandowsky, S., Cook, J., Ecker, U., Albarracín, D., Kendeou, P., Newman, E. J., et al. (2020). Debunking Handbook. Fairfax, VA: George Mason University Center for Climate Change Communication.
Lewandowsky, S., Ecker, U. K., Seifert, C. M., Schwarz, N., and Cook, J. (2012). Misinformation and its correction: continued influence and successful debiasing. Psychol. Sci. Pub. Interest 13, 106–131. doi: 10.1177/1529100612451018
Lewandowsky, S., Gignac, G. E., and Oberauer, K. (2013a). The Role of conspiracist ideation and worldviews in predicting rejection of science. PLoS ONE 8, e75637. doi: 10.1371/journal.pone.0075637
Lewandowsky, S., Oberauer, K., and Gignac, G. E. (2013b). NASA faked the moon landing—Therefore, (climate) science is a hoax: an anatomy of the motivated rejection of science. Psychol. Sci. 24, 622–633. doi: 10.1177/0956797612457686
Liebers, N., and Holger, S. (2019). Parasocial interactions and relationships with media characters- an inventory of 60 years of research. Commu. Res. Trends 38, 4.
LightWarrior (2021). Qanon and Antisemitism and Colorado Chiropractor, Oh My. Daily Kos. Available online at: https://www.dailykos.com/stories/2021/4/30/2028402/-Qanon-and-Antisemitism-and-Colorado-Chiropractor-Oh-My (accessed April 30 2021).
Lloyd, A. (2010). Information Literacy Landscapes: Information Literacy in eduCation, Workplace and Everyday Contexts. England: Chandos Publishing.
Lovelace, H. T. (2022). Xenophobic conspiracy theories and the long roots of January sixth. Law Contemp. Probs. 85, 19.
Lundsko, G., and MacMillen, S. L. (2023). QAnon and Other Replacement Realities. Lanham, MD: Lexington Books.
Mac, R., and Lytvynenko, J. (2020). Facebook Has Banned QAnon. BuzzFeed News. Available online at: https://www.buzzfeednews.com/article/ryanmac/facebook-bans-qanon (accessed October 6, 2020).
Mackey, T. P., and Jacobson, T. E. (2011). Reframing information literacy as a metaliteracy. College Res. Libr. 72, 62–78. doi: 10.5860/crl-76r1
Mackey, T. P., and Jacobson, T. E. (2018). Metaliterate Learning for the Post-Truth World. Chicago, IL: ALA Neal-Schuman.
Maertens, R., Roozenbeek, J., Basol, M., and van der Linden, S. (2021). Long-term effectiveness of inoculation against misinformation: three longitudinal experiments. J. Exp. Psychol. Appl. 27, 1–16. doi: 10.1037/xap0000315
Marín, V. I., and Castaneda, L. (2023). Developing Digital Literacy for Teaching and Learning. Handbook of Open, Distance and Digital Education. Singapore: Springer Nature Singapore, 1089–1108.
Marko, K. (2022). Extremist language in anti-COVID-19 conspiracy discourse on Facebook. Critical Discourse Stu. 22, 1–20. doi: 10.1080/17405904.2022.2110134
Martiny, C., and Lawrence, S. (2023). A Year of Hate: Anti-Drag Mobilization Efforts Targeting LGBTQ+ People in the US (Institute for Strategic Dialogue). Available online at: https://www.isdglobal.org/wp-content/uploads/2023/06/Anti-drag-Mobilization-Efforts-Targeting-LGBTQ-People-in-the-US.pdf (accesssed June 6, 2023).
Matsa, K. E. (2023). More Americans are Getting News on TikTok, Bucking the Trend Seen on Most Other Social Media Sites. Pew Research Center. Available online at: https://www.pewresearch.org/short-reads/2023/11/15/more-americans-are-getting-news-on-tiktok-bucking-the-trend-seen-on-most-other-social-media-sites/ (accessed November 15, 2023).
McCaffee, D. G. (2020). The Curious Persons Guide to Fighting Fake News. New York, NY: Pitchstone Publishing.
McFall-Johnsen, S. N., and Morgan, M. (2021). Several Groups of Extremists Stormed the Capitol on Wednesday. Here are Some of the Most Notable Individuals, Symbols, and Groups. Business Insider. Available online at: https://www.businessinsider.com/hate-symbols-and-extremist-groups-at-the-us-capitol-siege-2021-1 (accessed January 15 2021).
McIntyre, L. (2021). How to Talk to a Science Denier: Conversations With Flat Earthers, Climate Deniers, and Others Who Defy Reason. London: The MIT Press.
McKay, S., and Tenove, C. (2021). Disinformation as a threat to deliberative democracy. Political Res. Q. 74, 703–717. doi: 10.1177/1065912920938143
McPhedran, R., Ratajczak, M., Mawby, M., King, E., Yang, Y., Gold, N., et al. (2023). Psychological inoculation protects against the social media infodemic. Sci. Rep. 13, 62. doi: 10.1038/s41598-023-32962-1
Meltzer (2021). How New Age Spirituality and Sensitive Masculinity Led to QAnon. Washington Post Magazine. Available online at: https://www.washingtonpost.com/magazine/2021/03/29/qanon-new-age-spirituality/ (accessed March 29, 2021).
Mercier, H. (2021). How good are we at evaluating communicated information? Royal Inst. Philos. Supple. 89, 257–272. doi: 10.1017/S1358246121000096
Michel, A. H. (2022). What happens when everything becomes TikTok. The Atlantic. Available online at: https://www.theatlantic.com/technology/archive/2022/11/tiktok-instagram-video-feeds-ai-algorithm/672002/ (accessed February 15, 2024).
Modern Flat Earth. (n.d.) In Wikipedia. Available online at: https://en.wikipedia.org/wiki/Modern_flat_Earth_beliefs (accessed February 15 2024).
Modirrousta-Galian, A., and Higham, P. A. (2023). Gamified inoculation interventions do not improve discrimination between true and fake news: reanalyzing existing research with receiver operating characteristic analysis. J. Exp. Psychol. Gen. 152. doi: 10.31234/osf.io/4bgkd
Mogul, J. L., Ritchie, A. J., and Whitlock, K. (2011). Queer (In)Justice: The Criminalization of LGBT People in the United States. Boston, MA: Beacon Press.
Monroe, R. (2017). How Essential Oils Became the Cure for Our Age of Anxiety. The New Yorker. Available online at: https://www.newyorker.com/magazine/2017/10/09/how-essential-oils-became-the-cure-for-our-age-of-anxiety (accessed October 2 2017).
Mont'Alverne, C., Arguedas, A. R., Badrinathan, S., Toff, B., Fletcher, R., and Nielsen, R. K. (2023). Domain-specific influence on Facebook: how topic matters when assessing influential accounts in four countries. J. Quant. Des. Digital Media 3, 14. doi: 10.51685/jqd.2023.014
Myers, S. L., and Frenkel, S. (2022). How Disinformation Splintered and Became More Intractable. The New York Times. Available onlline at: https://www.nytimes.com/2022/10/20/technology/disinformation-spread.html (accessed October 20 2022).
Mylan, S., and Hardman, C. (2021). COVID-19, cults, and the anti-vax movement. The Lancet 397, 1181. doi: 10.1016/S0140-6736(21)00443-8
Neely, T. Y., and Montañez, M. (2022). Dismantling Constructs of Whiteness in Higher Education, 1st Edn. London: Routledge.
Neylan, J., Biddlestone, M., Roozenbeek, J., and van der Linden, S. (2023). How to inoculate against multimodal misinformation: a conceptual replication of Roozenbeek and van der Linden (2020). Sci. Rep. 13, 18273. doi: 10.1038/s41598-023-43885-2
Niewert, D. (2017). What the Kek: Explaining the Alt-Right Deity Behind Their Meme Magic. Southern Poverty Law Center. Available online at: https://www.splcenter.org/hatewatch/2017/05/08/what-kek-explaining-alt-right-deity-behind-their-meme-magic (accessed May 9 2017.).
Oremus, W., and Merrill, J. B. (2021). Five Points for Anger, One for a Like: How Facebooks Formula Fostered Rage and Misinformation. The Seattle Times. Available online at: https://www.seattletimes.com/nation-world/five-points-for-anger-one-for-a-like-how-facebooks-formula-fostered-rage-and-misinformation/ (accessed October 26, 2021).
Orenstein, W.alter A., Mark, J., and Melinda, E. (2004). Measles elimination in the United States. J. Inf. Dis. 189, S1–S3. doi: 10.1086/377693
Packer, J., and Stoneman, E. (2021). Where we produce one, we produce all: the platform conspiracism of QAnon. Cult. Polit. 17, 255–278. doi: 10.1215/17432197-9305338
Palmer, M. (2023). Digital Literacy in Academic Libraries: Frameworks, Case Studies, and Considerations. Cham: Springer.
Palmer, T. (2019). “When stories and pictures lie together–and you do not even know it,” in Metaliterate Learning for the Post-Truth World, eds T. P. Mackey and T. E. Jacobson (Chicago, IL: ALA Neal-Schuman), 103–142.
Peeters, S., and Willaert, T. (2022). Telegram and digital methods: mapping networked conspiracy theories through platform affordances. M/C J. 25, 2878. doi: 10.5204/mcj.2878
PEN America (2023). School Book Bans: The Mounting Pressure to Censor. Available online at: https://pen.org/report/book-bans-pressure-to-censor/ (accessed September 21, 2023).
Pepe the Frog (2023). In Wikipedia. Available online at: https://en.wikipedia.org/w/index.php?title=Pepe_the_Frogandoldid=1156293322 (accessed February 15, 2024).
Petersen, S. (2023). Momfluenced: Inside the Maddening, Picture-Perfect World of Mommy Influencer Culture. Boston, MA: Beacon Press.
Pierre, J. (2020). How Far Down the QAnon Rabbit Hole Did Your Loved One Fall? Psychology Today. Available online at: https://www.psychologytoday.com/blog/psych-unseen/202008/how-far-down-the-qanon-rabbit-hole-did-your-loved-one-fall (accessed August 21 2020).
Pierre, J. (2023). “Down the conspiracy theory rabbit hole: how does one become a follower of qanon?,” in The Social Science of QAnon: A New Social and Political Phenomenon, ed M. K. Miller (Cambridge: Cambridge University Press), 17–32.
Public Religion Research Institute (PRRI) (2023). Threats to American Democracy Ahead of an Unprecedented Presidential Election. Available online at: https://www.prri.org/research/threats-to-american-democracy-ahead-of-an-unprecedented-presidential-election/ (accessed February 21, 2024).
Rascoe, A. (2023). How Real is the Threat of AI Deepfakes in the 2024 Election? NPR: Weekend Edition Sunday. Available online at: https://www.npr.org/2023/07/30/1190970436/how-real-is-the-threat-of-ai-deepfakes-in-the-2024-election (accessed July 30, 2023).
Remski, M., and Brave NoiseCat, J. (2023). Special Report: QAnon Fantasies Look Like Colonial Realities. Conspirituality Podcast. Available online at: https://www.conspirituality.net/episodes/special-report-julian-brave-noisecat (accessed November 26, 2023).
Rieh, S. Y. (2010). Credibility and Cognitive Authority of Information. Encyclopedia of Library and Information Sciences, 3rd Edn. New York, NY: Taylor and Francis Group, LLC, 1337–1344.
Robertson, D. G., and Amarasingam, A. (2022). How conspiracy theorists argue: epistemic capital in the QAnon social media sphere. Popular Commun. 20, 193–207. doi: 10.1080/15405702.2022.2050238
Roose, K. (2020). Yoga Teachers Take On QAnon. The New York Times. Available online at: https://www.nytimes.com/2020/09/15/technology/yoga-teachers-take-on-qanon.html (accessed September 15, 2020).
Rousis, G. J., Richard, F. D., and Wang, D. Y. D. (2020). The truth is out there: the prevalence of conspiracy theory use by radical violent extremist organizations. Terror. Polit. Viol. 34, 1739–1757. doi: 10.1080/09546553.2020.1835654
Russo, A., Jankowski, A., Beene, S., and Townsend, L. (2019). Strategic source evaluation: addressing the container conundrum. Ref. Serv. Rev. 47, 297. doi: 10.1108/RSR-04-2019-0024
Scheufele, D. A., and Krause, N. M. (2019). Science audiences, misinformation, and fake news. Proc. Nat. Acad. Sci. 116, 7662–7669. doi: 10.1073/pnas.1805871115
Seddig, D., Maskileyson, D., Davidov, E., Ajzen, I., and Schmidt, P. (2022). Correlates of COVID-19 vaccination intentions: attitudes, institutional trust, fear, conspiracy beliefs, and vaccine skepticism. Soc. Sci. Med. 21, 114981. doi: 10.1016/j.socscimed.2022.114981
Serafini, F. (2013). Reading the Visual: An Introduction to Teaching Multimodal Literacy (Illustrated Edition). New York, NY: Teachers College Press.
Serafini, F. (2022). Beyond the Visual: An Introduction to Researching Multimodal Phenomena. New York, NY: Teachers College Press.
Sharlet, J. (2023). The Undertow: Scenes from a Slow Civil War. New York, NY: W. W. Norton and Company.
Sobo, E. J. (2021). Conspiracy theories in political-economic context: lessons from parents with vaccine and other pharmaceutical concerns. J. Cult. Res. 25, 51–68. doi: 10.1080/14797585.2021.1886425
Spring, M. (2020). Wayfair: The False Conspiracy About a Furniture Firm and Child Trafficking. BBC. Available online at: https://www.bbc.com/news/world-53416247 (accessed July 15, 2020).
Stein, R. A., Ometa, O., Pachtman Shetty, S., Katz, A., Popitiu, M. I., Brotherton, R., et al. (2021). Conspiracy theories in the era of COVID-19: a tale of two pandemics. Int. J. Clin. Prac. 75, e13778. doi: 10.1111/ijcp.13778
Stocking, G., Mitchell, A., Matsa, K. E., Widjaya, R., Jurkowitz, M., Ghosh, S., et al. (2022). The Role of Alternative Social Media in the News and Information Environment. Pew Research Centers Journalism Project. Available online at: https://www.pewresearch.org/journalism/2022/10/06/the-role-of-alternative-social-media-in-the-news-and-information-environment/ (accessed October 6, 2022).
Sullivan, D., Landau, M. J., and Rothschild, Z. K. (2010). An existential function of enemyship: Evidence that people attribute influence to personal and political enemies to compensate for threats to control. J. Pers. Soc. Psychol. 98, 434–449. doi: 10.1037/a0017457
Swanson, T. (2006). Information literacy, personal epistemology, and knowledge construction. College Undergraduate Libr. 13, 93–112. doi: 10.1300/J106v13n03_07
Swanson, T. (2023). Knowledge as a Feeling: How Neuroscience and Psychology Impact Human Information Behavior. Lanham: Rowman and Littlefield.
Swire-Thompson, B., Miklaucic, N., Wihbey, J. P., Lazer, D., and DeGutis, J. (2022). The backfire effect after correcting misinformation is strongly associated with reliability. J. Exp. Psychol. Gen. 151, 1655. doi: 10.31234/osf.io/e3pvx
Thaler, R. H., and Sunstein, C. R. (2009). Nudge: Improving Decisions About Health, Wealth, and Happiness. New York, NY: Penguin Books.
The Associated Press (2022). Hateful Slurs Soared on Twitter After Musk Took Over, Says Digital Anti-Hate Watchdog. CBC News. Available online at: https://www.cbc.ca/news/world/twitter-racial-slurs-hate-musk-1.6647974 (accessed November 10, 2022).
Thompson, D. S. (2019). Teaching students to critically read digital images: a visual literacy approach using the DIG method. J. Visual Liter. 38, 110–119. doi: 10.1080/1051144X.2018.1564604
Thompson, D. S., and Beene, S. (2023). “Reading images with a critical eye,” in Teaching Critical Reading Skills: Strategies for Academic Librarians, Vol. 2, eds. H. G. Rempel and R. Hamelers (Chicago, IL: ACRL), 423–434.
Thompson, D. S., Beene, S., Greer, K., Wegmann, M., Fullmer, M., Murphy, M., et al. (2022). A proliferation of images: trends, obstacles, and opportunities for visual literacy. J. Visual Liter. 41, 113–131. doi: 10.1080/1051144X.2022.2053819
Thorbecke, C. D., and Catherine. (2022). Elon Musk Says He Will Begin Restoring Previously Banned Twitter Accounts Next Week CNN. Available online at: https://www.cnn.com/2022/11/24/tech/elon-musk-amnesty-poll/index.html (accessed November 24, 2022).
Tinmaz, H., Lee, Y.T., Fanea-Ivanovici, M., and Baber, H. (2022). A systematic review on digital literacy. Smart Learning Environ. 9, 1–18. doi: 10.1186/s40561-022-00204-y
Topor, L. (2022). Phishing for Nazis: Conspiracies, Anonymous Communications and White Supremacy Networks on the Dark Web. London: Routledge.
Unknown (2021a). This Week in Fascism #107: Violence from Proud Boys Still Welcomed at Rallies Alongside Trump Supporters; Given Free Pass by Police. Its Going Down. Available online at: https://itsgoingdown.org/this-week-in-fascism-107/ (accessed February 15, 2024).
Unknown (2021b). Craig Longley. Stop Antisemitism (website). Available online at: https://www.stopantisemitism.org/antisemite-of-the-week5-1/antisemite-of-the-week-craig-longley-the-jew-hating-doctor (accessed May 8, 2021).
Uscinski, J. E., and Parent, J. M. (2014). American Conspiracy Theories. Oxford: Oxford University Press.
van Bavel, J. J., Baicker, K., Boggio, P. S., Druckman, J. N., Drury, J., Dube, O., et al. (2020). Using social and behavioural science to support COVID-19 pandemic response. Nat. Hum. Behav. 4, 460–471. doi: 10.1038/s41562-020-0884-z
van der Linden, S., Dixon, G., Clarke, C., and Cook, J. (2021). Inoculating against COVID-19 vaccine misinformation. Lancet 33:100772. doi: 10.1016/j.eclinm.2021.100772
van Der Linden, S., Roozenbeek, J., and Compton, J. (2020). Inoculating against fake news about COVID-19. Front. Psychol. 9, 5667908. doi: 10.3389/fpsyg.2020.566790
Voinea, C., Marin, L., and Vică, C. (2023). The moral source of collective irrationality during COVID-19 vaccination campaigns. Philos. Psychol. 32, 1–20. doi: 10.1080/09515089.2022.2164264
Walker, J. (2013). The United States of Paranoia: A Conspiracy Theory, 1st Edn. Palatine, IL: Harper. Available online at: https://www.jstor.org/stable/27007298
Walther, S., and McCoy, A. (2021). US extremism on telegram: fueling disinformation, conspiracy theories, and accelerationism. Persp. Terr. 15, 100–124.
Weil, K. (2022). Off the Edge: Flat Earthers, Conspiracy Culture, and Why People Will Believe Anything. Dunham, NC: Algonquin Books.
Weiner, S. (2013). Teaching Information Literacy for Life: Addressing the Issues (Paper 21). Purdue E-Pubs. Weiner, Sharon A., Teaching Information Literacy for Life: Addressing the Issues (2013). Libraries Faculty and Staff Presentations. Paper 21. Available online at: https://docs.lib.purdue.edu/lib_fspres/21 (accessed February 15, 2024).
Westphal, M. (2017). Hermeneutics as Epistemology. The Blackwell Guide to Epistemology. New York, NY: John Wiley and Sons, Ltd, 415–435.
Wilson, P. (1983). Second-Hand Knowledge: An Inquiry into Cognitive Authority. Westport, CT: Greenwood Press.
Wineburg, S., and McGrew, S. (2017). Lateral Reading: Reading Less and Learning More When Evaluating Digital Information, SSRN Scholarly Paper. Rochester, NY: Social Science Research Network.
Yasmin, S., and Spencer, C. (2020). But I Saw It on Facebook: Hoaxes Are Making Doctors Jobs Harder. The New York Times. Available online at: https://www.nytimes.com/2020/08/28/opinion/sunday/coronavirus-misinformation-faceboook.html (accessed August 28, 2020).
Zembylas, M. (2021). “The affective turn in educational theory,” in Oxford Research Encyclopedia of Education, ed G. W. Noblit (Oxford: Oxford University Press).
Ziegler, M. F., Paulus, T., and Woodside, M. (2014). Understanding informal group learning in online communities through discourse analysis. Adult Educ. Q. 64, 60–78. doi: 10.1177/0741713613509682
Keywords: information literacy, QAnon, metaliteracy, visual literacy, social media, conspiracy theories
Citation: Greer K and Beene S (2024) When belief becomes research: conspiracist communities on the social web. Front. Commun. 9:1345973. doi: 10.3389/fcomm.2024.1345973
Received: 28 November 2023; Accepted: 09 February 2024;
Published: 28 February 2024.
Edited by:
Christopher T. Conner, University of Missouri, United StatesReviewed by:
T. Kenny Fountain, University of Virginia, United StatesRobert Spinelli, American Baptist College, United States
Copyright © 2024 Greer and Beene. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Katie Greer, Z3JlZXImI3gwMDA0MDtvYWtsYW5kLmVkdQ==