This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Exoskeleton for the Mind: Exploring Strategies Against Misinformation with a Metacognitive Agent

Yeongdae Kim 0000-0001-5346-0041 Tokyo Institute of Technology2 Chome-12-1 OokayamaMeguroTokyoJapan152-8550 [email protected] Takane Ueno 0000-0001-6626-7945 Tokyo Institute of Technology2 Chome-12-1 OokayamaMeguroTokyoJapan152-8550 [email protected] Katie Seaborn 0000-0002-7812-9096 Tokyo Institute of Technology2 Chome-12-1 OokayamaMeguroTokyoJapan152-8550 [email protected] Hiroki Oura 0000-0001-5111-7864 Tokyo University of Science1 Chome-3 KagurazakaShinjukuTokyoJapan162-8601 [email protected] Jacqueline Urakami 0000-0002-2866-0807 Tokyo Institute of Technology2 Chome-12-1 OokayamaMeguroTokyoJapan152-8550 [email protected]  and  Yuto Sawa Tokyo Institute of Technology2 Chome-12-1 OokayamaMeguroTokyoJapan152-8550 [email protected]
(2018)
Abstract.

Misinformation is a global problem in modern social media platforms with few solutions known to be effective. Social media platforms have offered tools to raise awareness of information, but these are closed systems that have not been empirically evaluated. Others have developed novel tools and strategies, but most have been studied out of context using static stimuli, researcher prompts, or low fidelity prototypes. We offer a new anti-misinformation agent grounded in theories of metacognition that was evaluated within Twitter. We report on a pilot study (n=17) and multi-part experimental study (n=57, n=49) where participants experienced three versions of the agent, each deploying a different strategy. We found that no single strategy was superior over the control. We also confirmed the necessity of transparency and clarity about the agent’s underlying logic, as well as concerns about repeated exposure to misinformation and lack of user engagement.

Misinformation, Social media, Design study, Intelligent agent, Metacognition
copyright: acmcopyrightjournalyear: 2018doi: XXXXXXX.XXXXXXXconference: The Augmented Humans (AHs) International Conference; March 12-14, 2023; Glasgow, United Kingdombooktitle: Augmented Humans International Conference (AHs ’23), March 12-14, 2023, Glasgow, United Kingdomprice: 15.00isbn: 978-1-4503-XXXX-X/18/06ccs: General and reference Empirical studiesccs: Networks Social media networksccs: Human-centered computing User studies

Citation

Yeongdae Kim, Takane Ueno, Katie Seaborn, Hiroki Oura, Jacqueline Urakami, and Yuto Sawa. 2023. Exoskeleton for the Mind: Exploring Strategies Against Misinformation with a Metacognitive Agent. In Proceedings of the Augmented Humans International Conference 2023 (AHs ’23). Association for Computing Machinery, New York, NY, USA, 209–220. https://doi.org/10.1145/3582700.3582725

The final publication is available via ACM at https://dl.acm.org/doi/10.1145/3582700.3582725.

1. Introduction

The digital misinformation age is upon us. Fake news, inaccuracies, both accidental and intended, opinions presented as fact, and misinformed but viral memes are an everyday feature of the social media landscape (Allcott and Gentzkow, 2017; Allcott et al., 2019). Social media users are exposed to and have the ability to broadcast content of all kinds with little oversight. Professional news sources, published research, and expert opinion are mixed in with, and often indistinguishable form, unverified and disreputable content. A combination of proprietary algorithms and human cognitive biases have even created ”echo chambers” (Colleoni et al., 2014) rife with misinformation and polarized views. This has led to the present state of affairs: the ”infodemic” (Zarocostas, 2020). While the notion of misinformation is not new (Hernon, 1995), its scale and spread are a feature of the Internet-connected world in which most people on Earth currently live. For example, there is substantial evidence that social media was harnessed by bad actors in order to interfere with the US 2016 election (Allcott and Gentzkow, 2017) and galvanize violent action that led to the 2021 US capitol riot (Bleakley, 2021; Garrett, 2021). Misinformation has also been a feature of social media output on the COVID-19 pandemic (Kouzy et al., 2020). This has led to urgent calls for ways in which to regulate social media (Cusumano et al., 2021), understand how misinformation influences people (Geeng et al., 2020; Allcott et al., 2019; Shin et al., 2018; Wu et al., 2016), and provide people with tools to fight misinformation in these spaces (Vraga and Bode, 2017; Bautista et al., 2021; Walter et al., 2020; Mena, 2020; Ross et al., 2018).

Social media companies explored automated and manual content curation, filters, warning labels, and policy updates (Niemiec, 2020; Stjernfelt and Lauritzen, 2020). However, criticism and concerns have been raised about moderation inconsistency and censorship (Stjernfelt and Lauritzen, 2020), as well as conflicts of interest and a lack of known empirical testing when it comes to company fact-checking tools, such as Meta/Facebook’s Third-Party Fact-Checking Program111https://www.facebook.com/formedia/mjp/programs/third-party-fact-checking and Twitter’s Community Notes222https://twitter.github.io/communitynotes. Another promising direction is third-party anti-misinformation tools. Previous research has explored a variety of strategies, such as warning labels (Bhuiyan et al., 2018; Hartwig and Reuter, 2019; Pennycook et al., 2018), fact-checking (Vijjali et al., 2020; Kim and Walker, 2020; Chi and Liao, 2022), asking users to pause and reflect (Fazio, 2020; Bago et al., 2020; Pennycook et al., 2021), and so on. Yet, most of this work has been conducted in lab environments with artificial stimuli, using printouts of doctored social media posts. Others have relied on quantitative methods, limiting access to richer forms of experiential data, such as observations of how people attend and react to certain content over others. This also limits access to the ”why” of behaviour and the meaning of the quantitative results. Also, most work has focused on one strategy in isolation, which makes drawing comparisons among strategies to identify the best strategies very difficult. Finally, several strategies, especially well-studied and widely used strategies like warning labels, have been found ineffective. A different approach to the design and study of anti-misinformation tools in social media may strengthen this crucial area of work and lead to clearer results.

To this end, we created a new anti-misinformation agent for live social media contexts that deploys a range of strategies informed by theories of metacognition in the digital age (Kozyreva et al., 2020). Metacognition means ”thinking about one’s thinking” (Dunlosky and Metcalfe, 2008). A ”metacognitive agent,” then, attempts to raise the user’s awareness of their thinking. In naming our agent, we translated a common metaphor from physical human augmentation systems to the case of metacognition: Elemi, an ”exoskeleton for the mind.” We aimed to build on previous work by: (i) using real content in a live platform known to have a misinformation problem, i.e., Twitter (Allcott et al., 2019); (ii) directly comparing several metacognitive strategies; and (iii) combining quantitative and qualitative forms of inquiry, i.e., mixed methods, to increase clarity on agent and strategy efficacy. We asked: Can a metacognitive agent be effective at helping people tackle misinformation within a live social media context? We conducted a pilot study on the basic form of the agent (n=17) and a multi-phased experimental study (n=57 in Session 1 and n=49 in Session 2) comparing three strategies deployed by the agent—Cover, Camouflage, and Consider—against a control. We also evaluated the presence of an illusory truth effect (Pennycook et al., 2018) by checking for changes in perceptions of content accuracy influenced by exposure in the first session. We used a curated selection of real Tweets to increase participant comfort and honesty, in response to recent findings on social stigma about misinformation (Ren et al., 2023). We aimed to strike a balance between realism and experimental control against the state of our prototype. We contribute: a proof-of-concept prototype that offers three strategies and was built for a live environment; a methodological approach to evaluating such a tool in situ; and insights into the acceptance, value, behaviour, and user experience (UX) of such a tool for social media users. We end with implications for designing such tools and directions for future research.

2. Related Work

We discuss the tools and strategies being explored to combat misinformation and how our agent builds on this foundation.

2.1. Anti-Misinformation Tools That Warn Us

Algorithms can be (re)developed to warn users when they are potentially or actually exposed to false information (Bhuiyan et al., 2018; Sharevski et al., 2021; Hartwig and Reuter, 2019; Saltz et al., 2021; Pennycook et al., 2018). Bhuiyan et al. (Bhuiyan et al., 2018) explored the effect of warnings on the perceived credibility of news content in FeedReflect, a browser extension for Twitter. Content that contained false information was highlighted or blurred with a warning popup. Findings showed that the warning nudge treatment increased the perceived credibility of the mainstream content while decreasing the credibility of the non-mainstream, misinformed content. However, this work was a small pilot test. Similarly, Sharevski et al. (Sharevski et al., 2021) investigated the impact of warning labels in Twitter. They utilized two soft moderation methods: Warning Cover and Warning Tag. They found that a simple Warning Tag was insufficient to reduce the perceived accuracy of misleading content. Even so, a Warning Cover led users to perceive the accuracy of misleading content negatively, i.e., users trusted content with misinformation less. But they tested it with content that they created, rather than real content. Hartwig and Reuter (Hartwig and Reuter, 2019) created a browser extension for Twitter called TrustyTweet, which used a ”white box” approach to inform users about the reasons behind algorithmic fact-checking. The tool attached a warning icon in combination with an explanation when clicked. Findings suggested that the tool was generally considered helpful and maintained the user’s sense of autonomy. However, this was a small study and the guidelines need further confirmation. We chose the Warning Cover strategy (Sharevski et al., 2021) as our first strategy, evaluating it in our pilot study as the Cover version of the agent and then iterating its design based on user feedback for the experimental study. We used the ”white box” approach (Hartwig and Reuter, 2019) for the Camouflage version of the agent. We built on Bhuiyan et al. (Bhuiyan et al., 2018) by running a full study with a larger sample of participants.

2.2. Anti-Misinformation Tools That Ask Us to Pause and Reflect

In a ”pause and reflect” strategy, the user is prompted to stop for a moment and reflect on the content that they are consuming (Fazio, 2020; Bago et al., 2020). Fazio (Fazio, 2020) explored a prompt that asked participants to pause and then explain how they knew that the headline of a Facebook news article was true or false. She found that doing so reduced the likelihood that people would share false news. However, the effect was lost on re-exposure to the same news items. Moreover, the stimulus was artificial: paper printouts of real and artificially faked versions of Facebook posts and prompts created by the experimenter. Bago, Rand, and Pennycook (Bago et al., 2020) found that pausing increased the accuracy ratings of fake news that aligned with participants’ political leanings. Later, Pennycook et al. (Pennycook et al., 2021) explored having the user answer questions, aiming to prompt a shift in attention to the accuracy of news headlines/images. This increased the quality of the news shared. Also, people shared news that was not politically aligned with their own beliefs. While promising, this research was limited in the same way as that of Fazio (Fazio, 2020). Chen, Ciuccarelli, and Colombo (Chen et al., 2022) explored an indirect ”pause and reflect” tool. Inspired by the ”filter bubble” phenomenon, they created a browser extension for Facebook called VisualBubble. The extension provided posts that were similar and different in opinion to that of individual users, i.e., in or outside their bubble. Some became skeptical but were not able to distinguish ”false” news in general. Still, based on its potential, we incorporated the ”pause and reflect” strategy in the Consider form of our agent. We used real Tweets and a live Twitter setting with an agent rather than a human experimenter. We also compared this strategy directly with the ”cover and warn” or Cover strategy and the Camouflage strategy, alongside a control group.

2.3. Anti-Misinformation and Fact-Checking

Sharing the results of fact-checking with users has also been explored. Vraga, Kim, and Cook (Vraga et al., 2019) considered what style of corrections would be most effective: logic-based or humor-based. They found that success depended on the topic, with logic-based corrections being the most effective overall. Epstein, Pennycook, and Rand (Epstein et al., 2020) considered a crowdsourcing mechanism of aggregating laypeople’s evaluations of the trustworthiness of sources on social media. They found that people trusted mainstream sources more than hyper-partisan or fake news sources, indicating that crowdsourcing might be a viable fact-checking option. Diakopoulos, Trielli, and Lee (Diakopoulos et al., 2021) developed and investigated an AI-powered tool to draw journalists’ attention to data that was subjected to quality and newsworthiness evaluations, but this was not assessed with lay users. In short, fact-checking seems to be an acceptable strategy for a variety of users.

Nevertheless, live fact-checking remains a technical challenge. While we designed our agent, Elemi, to use fact-checking as a basis for decision-making about misinformation, we were unable to develop a real-time fact-checking system—a project in its own right. Moreover, we wished to maintain a degree of experimental control, i.e., to compare equal sets of real and fake content. We also had concerns, emphasized by recent research (Ren et al., 2023), about whether people would be comfortable sharing their true thoughts and feelings about misinformation from their own feeds, i.e., content they or those in their social network produce or share. Due to these technical, methodological, and social challenges, we opted to curate and manually fact-check the Tweets used by our agent. We discuss this more in 3.1 and 3.2.

3. Pilot Study of the Cover Strategy: Explainable and Controllable Filtering

We conducted a pilot study of the first version of our agent, Elemi, which used the Cover strategy. Elemi hid Tweets containing misinformation, provided an explanation for the its decision based on fact-checking, and maintained user autonomy by allowing access to the content. We aimed to verify the positive results found in Sharevski et al. (Sharevski et al., 2021) with real rather than artificial content, increase efficacy with explanations based on fact-checking (Vraga et al., 2019; Epstein et al., 2020; Diakopoulos et al., 2021; Ren et al., 2023), and determine the usability and UX of the agent within the actual context of use: live Twitter. We asked: (RQ1) How user-friendly is the agent and what kind of experience does it provide? and (RQ2) How effective is the ”cover” strategy when deployed alongside a transparent, explainable AI system characterized by fact-checking and user autonomy in accessing hidden content? We explored these questions in a user study format with direct evaluation and interview phases.

3.1. System Design

Our agent Elemi (Figure 1) was a browser-based tool designed in and for Twitter. It did not use the Twitter API, so it was not subject to the ongoing shifts in restrictions placed on third party apps333https://www.engadget.com/twitter-new-developer-terms-ban-third-party-clients-211247096.html. The first version of the agent, Cover, deployed the ”cover” strategy of hiding tweets containing misinformation in a transparent and explainable way. The agent provided indicators that this was done and explained why with reference to fact-checking sources.

Refer to caption
Figure 1. Overview of the Cover version of the agent. Tweets with misinformation are hidden automatically by the agent (1). Users can access an explanation from the agent about why the tweet was hidden (2). They can also manually view the hidden tweet (3).

Elemi was designed to use fact-checking resources, but as a research prototype, it relied on the manual curation and fact-checking of tweets. When enabled, Elemi appeared as a cat-like, blue character icon on hidden tweets deemed to contain misinformation. The user could toggle the visibility of the tweet (user autonomy) and access an explanation about how Elemi based its decision on the results of fact-checking (transparency). Two researchers compiled the tweets used by the agent; refer to 3.2.3. Being semi-autonomous, Elemi automatically labelled tweets within the live environment of Twitter, without direct human assistance. This was concealed from participants so as to simulate how it would actually work, i.e., without direct human involvement, allowing for a more accurate evaluation of the UX.

3.2. Methods

We conducted the pilot test on the Cover version of Elemi between December 8th to 23rd, 2021.

3.2.1. Participants

Seventeen Japanese participants (13 men, 4 women, no one of another gender identity; aged 22-62) who had used Twitter at least once participated in the research. The research protocol was approved by the university ethics committee (approval #2022170). All participants signed a consent form after receiving an explanation of the research. They were given compensation of equivalent to roughly $11 USD.

3.2.2. Location

We conducted the study online and offline, due to restrictions arising from the COVID-19 pandemic. The offline version was conducted through a computer in the lab. The online version used Zoom and the Quick Assist program available in Windows 10. This allowed the participant to remotely control the lab computer, which allowed us to maintain the same virtual environment online and off, especially in terms of the browser environment.

3.2.3. Materials

The pre-screened tweets were fact-checked and categorized as either fake or true. For this, two researchers searched for real tweets about COVID-19 posted on Twitter between March 2020 and October 2021. 26 keywords were used as search terms, e.g., Vaccine, Corona, Corona graphene, Vaccine, Vaccine poison, Vaccine mask, mRNA spike, mask germs, mRNA poison, mRNA vaccine, Corona Electromagnetic Wave, Vaccine Danger, etc. Factual tweets usually referenced source data from the government or research articles, while fake tweets tended to lack evidence or media fact-checking. This led to more manual fact-checking efforts on the misinformation side. Fact-checking involved reviewing government materials and media articles. Tweets that both agreed to be true or fake were selected. From these, 15 factual and 15 fake tweets were randomly selected for the agent’s use.

3.2.4. Measures, Instruments, and Data collection

Qualitative and quantitative data were collected. Participants’ voices and faces during the think-aloud protocol were recorded. Audio was transcribed by a researcher who spoke fluent Japanese. The post-test questionnaire consisted of fifteen items using Likert scales and three open questions. Perceived usefulness (PU) was used as a measure of initial trust (Benbasat and Wang, 2005) and assessed through four items from the instrument by Zhou et al. (Wang et al., 2016) on a 7-point Likert scale. Usability was assessed through the 10-item System Usability Scale (SUS) (Brooke et al., 1996) on a 5-point Likert scale. We created one Likert-style item to assess frequency of Twitter use (no longer a user, never used, used once a month, once a week, daily). The open-ended questions asked about likes, dislikes, and what could be improved about the agent. For the interview, we asked about experience with the agent, use of Twitter, thoughts about Elemi’s design, understanding of the agent’s capabilities, and willingness to use the agent again or not. All items were presented in Japanese; all English scales were translated into Japanese by one of the researchers who spoke fluent Japanese and English, and back-translated by two or more other researchers whose native language was Japanese.

3.2.5. Procedure

Participants were given a description of the study and asked for consent. Next, they practiced the think-aloud protocol (Ericsson and Simon, 1984). For this, they read cat-related tweets for  5 mins. They then completed the main task using the think-aloud protocol to narrate their experience. For this, they read tweets from a timeline comprising the curated 15 factual and 15 fake tweets, while the agent was activated. Participants were not directed to engage with the agent; they were free to use it or ignore it. Afterwards, they answered a post-test questionnaire and participated in an interview. We then ended the session. Participants were thanked and compensated. The study took about one hour.

3.2.6. Data Analysis

We generated descriptive statistics and scored the measures according to the instrument instructions. We used a two-tailed Student’s t-test to compare the usability and PU scores by willingness or unwillingness to use the agent. For the qualitative data, we used reflexive thematic analysis (Braun and Clarke, 2006). An inductive approach was used in the absence of a predefined thematic framework. For this, one researcher read the corpus of data drawn from the think-aloud protocol, interview transcripts, and questionnaires three times. The researcher then developed 41 codes, later categorized under eight high-level codes: (1) experimental environment, (2) user interface, (3) feedback on the agent’s functions, (4) inferences about the agent’s logic, (5) UX, (6) feedback on the content, (7) opinion on the agent’s information, and (8) agent improvements. The researcher then grouped these into three themes, in light of the two RQs. These were reviewed several times with another researcher for clarity and refinement.

3.3. Results

3.3.1. Quantitative Results

Most participants identified the agent’s role as an anti-misinformation tool correctly; however, two did not. These were explainable: a non-frequent Twitter user and a non-student who believed that the timeline itself was created by the agent. We used the data from the fifteen participants who did understand the agent’s role. On average, they opened 43% of the tweets marked as misinformation. The average normalized score for the SUS was 66.91, indicating marginal usability (Bangor et al., 2008). There were no significant differences in usability scores between the eight participants willing to use the agent (70.25±19.8870.25\pm 19.88) and the seven who were unwilling (62.14±10.1562.14\pm 10.15). The mean PU score was 4.53 (out of 7). A significant difference between the willing (5.10±0.655.10\pm 0.65) and unwilling (3.71±0.963.71\pm 0.96) groups, t(15) = 3.576, p ¡ .01, indicated that those willing to use it found it more useful, as expected.

3.3.2. Qualitative Findings

We now turn to a review of our inductive findings from the thematic analysis.

Theme 1: Desire for transparency. Sixteen participants (all but one) wondered about the logic behind the agent. They investigated author profiles, the presence of quotes, the tone of the content, slang use, and other indicators to understand how the agent identified inaccurate tweets. Nine participants found that the filtering was not in line with their criteria and raised it as a problem. P1: ”Well, this tweet hasn’t been deleted, but I honestly don’t believe it.” P8: ”He doesn’t seem like a person who would be likely to publish inaccurate information, so I’m surprised that his tweet is inaccurate, and he’s using foreign data in English, or a summary of a paper, which seems to contain credible information.” Seven participants also expressed concern that the agent was biased and used only official or anti-COVID-19 information. P17: ”I’m a little scared that the information that the agent is letting me see or blocks is decided arbitrarily.”

Theme 2: Trust and reliance. Even when the agent was perceived as mistaken, some participants still trusted its decision. P11: ”Even if the information is correct and the agent is wrong, I’d still accept the agent’s decision.”. In fact, four participants hesitated to trust some tweets, stating that they would follow the agent’s decision without knowing why it was made. P1: ”I may believe this tweet because it’s not hidden by this tool.” P6: ”It’s easy to believe because the agent didn’t hide the tweet.”. Ten responded that they would use the agent in future.

Theme 3: Orientations to visual filtering. Impressions of the visual filtering changed over time with continued use. Nine participants found that the more they used the agent, the more it attracted their attention. P1: ”I want to see what the conspiracy theory is by looking at the hidden tweets.” P4: ”If such radical content is hidden, this might be a cue to click it, so I thought it might be a ploy by someone who wants to spread such radical ideas.”. Five participants also noticed and appreciated that the agent prepared them to consider the content before they opened it. P11: ”It was good to embody the point of stopping and thinking about whether or not it was fake information.”

Still, about half did not necessarily want the agent to take action on misinformation. Eight participants enjoyed reading misinformation, finding it humorous. P13 criticized the filtering: ”It’s fun to hear various opinions, and I gain something from them. We must to decide ourselves whether to accept a specific opinion or not. But this agent takes away such a process. It seems for me there’s no difference between a person deeply immersed in conspiracy theories and users of the agent.”. One participant suggested that it would be better to make the right information stand out rather than hide the wrong information. P3: ”I’d like the agent to comment on something rather than just excluding negative tweets.”.

3.4. Discussion

We now discuss our findings from the pilot study based on the research questions.

3.4.1. Agent Usability and UX (RQ1)

The agent was generally usable, based on the SUS score (66.91) (Bangor et al., 2008) and qualitative feedback. Frequent Twitter users immediately understood the agent’s purpose and the ”one click” toggle generated trust. Still, the presence of the agent iself drew attention to the tweets containing misinformation, leading many to focus on these fake tweets rather than stay vigilant. This points to the Pandora effect (Hsee and Ruan, 2016): a desire to resolve uncertainty even while knowing that it will have a negative effect. Similar to Pennycook, Cannon, and Rand (Pennycook et al., 2018), the ”fact-checked” explanations provided by the agent served to increase perceived credibility, regardless of factuality. However, the unintentional, negative side-effects need addressing, perhaps with metacognitive scaffolds or a means of allowing the user to guess what the hidden content might be even while not displaying the content directly (Stevens, 2007).

3.4.2. Agent Effectiveness (RQ2)

Observations showed participants skipped nearly half of the content containing misinformation. In this sense, the agent was effective at preventing exposure to misinformation in live Twitter. But this experience was not pleasant for everyone. PU was linked to a desire to use the agent in the future. Some participants wished for a greater understanding of how the agent made decisions, while some were satisfied even when they thought that the agent made a mistake. Information that lacks transparency and a clear objective is unacceptable, even when true (Hartwig and Reuter, 2019). Others wished to view misinformation regardless, which may defeat the purpose of hiding it due to the threat of repeated exposure, where familiarity breeds a sense of accuracy, i.e., an illusory truth effect (Hassan and Barber, 2021).

Overall, most people found the surface-level explanations about fact-checking to be insufficient when the agent covered up content. We also needed to tackle the challenge of repeated exposure. We thus returned to the literature, drawing on theories of metacognition as a means of assisting users who wish to engage with content containing misinformation directly. We developed two new versions of the agent, each representing a high-level strategy suggested as promising by previous work: Camouflage (Stevens, 2007), a new form of Cover that blurs out (rather than hides) the content and specifies what parts were misinformed and why, as a metacognitive cue; and Consider, which uses the metacognitive ”pause and reflect” strategy investigated in numerous studies (Fazio, 2020; Bago et al., 2020). We report on the study next.

4. Experimental Study: The Metacognitive Camouflage and Consider Strategies

We experimentally evaluated the new versions of Elemi, Camouflage and Consider, focusing on whether and how user behaviors and perceptions of misinformation changed with each strategy and over time. We used a between-subjects design with conditions for each version of the agent and a control condition (no Elemi, just regular Twitter). Two sessions were conducted two weeks apart to explore time-based effects, notably the illusory truth effect, where prior exposure to fake news can increase perceptions of accuracy and willingness to share that news (Pennycook et al., 2018). As before, the study took place in a lab setting but used real content in live Twitter. We asked: (RQ1) Does the Camouflage strategy work better than the Consider strategy at helping people identify misinformation? Given our shift from hiding content to enabling a way of productively engaging with content using a metacognitive approach deployed over time, we needed to check for an illusory truth effect. We asked: (RQ2) Does an agent employing a metacognitive approach mitigate the illusory truth effect and to what degree? We begin by describing the new forms of Elemi, the agent.

4.1. System Design

We (re)designed the agent around the model of metacognition by Dunlosky and Metcalfe (Dunlosky and Metcalfe, 2008), focusing on the metacognitive strategies component as a method users can employ when exposed to or choosing to engage with misinformation. This component, targeting the regulation of cognition, is made up of three activities: planning, or deciding to be conscious of one’s thinking and how to do so; monitoring, or keeping track of thinking about one’s thinking in the moment; and evaluating one’s success at the thinking task. In designing the agent to elicit certain activities, we adapted the framework of behavioural and cognitive interventions by Kozyreva et al. (Kozyreva et al., 2020). We explain below.

The Camouflage version of the agent (Figure 2.a) restricted the content users were exposed to using a blur filter rather than hiding the content entirely, i.e., the Cover strategy). Based on the pilot test, we expected this would act as a nudge (Kozyreva et al., 2020), but we could not predict in which direction. If the blurring effect reduced the visual saliency of the content, it could act as a nudge away from the content (Stevens, 2007). But if it drew attention as a novel visual element in the context of Twitter, it could nudge users towards the content—a possibility that we could only evaluate through our novel experimental approach in live Twitter. The agent stimulated the planning stage of metacognition by explaining why the content was hidden, i.e., based on specific fact-checking sources, and allowing the user to unblur it. The agent also indicated where misinformation was detected in the text of the content with striked red text (Vraga et al., 2020) and explanations on hover (Hartwig and Reuter, 2019). In essence, we aimed to improve the design of the Cover strategy using a metacognition-inspired approach with a visual Camouflage treatment and a more robust explanation, all the while retaining user control and transparency.

The new version Elemi, Consider (Figure 2.b), offered a ”technocognition” intervention (Kozyreva et al., 2020) in the form of a ”pause and reflect” strategy. Technocognition is about ”directed friction,” or encouraging deliberation of the content being consumed. This approach embodies the full metacognition cycle: planning, monitoring, and evaluating. An icon of the agent was placed next to the share button in each tweet, a critical zone in that UI; we aimed to divert attention from that button to the agent instead. The icon drew attention by turning left and right at a ±5\pm 5 angle; its speech bubble also changed in size by 10%. When clicking the agent, the user was asked about the accuracy of the content—prompting a planning activity—with four response options: accurate, inaccurate, not sure, don’t care (Pennycook et al., 2021). The agent then asked for the reason behind the response (Bago et al., 2020), prompting a monitoring activity. Then the agent provided a graph of other users’ choices444The user choice data was predefined based on the fact-checking sources used by the agent (Kim and Walker, 2020)., prompting an evaluation activity and completing the metacognitive cycle of engagement for that tweet.

Refer to caption
(a) Camouflage
Refer to caption
(b) Consider
Figure 2. User flow for (a) Camouflage: Tweets with misinformation are hidden (1). The user can engage the agent (2) for an explanation (3) and access the blurred out tweet (4). They can view the misinformation indicators (5) and fact-checking sources (6). User flow for (b) Consider: The agent is next to the share button (1). The user clicks the agent and is asked for their opinion on the accuracy of the content (2). The user responds (3). The agent asks for the reason (4). The agent displays a graph of response data from other users (5).

4.2. Methods

We conducted a phased 2x3 controlled experimental study in which participants were pseudo-randomly assigned to one of three between-subjects conditions (Camouflage, Consider, or the control) and experienced the agent in two phases, roughly two weeks apart (within-subjects). We preregistered our methods and hypothesis with OSF555https://osf.io/ucy9p on April 22nd, 2022 and updated it on June 21st, 2022. The experiment ran from May 12th to September 13th, 2022.

4.2.1. Participants

We recruited people experienced with Twitter (Table 1), with 57 in the first session and 49 completing the second. They were asked about their age and level of education after the second session to avoid stereotype threats (Kray et al., 2002; Mazerolle et al., 2012). Participants were recruited through word-of-mouth, our social networks, and Prolific666https://www.prolific.co/. They were compensated with at roughly $11 USD per session. This study was approved by the university ethics committee (approval #2022170).

Table 1. Age range and education level of participants
Age Range Overall Control Camouflage Consider
182418\sim 24 25 8 8 9
253425\sim 34 17 6 5 6
354435\sim 44 3 0 2 1
455445\sim 54 2 1 1 0
556455\sim 64 2 0 1 1
Education Level
High school degree or equivalent 11 2 6 3
Associate degree 4 1 2 1
Bachelor’s degree 14 5 4 5
Graduate degree 20 7 5 8
Total 49 15 17 17

4.2.2. Location

The study was conducted online and offline. Participants remotely controlled a lab computer in Zoom. This allowed us to control the study and ensure that all participants experienced the same environment.

4.2.3. Materials

The agent used the same fact-checking resources as in the pilot study, but the pre-screened tweets were changed to ensure up-to-date content. 30 factual and 10 fake tweets posted between January and May 2022 were selected. Unlike the pilot study, we allowed participation in either Japanese or English. For this, two researchers translated and back-translated the tweets. Also, all UI text and other content, such as ads and response tweets, were translated in realtime using the Google Translate API. English participants were notified that the original content was in Japanese to avoid confusion due to missed translations or oddities in the machine translation.

4.2.4. Data Collection

Tweet accuracy evaluations, misinformation exposure time, fact-checking activities, and usage logs were collected. Screen recordings with Zoom were used to capture observations of user activities. We verbally prompted participants to judge whether a tweet contained misinformation (Chen et al., 2022). These assessments, as well as exposure time to misinformation content and other user activities, were transcribed from the recordings. The post-questionnaire used the same items as in the pilot study (refer to 3.2.4), in Japanese or English.

4.2.5. Procedure

There were two sessions, two weeks apart, with a different Twitter timeline each time that included 30 tweets: 20 factual and 10 fake. The 10 factual tweets changed between the first and the second session. To start, participants were given a description of the study and asked for consent. They then carried out the main task, verbally speaking aloud their assessment of each tweet as they went. Unlike in the pilot study, they were asked to read the timeline to the end. Participants were not given any information of the agent and were free to engage with it or ignore it. After the main task, they completed the post-questionnaire, and then were thanked and compensated.

4.2.6. Data Analysis

Descriptive statistics were generated. An individual accuracy score was obtained by dividing the number of accurate assessments by the number of tweets (0 to 1), one each per set of factual and fake tweets. We compared the time spent per fake tweets and during fact-checking activities by agent version and session to assess the effectiveness of each strategy and the presence of an illusory truth effect (Pennycook et al., 2018). When groups contained less than five data points, Fisher’s exact test was used instead of a Chi-square test. In addition, we used a Kruskal-Wallis test when comparing the accuracy score ratio, SUS, PU, and the time spent per fake tweet across the three conditions. For the qualitative data, we used an applied thematic analysis approach (Guest et al., 2011) with the goal of generalizing the findings by agent version and across participants. For this, two raters analyzed the data from the three open-ended questions. The primary rater added on new themes and then each rater coded 20% of the data separately. Inter-rater reliability was assessed with Cohen’s kappa, with 0.8 as the criterion for agreement (Landis and Koch, 1977). Themes that did not meet this criterion were discarded or modified. Borderline cases were resolved by discussion. Similar themes were grouped together.

4.3. Quantitative Results

Table 2. Participants’ fact-checking activities per session and agent version
Category(Checklist) Control Camouflage Consider
Author’s profile (11/19) (12/19) (13/19)
Twitter user reaction (14/19) (13/19) (17/19)
Content attachment (15/19) (14/19) (17/19)
Web searching (5/19) (6/19) (7/19)
No fact-checking (3/19) (3/19) (1/19)
Author’s profile (12/15) (12/17) (13/17)
Twitter user reaction (12/15) (17/17) (14/17)
Content attachment (12/15) (15/17) (17/17)
Web searching (6/15) (7/17) (8/17)
No fact-checking (2/15) (0/17) (0/17)
Table 3. Agent engagement frequency per session and agent version
Engagement Session 1 Session 2
Camouflage Consider Camouflage Consider
Skipped 12 18 5 3
Engaged 7 1 12 14
Active 7 0 8 7
Total 19 19 17 17

4.3.1. Influence of Agent Version on Fact-Checking Activities

Participants engaged in fact-checking activities (Table 2). For all groups, the most frequent activity in Session 1 was checking ”content attachments,” followed by ”Twitter users reactions” and ”author profiles.” Notably, those in the Consider group engaged at the same frequency in ”content attachment” and ”Twitter users reaction” activities. One-third conducted ”web searches” outside of Twitter. In the second session, the trend was reversed. All in the Camouflage group checked ”Twitter users reactions,” followed by ”content attachments,” ”author profiles,” and ”web searching.” All in the Consider group checked ”content attachments,” followed by ”Twitter users reactions,” ”author profiles,” and ”web searching.” All in the agent conditions performed at least one fact-checking activity in the second session. Regardless of condition, regular Twitter users actively sought out additional information before they judged whether the content was factual or fake.

4.3.2. Influence of Agent Version on Perceived Accuracy of Tweets

Accuracy scores for fake and factual tweets alongside exposure time to fake tweets per session and agent version are in Table 4. In the Camouflage group, seven (of 19) identified misinformation in the first session and twelve (of 17) did so in the second session. These scores were compared by agent version and across time. No statistical significance was found for accuracy score or exposure time.

4.3.3. Influence of Agent Version on Agent Engagement

Agent engagement frequencies are in Table 3. In the first session, seven (of 19) participants opened the Camouflage agent’s filter and read all false tweets. In the second session, twelve (of 17) opened up the filter and read the content at least once, and eight (of 17) uncovered nearly all content. In the Consider group, one (of 19) interacted with the agent in the first session. Fourteen (of 17) in the second session used the agent at least once, and seven (of 17) interacted with it continuously. In the first session, the Camouflage version of Elemi was interacted with more than the Consider version (p = .042), but not in the second session (p = .688).

4.3.4. Usability by Agent Version

More than half of participants in Session 1 did not interact with Elemi As such, only the scores for Session 2 were compared. The SUS for Camouflage was 5.24±1.145.24\pm 1.14 and for Consider was 4.37±0.794.37\pm 0.79. There was a statistically significant difference found between Camouflage and Consider, t(32) = 2.621, p ¡ .05, favouring Camouflage. For PU, Camouflage was 68.3±14.468.3\pm 14.4 and Consider was 53.8±11.453.8\pm 11.4. There was also a statistically significant difference found, t(32) = 3.250, p ¡ .01, indicating Camouflage was preferred.

Table 4. Content evaluations for factual and fake content, exposure time, PU, and usability by session and agent version
Session - Content Measures control Camouflage Consider
Session1 - Factual Number of participants engaged 19 19 19
Accuracy Score ratio (01)(0\sim 1) 0.75±0.160.75\pm 0.16 0.81±0.130.81\pm 0.13 0.76±0.170.76\pm 0.17
Session1 - Fake Number of participants engaged 19 7 19
Accuracy Score ratio (01)(0\sim 1) 0.85±0.150.85\pm 0.15 0.96±0.050.96\pm 0.05 0.87±0.140.87\pm 0.14
Session2 - Factual Number of participants engaged 15 17 17
Accuracy Score ratio (01)(0\sim 1) 0.75±0.160.75\pm 0.16 0.74±0.190.74\pm 0.19 0.77±0.160.77\pm 0.16
Session2 - Fake Number of participants engaged 15 12 17
Accuracy Score ratio (01)(0\sim 1) 0.83±0.170.83\pm 0.17 0.81±0.340.81\pm 0.34 0.85±0.170.85\pm 0.17
Session1 Perceived Usefulness (PU) - 5.43±1.285.43\pm 1.28 4.90±1.354.90\pm 1.35
System Usability Scale (SUS) - 72.5±14.272.5\pm 14.2 68.2±17.868.2\pm 17.8
Time spent per fake content (second) 55±3855\pm 38 49±2649\pm 26 52±3452\pm 34
Session2 Perceived Usefulness (PU) - 5.24±1.145.24\pm 1.14 4.37±0.794.37\pm 0.79
System Usability Scale (SUS) - 68.3±14.468.3\pm 14.4 53.8±11.453.8\pm 11.4
Time spent per fake content (second) 60±4760\pm 47 75±3575\pm 35 69±3669\pm 36

4.4. Qualitative Findings

Here we present a summary of our thematic analysis findings, including a frequency analysis for each theme.

4.4.1. Dis/misuse

Use of Elemi varied by version, i.e., the metacognitive strategy it deployed. When it came to the Camouflage version of Elemi, two participants were not aware of its existence. According to P40, ”I wasn’t even aware of the presence of an agent.” Among those who used Elemi, one (P18) expressed concern about dependence on the agent’s judgments when in Camouflage form: ”I tend to judge by the content of the agent and less by the content of the tweets.” In contrast, the Consider version was intentionally not used by five participants. According to P50: ”I dislike it because it just provides opinions, there’s little objective about it.”

4.4.2. Usability and UX

Participants had various reactions to the user friendliness of the agent forms. Nine users of the Camouflage version expressed that the agent was easy to use. As per P54: ”It’s lightweight and easy to learn.” Regarding its transparency and explainability, two user who never opened up its filter expressed positive opinions about its explainability. According to P41: ”It’s easy to use and integrated well with the social media platform.” Still, eight (of 12) who engaged with the agent expressed suspicion about its transparency and explainability. For instance, P13 wrote: ”I doubt the authenticity of the agent’s judgment …” Another advised that the Camouflage version of Elemi may be a fit with other platforms, especially educational ones: ”I think that detecting lies is a necessary skill when you grow up, so I feel it would be good to make it for children” (P26). For two Consider users, the agent was friendly. As per P8: ”The design is catchy and friendly.” But when it came to ease-of-use, eleven (of 14) participants noted that the agent was not so easy to use. They all criticized the reasoning process. As P11 described: ”I dislike having to go through each step …”

4.4.3. Agent Performance

Ten participants spoke to the Camouflage version of Elemi. While six said that the accuracy of the fact-checking was high, the other four indicated that some of the agent’s judgments were incorrect. According to P16: ”I feel that the sources for judging misinformation were somewhat scarce …” In short, more fact-checking was desired. At the same time, no comments were made about the Consider version of the agent or other aspects of its performance.

4.4.4. Dis/advantages

Both versions of the agent were deemed good for decision-making support, i.e., as metacognitive tools. Still, the perceptions of Elemi’s capability differed by version. Participants viewed the Camouflage version of Elemi as an agent that confirmed the correctness or incorrectness of information. In contrast, Consider was characterized as an agent that encouraged a rational decision: a different method of assistance. For Camouflage, six participants commented that making incorrect information invisible would lead to platform comfort. According to P2: ”The collection of tweets with credible sources was pleasant to see.”. Two said that the Camouflage version of Elemi led them to be more aware of fake news. As per P33: ”The warning made me more alert.”. Two others said that, on the contrary, their interest in the content of fake news rose. P31 explained: ”Using an agent arouses interest in the fake news it’s hiding.” Five Consider participants said that the agent inspired them to think logically. According to P7: ”It helped me to judge logically rather than intuitively.” However, seven participants felt that there was a lack of objectivity in the agent’s decisions. As per P52: ”It’s mainly or only opinion, there’s little objective about it.” For P18, the Consider version made it difficult to use the platform as casually as usual because it encouraged them to think about the information: …”I agree that fact-checking is important, but I felt that the value of Twitter as entertainment decreased …”

4.5. Discussion

We now discuss our findings based on the research questions.

4.5.1. Does Camouflage work better than Consider at helping people identify misinformation? (RQ1)

Leaving engagement with the agent up to the participant led to low engagement in the first session. Still, the accuracy score ratio of the Consider version of the agent did not increase in the second session. As such, this strategy did not lead people to deliberate on the content as described by Bago, Rand, and Pennycook (Bago et al., 2020). As for Camouflage, while it worked in the first session, its efficacy was disrupted in the second, apparently when tweets were repeated over time, i.e., repeated exposure, in terms of both accuracy score and user doubt. Transparency and clear explanations about the agent’s decisions (Hartwig and Reuter, 2019), such as misinformation keywords and fact-checking references, did not prevent perceptions of the performance of the Camouflage version of the agent from degrading on re-exposure to the same misinformation.

A phased, time-based approach to assessing the agent revealed the weakness of the Camouflage version, indicating that it would not be superior to the Consider strategy in the wild, at least in terms of helping people distinguish between factual and misinformed tweets to which they were exposed multiple times. Indeed, neither Camouflage nor Consider alone were ideal as an anti-misinformation strategy over time. Camouflage, which provided fact-checking data from the agent, was increasingly questioned by users from session to session. Given that agent behaviour can make users skeptical (Chen et al., 2022), repeated usage could increase the chance of users encountering inconsistencies with agent decision-making. These inconsistencies could exacerbate perceptions of distrust towards agent performance. Consider, which required cognitive effort and made people think for themselves, was described as hard to use and inconvenient, having marginal acceptability and decreasing the entertainment value of social media. Consider users also wanted objective information to support their decision-making rather than engaging in reasoning activities while reading the content. In the absence of objective feedback, they rated the agent’s usefulness as poor, equivalent to the level of the ”will not use” group in the pilot study. In Twitter, where content is instantaneous and unchallenged (Starbird et al., 2010), taking on cognitive load was opposed by participants. Therefore, a more practical approach could be to combine strategies, harmonizing the two agent versions as one and potentially including other strategies that we did not assess.

4.5.2. Does an agent employing a metacognitive approach mitigate the illusory truth effect and to what degree? (RQ2)

Changes in participants behaviour were captured over the course of two sessions. Unlike in previous research (Pennycook et al., 2018; Hassan and Barber, 2021), the illusory truth effect was not found. Still, a notable difference was found in the Camouflage group. Although it did not reach statistical significance as in Chen et al. (Chen et al., 2022), the Camouflage group tended to be more skeptical than those in the Consider and control groups over time. The metacognitive cue that Camouflage used, blurring certain tweets as an indicator of misinformation and thereby implying that the rest was factual, was questioned by participants. The more inconsistent it was felt to be over time, the more that participants felt distrust towards the agent. Some even felt that the agent was misguiding them on purpose. The number of participants who disagreed with the agent doubled from the first to the second session (2 to 4). We might call this an amplified disbelief effect. In short, those who had doubts about the agent’s performance the first time were inclined to feel twice as doubtful the second time around. Thus, there is a chance that those who mistrust an agent from the start will further mistrust the agent in subsequent sessions.

5. General Discussion

This research aimed at empirically evaluating an ”exoskeleton for the mind” in the form of Elemi, a metacognitive agent deploying several anti-misinformation strategies in the live context of Twitter. As a first step, we drew from existing work to create a Cover version of the agent that filtered content. Pilot testing led us to improve this design and extend it with theories of metacognition applied to the digital realm of social media. Cover became Camouflage, which we empirically evaluated in a phased experimental study alongside a new version of the agent, Consider. This version deployed a ”pause and reflect” strategy that captured the full cycle of theorized metacognitive activities. We compared each version to a control. Overall, we found no statistically significant differences in terms of participant activities, misinformation exposure time, and accuracy score rating. However, people rated the Camouflage version of Elemi much higher than its predecessor and its counterpart, Consider. We also did not find evidence of an illusory truth effect, although we did find evidence of another effect for those who claimed reason to distrust the agent from the start: what we call amplified disbelief effect. We discuss the implications below.

5.1. Still In Search of Strategies

Classic countermeasures for misinformation, namely warning and filtering, are deployed based on the severity of the content in question. Yet, in modern social media contexts, countless people become sources of information on the fly, making it difficult to identify and respond to misinformation in a timely and effective manner (Grinberg et al., 2019). Automated methods, such as Meta/Facebook’s approach777https://www.facebook.com/formedia/mjp/programs/third-party-fact-checking, are often deployed, but such methods are known to be slow (Pennycook et al., 2018; Fazio, 2020) and require trust in the company. Likewise, the Camouflage version of our agent revealed cases where users questioned the agent’s filtering and had doubts about whether the filtered information was misinformation, thus inspiring direct engagement with the content the second time around. This amplified disbelief effect decreases the utility of such agents, even biasing user decisions towards misinformation. A counter strategy is needed to reap the rewards of a Camouflage strategy.

However, this does not necessarily mean that user-driven strategies work better than the agent-driven ones. Several anti-misinformation strategies and psychological studies on the ”pause and reflect” approach have shown high effectiveness for user vigilance towards fake news (Bago et al., 2020; Fazio, 2020). However, it seems that inculcating ”metacognitive behaviour” into social media platforms requires a number of hurdles to be overcome. The scoring ratio patterns between the control and Consider conditions in our work indicate that forcing deliberation within the social media context went nowhere (Bago et al., 2020). The Consider version of Elemi was rated as unfit for a casual social media context, and forced cognitive engagement was tiresome. Future anti-misinformation and/or metacognitive agent designs should take into account how to motivate users. As we confirmed, when given leave to act freely, Twitter users were uncooperative, especially when it came to reasoning activities and even under the observation a researcher. This implies that they could do worse when alone.

These findings suggest that anti-misinformation and/or metacognitive strategies should be integrated into agents in a way that complements user needs and expectations with the given social media platform. Still, hiding fake content, using either a Cover or Camouflage strategy, was particularly useful for preventing exposure to false information, at least at first. Positive feedback alongside high PU and usability scores were maintained as long as the filtering allowed for user autonomy and accessibility. Still, we need to find a way to motivate users to not open the filter needlessly. One example could be combining a user-driven evaluation approach, such as letting users rate on the first encounter, and then reminding them of this rating on subsequent exposure to the content, ideally with supporting articles and sources for fact-checking purposes. This would also address user feedback on the Consider form of the agent, in terms of a desire for more objective feedback. The agent should certainly prepare sufficient evidence as to why a given piece of content is fake (or not) before users are asked to judge that content, so as not to generate suspicion.

Another point to bear in mind is when to have the agent help the user confront misinformation. Chen et al. (Chen et al., 2022) showed that users became more skeptical of information while iterating through judgments of whether the information was true or false. This was also found for the case of the Camouflage version of Elemi. In this respect, the role of Camouflage is limited. Judgments involve cognitive load, so user-driven methods should consider how to exercise cognition in a way that leads to metacognition activities that are effective and not overwhelming.

5.2. Implications for Design

We now summarize our findings in a set of general design implications for similar in situ anti-misinformation tools that aim to include elements of metacognition:

  • Hidden misinformation should still be accessible to users for reasons of trust, autonomy, and transparency.

  • Filtering should be accompanied by clear scaffolding to prevent repeated access to the misinformation, or at least contextualize the user’s previous reactions to that content, such as by providing the user’s original ratings.

  • A repeated mismatch between user and agent judgments and decisions on the misinformed nature of content or fact-checking activities could induce an amplified disbelief effect that biases the user towards, rather than away from, misinformation. Counter strategies, personalization, and reliance on objective sources should be explored.

  • Forced cognitive engagement may not be welcome by most users, especially regular Twitter users.

  • How to engage users, especially in consideration of cognition load within the casual context of many social media platforms, should be considered and tested with respect to helping users remain vigilant to misinformation. Users should be guided towards metacognitive activities, if possible.

5.3. Limitations

Not all participants completed the second session, limiting the time-based results. Additionally, most participants were students in engineering, so future research should recruit a more diverse sample. It is also possible that the topic of COVID-19, a known polarizing topic, influenced impressions of the agent itself. Future work can include different topics, hashtags, and tweets. We also made important decisions about the design of the agent and the research design, notably opting for a balance of realism and experimental control in light of practical limitations with the prototype. Our findings indicate that this was beneficial in allowing participants to feel comfortable and freely provide their opinions, as well as revealing unusual results, especially over time with the nullified illusory truth effect. Nevertheless, future work should fully automate the agent with a live fact-checking service and re-test with user’s real accounts and timelines.

6. Conclusion

Anti-misinformation tools are on the rise. Many strategies and ideas have been tested experimentally, primarily outside of the wilderness of social media platforms like Twitter, with middling or unclear results. We aimed to increase the ecological validity of this body of work and directly compare the efficacies of promising strategies. Our work on Elemi shows that such tools may need special care to work in live social media environments with a range of potential end-users. Future work can explore the design implications proposed with diverse personalities on a greater range of platforms in an autonomous form distributed and tested in live environments with non-curated content.

Acknowledgements.
This work was funded by a DLab Challenge: Laboratory for Design of Social Innovation in Global Networks (DLab) Research Grant (Tokyo Institute of Technology). We thank the Aspirational Computing Lab for early feedback.

References

  • (1)
  • Allcott and Gentzkow (2017) Hunt Allcott and Matthew Gentzkow. 2017. Social media and fake news in the 2016 election. Journal of Economic Perspectives 31, 2 (2017), 211–36.
  • Allcott et al. (2019) Hunt Allcott, Matthew Gentzkow, and Chuan Yu. 2019. Trends in the diffusion of misinformation on social media. Research & Politics 6, 2 (2019), 2053168019848554.
  • Bago et al. (2020) Bence Bago, David G Rand, and Gordon Pennycook. 2020. Fake news, fast and slow: Deliberation reduces belief in false (but not true) news headlines. Journal of Experimental Psychology: general 149, 8 (2020), 1608–1613.
  • Bangor et al. (2008) Aaron Bangor, Philip T Kortum, and James T Miller. 2008. An empirical evaluation of the system usability scale. Intl. Journal of Human–Computer Interaction 24, 6 (2008), 574–594.
  • Bautista et al. (2021) John Robert Bautista, Yan Zhang, and Jacek Gwizdka. 2021. Healthcare professionals’ acts of correcting health misinformation on social media. International Journal of Medical Informatics 148 (2021), 104375.
  • Benbasat and Wang (2005) Izak Benbasat and Weiquan Wang. 2005. Trust in and adoption of online recommendation agents. Journal of the Association for Information Systems 6, 3 (2005), 72–101.
  • Bhuiyan et al. (2018) Md Momen Bhuiyan, Kexin Zhang, Kelsey Vick, Michael A Horning, and Tanushree Mitra. 2018. FeedReflect: A Tool for Nudging Users to Assess News Credibility on Twitter. In Companion of the 2018 ACM Conference on Computer Supported Cooperative Work and Social Computing. ACM, Jersey City NJ USA, 205–208.
  • Bleakley (2021) Paul Bleakley. 2021. Panic, pizza and mainstreaming the alt-right: A social media analysis of Pizzagate and the rise of the QAnon conspiracy. Current Sociology (2021), 00113921211034896.
  • Braun and Clarke (2006) Virginia Braun and Victoria Clarke. 2006. Using thematic analysis in psychology. Qualitative Research in Psychology 3, 2 (2006), 77–101.
  • Brooke et al. (1996) John Brooke et al. 1996. SUS-A quick and dirty usability scale. Usability evaluation in industry 189, 194 (1996), 4–7.
  • Chen et al. (2022) Guangyu Chen, Paolo Ciuccarelli, and Sara Colombo. 2022. VisualBubble: Exploring How Reflection-Oriented User Experiences Affect Users’ Awareness of Their Exposure to Misinformation on Social Media. In CHI Conference on Human Factors in Computing Systems Extended Abstracts. ACM, New Orleans LA USA, 1–7.
  • Chi and Liao (2022) Haixiao Chi and Beishui Liao. 2022. A quantitative argumentation-based Automated eXplainable Decision System for fake news detection on social media. Knowledge-Based Systems 242 (2022), 108378.
  • Colleoni et al. (2014) Elanor Colleoni, Alessandro Rozza, and Adam Arvidsson. 2014. Echo chamber or public sphere? Predicting political orientation and measuring political homophily in Twitter using big data. Journal of communication 64, 2 (2014), 317–332.
  • Cusumano et al. (2021) Michael A Cusumano, ANNABELLE REINA GAWER, and David B Yoffie. 2021. Social Media Companies Should Self-Regulate. Now. Harvard Business Review (2021).
  • Diakopoulos et al. (2021) Nicholas Diakopoulos, Daniel Trielli, and Grace Lee. 2021. Towards understanding and supporting journalistic practices using semi-automated news discovery tools. Proceedings of the ACM on Human-Computer Interaction 5, CSCW2 (2021), 1–30.
  • Dunlosky and Metcalfe (2008) John Dunlosky and Janet Metcalfe. 2008. Metacognition. Sage Publications, Los Angeles.
  • Epstein et al. (2020) Ziv Epstein, Gordon Pennycook, and David Rand. 2020. Will the crowd game the algorithm? Using layperson judgments to combat misinformation on social media by downranking distrusted sources. In Proceedings of the 2020 CHI conference on Human Factors in Computing Systems. ACM, Honolulu HI USA, 1–11.
  • Ericsson and Simon (1984) K Anders Ericsson and Herbert A Simon. 1984. Protocol Analysis: Verbal Reports as Data. MIT Press, Cambridge, MA, US.
  • Fazio (2020) Lisa Fazio. 2020. Pausing to consider why a headline is true or false can help reduce the sharing of false news. Harvard Kennedy School Misinformation Review 1, 2 (2020).
  • Garrett (2021) Major Garrett. 2021. ” It was a drug”: Capitol riot exposes reach of QAnon disinformation.”. CBS News 31 (2021).
  • Geeng et al. (2020) Christine Geeng, Savanna Yee, and Franziska Roesner. 2020. Fake News on Facebook and Twitter: Investigating How People (Don’t) Investigate. In Proceedings of the 2020 CHI conference on human factors in computing systems. ACM, New York, NY, USA, 1–14.
  • Grinberg et al. (2019) Nir Grinberg, Kenneth Joseph, Lisa Friedland, Briony Swire-Thompson, and David Lazer. 2019. Fake news on Twitter during the 2016 US presidential election. Science 363, 6425 (2019), 374–378.
  • Guest et al. (2011) Greg Guest, Kathleen M MacQueen, and Emily E Namey. 2011. Applied Thematic Analysis. Sage, Thousand Oaks, CA, USA.
  • Hartwig and Reuter (2019) Katrin Hartwig and Christian Reuter. 2019. TrustyTweet: An Indicator-based Browser-Plugin to Assist Users in Dealing with Fake News on Twitter. In Proceedings of the International Conference on Wirtschaftsinformatik (WI). Association for Information Systems (AIS).
  • Hassan and Barber (2021) Aumyo Hassan and Sarah J Barber. 2021. The effects of repetition frequency on the illusory truth effect. Cognitive research: principles and implications 6, 1 (2021), 1–12.
  • Hernon (1995) Peter Hernon. 1995. Disinformation and misinformation through the internet: Findings of an exploratory study. Government information quarterly 12, 2 (1995), 133–139.
  • Hsee and Ruan (2016) Christopher K Hsee and Bowen Ruan. 2016. The Pandora effect: The power and peril of curiosity. Psychological science 27, 5 (2016), 659–666.
  • Kim and Walker (2020) H. Kim and D. Walker. 2020. Leveraging volunteer fact checking to identify misinformation about COVID-19 in social media. Harvard Kennedy School (HKS) Misinformation Review (2020). https://doi.org/10.37016/mr-2020-021
  • Kouzy et al. (2020) Ramez Kouzy, Joseph Abi Jaoude, Afif Kraitem, Molly B El Alam, Basil Karam, Elio Adib, , Jabra Zarka, Cindy Traboulsi, Elie W Akl, and Khalil Baddour. 2020. Coronavirus goes viral: Quantifying the covid-19 misinformation epidemic on twitter. Cureus 12, 3 (2020), e7255. https://doi.org/10.7759/cureus.7255
  • Kozyreva et al. (2020) Anastasia Kozyreva, Stephan Lewandowsky, and Ralph Hertwig. 2020. Citizens versus the internet: Confronting digital challenges with cognitive tools. Psychological Science in the Public Interest 21, 3 (2020), 103–156.
  • Kray et al. (2002) Laura J Kray, Adam D Galinsky, and Leigh Thompson. 2002. Reversing the gender gap in negotiations: An exploration of stereotype regeneration. Organizational behavior and human decision processes 87, 2 (2002), 386–409.
  • Landis and Koch (1977) J Richard Landis and Gary G Koch. 1977. The measurement of observer agreement for categorical data. Biometrics (1977), 159–174.
  • Mazerolle et al. (2012) Marie Mazerolle, Isabelle Régner, Pauline Morisset, François Rigalleau, and Pascal Huguet. 2012. Stereotype threat strengthens automatic recall and undermines controlled processes in older adults. Psychological science 23, 7 (2012), 723–727.
  • Mena (2020) Paul Mena. 2020. Cleaning up social media: The effect of warning labels on likelihood of sharing false news on Facebook. Policy & Internet 12, 2 (2020), 165–183.
  • Niemiec (2020) Emilia Niemiec. 2020. COVID-19 and misinformation: Is censorship of social media a remedy to the spread of medical misinformation? EMBO reports 21, 11 (2020), e51420.
  • Pennycook et al. (2018) Gordon Pennycook, Tyrone D Cannon, and David G Rand. 2018. Prior exposure increases perceived accuracy of fake news. Journal of Experimental Psychology: General 147, 12 (2018), 1865.
  • Pennycook et al. (2021) Gordon Pennycook, Ziv Epstein, Mohsen Mosleh, Antonio A Arechar, Dean Eckles, and David G Rand. 2021. Shifting attention to accuracy can reduce misinformation online. Nature 592, 7855 (2021), 590–595.
  • Ren et al. (2023) Zhiying Bella Ren, Eugen Dimant, and Maurice Schweitzer. 2023. Beyond belief: How social engagement motives influence the spread of conspiracy theories. Journal of Experimental Social Psychology 104 (2023), 104421.
  • Ross et al. (2018) Björn Ross, Anna Jung, Jennifer Heisel, and Stefan Stieglitz. 2018. Fake news on social media: The (in) effectiveness of warning messages. In International Conference on Information Systems (ICIS). AIS, Atlanta, GA, USA, Article 16.
  • Saltz et al. (2021) Emily Saltz, Claire R Leibowicz, and Claire Wardle. 2021. Encounters with Visual Misinformation and Labels Across Platforms: An Interview and Diary Study to Inform Ecosystem Approaches to Misinformation Interventions. In Extended Abstracts of the 2021 CHI Conference on Human Factors in Computing Systems. ACM, New York, NY, USA, 1–6.
  • Sharevski et al. (2021) Filipo Sharevski, Raniem Alsaadi, Peter Jachim, and Emma Pieroni. 2021. Misinformation Warning Labels: Twitter’s Soft Moderation Effects on COVID-19 Vaccine Belief Echoes. arXiv preprint arXiv:2104.00779 (2021).
  • Shin et al. (2018) Jieun Shin, Lian Jian, Kevin Driscoll, and François Bar. 2018. The diffusion of misinformation on social media: Temporal pattern, message, and source. Computers in Human Behavior 83 (2018), 278–287.
  • Starbird et al. (2010) Kate Starbird, Leysia Palen, Amanda L Hughes, and Sarah Vieweg. 2010. Chatter on the red: what hazards threat reveals about the social life of microblogged information. In Proceedings of the 2010 ACM conference on Computer supported cooperative work. ACM, New York, NY, USA, 241–250.
  • Stevens (2007) Martin Stevens. 2007. Predator perception and the interrelation between different forms of protective coloration. Proceedings of the Royal Society B: Biological Sciences 274, 1617 (2007), 1457–1464.
  • Stjernfelt and Lauritzen (2020) Frederik Stjernfelt and Anne Mette Lauritzen. 2020. Your post has been removed: Tech giants and freedom of speech. Springer Nature, London, UK.
  • Vijjali et al. (2020) Rutvik Vijjali, Prathyush Potluri, Siddharth Kumar, and Sundeep Teki. 2020. Two stage transformer model for COVID-19 fake news detection and fact checking. arXiv preprint arXiv:2011.13253 (2020).
  • Vraga and Bode (2017) Emily K Vraga and Leticia Bode. 2017. Using expert sources to correct health misinformation in social media. Science Communication 39, 5 (2017), 621–645.
  • Vraga et al. (2019) Emily K Vraga, Sojung Claire Kim, and John Cook. 2019. Testing logic-based and humor-based corrections for science, health, and political misinformation on social media. Journal of Broadcasting & Electronic Media 63, 3 (2019), 393–414.
  • Vraga et al. (2020) Emily K. Vraga, Sojung Claire Kim, John Cook, and Leticia Bode. 2020. Testing the effectiveness of correction placement and type on Instagram. The International Journal of Press/Politics 25, 4 (2020), 632–652. https://doi.org/10.1177/1940161220919082
  • Walter et al. (2020) Nathan Walter, John J Brooks, Camille J Saucier, and Sapna Suresh. 2020. Evaluating the impact of attempts to correct health misinformation on social media: A meta-analysis. Health Communication 36 (2020), 1–9. Issue 13.
  • Wang et al. (2016) Weiquan Wang, Lingyun Qiu, Dongmin Kim, and Izak Benbasat. 2016. Effects of rational and social appeals of online recommendation agents on cognition-and affect-based trust. Decision Support Systems 86 (2016), 48–60.
  • Wu et al. (2016) Liang Wu, Fred Morstatter, Xia Hu, and Huan Liu. 2016. Mining misinformation in social media. Chapman and Hall, UK, 135–162.
  • Zarocostas (2020) John Zarocostas. 2020. How to fight an infodemic. The Lancet 395, 10225 (2020), 676.