This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

CrowdChecked: Detecting Previously Fact-Checked Claims in Social Media

Momchil Hardalov1  Anton Chernyavskiy2
Ivan Koychev1  Dmitry Ilvovsky2  Preslav Nakov3
1Sofia University “St. Kliment Ohridski”, Bulgaria
2HSE University, Russia
3Mohamed bin Zayed University of Artificial Intelligence, UAE
{hardalov, koychev}@fmi.uni-sofia.bg
{acherniavskii, dilvovsky}@hse.ru

[email protected]
Abstract

While there has been substantial progress in developing systems to automate fact-checking, they still lack credibility in the eyes of the users. Thus, an interesting approach has emerged: to perform automatic fact-checking by verifying whether an input claim has been previously fact-checked by professional fact-checkers and to return back an article that explains their decision. This is a sensible approach as people trust manual fact-checking, and as many claims are repeated multiple times. Yet, a major issue when building such systems is the small number of known tweet–verifying article pairs available for training. Here, we aim to bridge this gap by making use of crowd fact-checking, i.e., mining claims in social media for which users have responded with a link to a fact-checking article. In particular, we mine a large-scale collection of 330,000 tweets paired with a corresponding fact-checking article. We further propose an end-to-end framework to learn from this noisy data based on modified self-adaptive training, in a distant supervision scenario. Our experiments on the CLEF’21 CheckThat! test set show improvements over the state of the art by two points absolute. Our code and datasets are available at https://github.com/mhardalov/crowdchecked-claims

CrowdChecked: Detecting Previously Fact-Checked Claims in Social Media


Momchil Hardalov1  Anton Chernyavskiy2 Ivan Koychev1  Dmitry Ilvovsky2  Preslav Nakov3 1Sofia University “St. Kliment Ohridski”, Bulgaria 2HSE University, Russia 3Mohamed bin Zayed University of Artificial Intelligence, UAE {hardalov, koychev}@fmi.uni-sofia.bg {acherniavskii, dilvovsky}@hse.ru [email protected]


1 Introduction

The massive spread of disinformation online, especially in social media, was counter-acted by major efforts to limit the impact of false information not only by journalists and fact-checking organizations but also by governments, private companies, researchers, and ordinary Internet users. This includes building systems for automatic fact-checking  (Zubiaga et al., 2016; Derczynski et al., 2017; Nakov et al., 2021a; Gu et al., 2022; Guo et al., 2022; Hardalov et al., 2022), fake news (Ferreira and Vlachos, 2016; Nguyen et al., 2022), and fake news website detection Baly et al. (2020); Stefanov et al. (2020); Panayotov et al. (2022).

Refer to caption
Figure 1: Crowd fact-checking thread on Twitter. The first tweet (Post w/ claim) makes the claim that Ivermectin causes sterility in men, which then receives replies. A (crowd) fact-checker replies with a link to a verifying article from a fact-checking website. We pair the article with the tweet that made this claim (the first post ), as it is irrelevant () to the other replies.

Unfortunately, fully automatic systems still lack credibility, and thus it was proposed to focus on detecting previously fact-checked claims instead: Given a user comment, detect whether the claim it makes was previously fact-checked with respect to a collection of verified claims and their corresponding articles (see Table 1). This task is an integral part of an end-to-end fact-checking pipeline Hassan et al. (2017), and also an important task on its own right as people often repeat the same claim Barrón-Cedeño et al. (2020b); Vo and Lee (2020); Shaar et al. (2021). Research on this problem is limited by data scarceness, with datasets typically having about a 1,000 tweet–verifying article pairs Barrón-Cedeño et al. (2020b); Shaar et al. (2020, 2021), with the notable exception of Vo and Lee (2020), which contains 19K claims about images matched against 3K fact-checking articles.

We propose to bridge this gap using crowd fact-checking to create a large collection of tweet–verifying article pairs, which we then label (if the pair is correctly matched) automatically using distant supervision. An example is shown in Figure 1.

Our contributions are as follows:

  • we mine a large-scale collection of 330,000 tweets paired with fact-checking articles;

  • we propose two distant supervision strategies to label the CrowdChecked dataset;

  • we propose a novel method to learn from this data using modified self-adaptive training;

  • we demonstrate sizable improvements over the state of the art on a standard test set.

2 Our Dataset: CrowdChecked

2.1 Dataset Collection

We use Snopes as our target fact-checking website, due to its popularity among both Internet users and researchers Popat et al. (2016); Hanselowski et al. (2019); Augenstein et al. (2019); Tchechmedjiev et al. (2019). We further use Twitter as the source for collecting user messages, which could contain claims and fact-checks of these claims.

Our data collection setup is similar to the one in Vo and Lee (2019). First, we form a query to select tweets that contain a link to a fact-check from Snopes (url:snopes.com/fact-check/), which is either a reply or a quote tweet, and not a retweet. An example result from the query is shown in Figure 1, where the tweet from the crowd fact-checker contains a link to a fact-checking article. We then assess its relevance to the claim (if any) made in the first tweet (the root of the conversation) and the last reply in order to obtain tweet–verified article pairs. We analyze in more detail the conversational structure of these threads in Section 2.2.

User Post w/ Claim: Sen. Mitch McConnell: “As recently as October, now-President Biden said you can’t legislate by executive action unless you are a dictator. Well, in one week, he signed more than 30 unilateral actions.” [URL] — Forbes (@Forbes) January 28, 2021
Verified Claims and their Corresponding Articles
(1) When he was still a candidate for the presidency in October 2020, U.S. President Joe Biden said, “You can’t legislate by executive order unless you’re a dictator.” http://snopes.com/fact-check/biden-executive-order-dictator/
(2) U.S. Sen. Mitch McConnell said he would not participate in 2020 election debates that include female moderators. http://snopes.com/fact-check/mitch-mcconnell-debate-female/
Table 1: Illustrative examples for the task of detecting previously fact-checked claims. The post contains a claim (related to legislation and dictatorship), the Verified Claims are part of a search collection of previous fact-checks. In row (1), the fact-check is a correct match for the claim made in the tweet (), whereas in (2), the claim still discusses Sen. Mitch McConnell, but it is a different claim (), and thus this is an incorrect pair.

We collected all tweets matching our query from October 2017 till October 2021, obtaining a total of 482,736 unique hits. We further collected 148,503 reply tweets and 204,250 conversation (root) tweets.111The sum of the unique replies and of the conversation tweets is not equal to the total number of fact-checking tweets, as more than one tweet might reply to the same comment. Finally, we filter out malformed pairs, i.e., tweets linking to themselves, empty tweets, non-English ones, such with no resolved URLs in the Twitter object (‘entities’), with broken links to the fact-checking website, and all tweets in the CheckThat ’21 dataset. We ended up with 332,660 unique tweet–article pairs (shown in first row in Table 5), 316,564 unique tweets, and 10,340 fact-checking articles from Snopes they point to.

More detail about the process of collecting fact-checking articles as well as detailed statistics are given in Appendix B.1 and on Figure 2.

2.2 Tweet Collection

(Conversation Structure) It is important to note that the ‘fact-checking’ tweet can be part of a multiple-turn conversational thread, therefore taking the post that it replies to (previous turn), does not always express a claim which the current tweet targets. In order to better understand this, we performed manual analysis of some conversational threads. Conversational threads in Twitter are organized as shown Figure 1: the root is the first comment, then there can be a long discussion, followed by a fact-checking comment (i.e., the one with a link to a fact-checking article on Snopes). In our analysis, we identify four patterns: (i) the current tweet verifies a claim in the tweet it replies to, (ii) the tweet verifies the root of the conversation, (iii) the tweet does not verify any claim in the chain (a common scenario), and (iv) the fact-check targets a claim that was not expressed in the root or in the closest tweet (this was in very few cases). This analysis suggests that for the task of detecting previously fact-checked claims, it is sufficient to collect the triplet of the fact-checking tweet, the root of the conversation (conversation), and the tweet that the target tweet is replying to (reply).

Dataset Tweets Words Vocab
||Unique|| Mean 50% Max ||Unique||
CrowdChecked (Ours) 316,564 12.2 11 60 114,727
CheckThat ’21 1,399 17.5 16 62 9,007
Table 2: Statistics about our dataset vs. CheckThat ’21. The number of unique tweets is lower than the total number of tweet–article pairs, as an input tweet could be fact-checked by multiple articles.

2.3 Comparison to Existing Datasets

We compare our dataset to a closely related dataset from the CLEF-2021 CheckThat ’21 on Detecting Previously Fact-Checked Claims in Tweets (Shaar et al., 2021), to which we will refer as CheckThat ’21 in the rest of the paper. There exist other related datasets that are smaller Barrón-Cedeño et al. (2020b), come from a different domain Shaar et al. (2021), are not in English Elsayed et al. (2019), or are multi-modal (Vo and Lee, 2020).

Table 2 compares our CrowdChecked to CheckThat ’21 in terms of number of examples, length of the tweets, and vocabulary size. Before calculating these statistics, we lowercased the text and we removed all URLs, Twitter handlers, English stop words, and punctuation. We can see in Table 2 that CrowdChecked contains two orders of magnitude more examples, slightly shorter tweets (but the maximum length stays approximately the same, which can be explained by the word limit of Twitter), and has a vocabulary size that is an order of magnitude larger. Note, however, that many examples in CrowdChecked are incorrect matches (see Section 2.1), and thus we use distant supervision to label them (see Section 2.4), with the resulting dataset sizes of matching pairs shown in Table 5. Here, we want to emphasize that there is absolutely no overlap at all between CrowdChecked and CheckThat ’21 in terms of tweets/claims.

In terms of topics, the claims in both our dataset and CheckThat ’21 are quite diverse, including fact-checks for a broad set of topics related, but not limited to politics (e.g., the Capitol Hill riots, US elections), pop culture (e.g., famous performers and actors such as Drake and Leonardo di Caprio), brands (e.g., McDonald’s and Disney), and COVID-19, among many others. Illustrative examples of the claim/topic diversity can be found in Tables 1 and 10 (in the Appendix). Moreover, the collection of Snopes articles contains almost 14K different fact-checks on an even wider range of topics, which further diversifies the set of tweet–article pairs.

Finally, we compare the set of Snopes fact-checking articles referenced by the crowd fact-checkers to the ones included in the CheckThat ’21 competition. We can see that the tweets in CrowdChecked refer to less articles (namely 10,340), compared to CheckThat ’21, which consists of 13,835 articles. A total of 8,898 articles are present in both datasets. Since the CheckThat ’21 is collected earlier, it includes less articles from recent years compared to CrowdChecked, and peaks at 2016/2017. Nevertheless, for CheckThat ’21, the number of Snopes articles included in a claim–article pair is far less compared to our dataset (even after filtering out unrelated pairs), as it is capped at the number of tweets included in that dataset (which is 1.4K).

Refer to caption
Figure 2: Histogram of the year of publication of the Snopes articles included in CrowdChecked (our dataset) vs. those in CheckThat ’21.

More detail about the process of collecting the fact-checking articles is given in Appendix B.1.

2.4 Data Labeling (Distant Supervision)

To label our examples, we experiment with two distant supervision approaches: (i) based on the Jaccard similarity between the tweet and the target fact-checking article, and (ii) based on the predictions of a model trained on CheckThat ’21.

Jaccard Similarity

In this approach, we first pre-process the texts by converting them to lowercase, removing all URLs and replacing all numbers with a single zero. Then, we tokenize them using NLTK’s Twitter tokenizer Loper and Bird (2002), and we strip all handles and user mentions. Finally, we filter out all stop words and punctuation (including quotes and special symbols) and we stem all tokens using the Porter stemmer Porter (1980).

In order to obtain a numerical score for each tweet–article pair, we calculate the Jaccard similarity (jac) between the normalized tweet text and each of the title and the subtitle from the Snopes article (i.e., the intersection over the union of the unique tokens). Both fields present a summary of the fact-checked claim, and thus should include more compressed information. Finally, we average these two similarity values to obtain a more robust score. Statistics are shown in Table 3.

Range Examples Correct Pairs Correct Pairs
(Jaccard) (%) Reply (%) Conv. (%)
[0.0;0.1) 62.57 5.88 0.00
[0.1;0.2) 18.98 36.36 14.29
[0.2;0.3) 10.21 46.67 50.00
[0.3;0.4) 4.17 76.47 78.57
[0.4;0.5) 2.33 92.86 92.86
[0.5;0.6) 1.08 94.12 94.12
[0.6;0.7) 0.43 80.00 80.00
[0.7;0.8) 0.11 92.31 92.31
[0.8;0.9) 0.05 91.67 92.86
[0.9;1.0] 0.02 100.00 100.00
Table 3: Proportion of examples in different bins based on average Jaccard similarity between the tweet and the title/subtitle. Manual annotations of the correct pairs.

Semi-Supervision

Here, we train a SentenceBERT Reimers and Gurevych (2019) model, as described in Section 3, using the manually annotated data from CheckThat ’21. The model shows strong performance on the testing set of CheckThat ’21 (see Table 6), and thus we expect it to have good precision at detecting matching fact-checked pairs. In particular, we calculate the cosine similarity between the embeddings of the fact-checked tweet and the fields from the Snopes article. Statistics about the scores are shown in Table 4.

Range Examples Correct Pairs
(Cosine) (%) (%)
[-0.4;0.1) 37.83 0.00
[0.1;0.2) 16.50 6.67
[0.2;0.3) 12.28 41.46
[0.3;0.4) 10.12 36.36
[0.4;0.5) 8.58 63.16
[0.5;0.6) 6.69 70.00
[0.6;0.7) 4.47 84.21
[0.7;0.8) 2.48 96.15
[0.8;0.9) 0.97 93.10
[0.9;1.0] 0.08 100.00
Table 4: Proportion of examples in different bins based on cosine similarity using Sentence-BERT trained on CheckThat ’21. Manual annotations of the correct pairs.

2.5 Feasibility Evaluation

To evaluate the feasibility of the obtained labels, we performed manual annotation, aiming to estimate the number of correct pairs (i.e., tweet–article pairs, where the article fact-checks the claim in the tweet). Our prior observations of the data suggested that unbiased sampling from the pool of tweets was not suitable, as it would include mostly pairs that have very few overlapping words, which is often an indicator that the texts are not related. Thus, we sample the candidates for annotation based on their Jaccard similarity.

We divided the range of possible values [0;1] into 10 equally sized bins and we sampled 15 examples from each bin, resulting into 150 conversation–reply–tweet triples. Afterwards, the appropriateness of each reply-article and conversation-article pair is annotated by three annotators independently. The annotators had a good level of inter-annotator agreement: 0.75 in terms of Fleiss Kappa Fleiss (1971) (see Appendix C).

Tables 3 and 4 show the resulting estimates of correct pairs for both Jaccard and cosine-based labeling. In the case of Jaccard, we can see that the expected number of correct examples is very high (over 90%) in the range of [0.4–1.0], and then it drastically decreases, going to almost zero when the similarity is less than 0.1. Similarly, for the cosine score, we can see high number of matches in the top 4 bins ([0.6–1.0]), albeit the number of matches remains relatively high in the following interval of [0.2–0.6) between 36% and 63%, and again gets close to zero for the lower-score bins. We analyze the distribution of the Jaccard scores in CheckThat ’21 in more detail in Appendix B.2.

3 Method

General Scheme

As a base for our models, we use Sentence-BERT (SBERT). It uses a Siamese network trained with a Transformer Vaswani et al. (2017) encoder to obtain sentence-level embeddings. We keep the base architecture proposed by Reimers and Gurevych (2019), but we use additional features, training tricks, and losses described in the next sections.

Our input is a pair of a tweet and a fact-checking article, which we encode as follows:

  • Tweet: [CLS] Tweet Text [SEP]

  • Verifying article: [CLS] Title [SEP] Subtitle [SEP] Verified Claim [SEP]

We train the models using the Multiple Negatives Ranking (MNR) loss Henderson et al. (2017) (see Eq. 1), instead of the standard cross-entropy (CE) loss, as the datasets contain only positive (i.e., matching) pairs. Moreover, we propose a new variant of the MNR loss that accounts for the noise in the dataset, as described in detail in Section 3.1.

Enriched Scheme

In the enriched scheme of the model, we adopt the pipeline proposed in the best-performing system from the CheckThat ’21 competition Chernyavskiy et al. (2021). Their method consists of independent components for assessing lexical (TF.IDF-based) and semantic (SBERT-based) similarities. The SBERT models use the same architecture and input format as described in the General Scheme above. However, Chernyavskiy et al. (2021) use an ensemble of models, i.e., instead of calculating a single similarity between the tweet and the joint title/subtitle/verified claim, the similarities between the tweet and the claim, the joint title/claim, and the three together are obtained from three models, one using TF.IDF and one using SBERT, for each combination. These similarities are combined via a re-ranking model (see Section 3.2). In our experiments, the TF.IDF and the model ensembles are included only in the models with re-ranking.

Shuffling and Temperature

Additionally, we adopt a temperature parameter (τ\tau) in the MNR loss. We also make it trainable in order to stabilize the training process as suggested in (Chernyavskiy et al., 2022). This forces the loss to focus on the most complex and most important examples in the batch. Moreover, this effect is amplified after each epoch by an additional data shuffling that composes batches from several groups of the most similar examples. This shuffling, in turn, increases the temperature significance. The nearest neighbors forming the groups are found using the model predictions. More detail about the training and the models themselves can be found in (Chernyavskiy et al., 2021).

3.1 Training with Noisy Data

Self-Adaptive Training

To account for possible noise in the distantly supervised data, we propose a new method based on self-adaptive training Huang et al. (2020), which was introduced for classification tasks and the CE loss; however, it needs to be modified in order be used with the MNR loss. We iteratively refurbish the labels yy using the predictions of the current model starting after an epoch of choice, which is a hyper-parameter:

yrαyr+(1α)y^y^{r}\leftarrow\alpha\cdot y^{r}+(1-\alpha)\cdot\hat{y},

where yry^{r} is the current refurbished label (yr=yy_{r}=y initially), y^\hat{y} is the model prediction, and α\alpha is a momentum hyper-parameter (we set α\alpha to 0.9).

Since the MNR loss operates with positive pairs only (it does not operate with labels), to implement this approach, we had to modify the loss function. Let {ci,vi}i=1,,m\{c_{i},v_{i}\}_{i=1,\ldots,m} be the batch of input pairs, where mm is the batch size, C,Vm×hC,V\in\mathbb{R}^{m\times{h}} are the matrices of embeddings for the tweets and for the fact-checking articles (hh is the embeddings’ hidden size), and C,VC,V are normalized to the unit hyper-sphere (we use cosine similarity), then:

=1mi=1myri(ciTviτlogj=1mexp(ciTvjτ))\begin{split}\mathcal{L}&=-\frac{1}{m}\sum_{i=1}^{m}{y^{r}}_{i}\Big{(}\frac{c_{i}^{T}v_{i}}{\tau}-\log\sum_{j=1}^{m}\exp(\frac{c_{i}^{T}v_{j}}{\tau})\Big{)}\end{split} (1)

If we set yir=1y^{r}_{i}=1, then Eq. 1 resembles the MNR loss definition. The parameter τ\tau is the temperature, discussed in Section 3 Shuffling and Temperature.

Weighting

In the self-adaptive training approach, Huang et al. (2020) introduce weights wi=maxj{1,..,L}ti,jw_{i}=\max_{j\in\{1,..,L\}}{t_{i,j}}, where tit_{i} is the corrected one-hot encoded target vector in a classification task with LL classes. The goal is to ensure that noisy labels will have a lower influence on the training process compared to correct labels. Instead of a classification task with one-hot target vectors ti,jt_{i,j}, here we have real targets yiry^{r}_{i}. Therefore, we take these probabilities as weights: wi=yirw_{i}=y^{r}_{i}. After applying both modifications with the addition of labels and weights, the impact of each training example is proportional to the square of the corrected label, i.e., in Eq. 1 yiry^{r}_{i} is now squared.

3.2 Re-ranking

Re-ranking has shown major improvements for detecting previously fact-checked claims Shaar et al. (2020, 2021); Mihaylova et al. (2021); Chernyavskiy et al. (2021), and we include it as part of our model.

In particular, we adopt the re-ranking procedure from Chernyavskiy et al. (2021), which uses LambdaMART Wu et al. (2010) for re-ranking. The inputs are the reciprocal ranks (position in the ranked list of claims) and the predicted relevance scores (two factors) based on the scores of the TF.IDF and S-BERT models (two models), between the tweet and the claim, claim+title, and claim+title+subtitle (three combinations), for a total of twelve features in the ensemble and four in the single model.

4 Experiments

In this section, we describe our experimental setup, baselines, and experimental results. The training procedure and the hyper-parameters are described in more detail in Appendix A.

4.1 Experimental Setup

Datasets

Table 5 shows statistics about the data split sizes for CrowdChecked and CheckThat ’21. We use these splits in our experiments, albeit sometimes mixed together.

The first group (CrowdChecked) is the data splits obtained using distant supervision. As the positive pairs are annotated with distant supervision and not by humans, we include them as part of the training set. Each shown split is obtained using a different similarity measure (Jaccard or Cosine) or threshold. From the total number of 332K collected tweet–article pairs in CrowdChecked, we ended up with subsets of sizes between 3.5K and 49K examples.

The second group describes the CheckThat ’21 dataset. We preserve the original training, development, and testing splits. In each of our experiments, we validate and test on the corresponding subsets from the CheckThat ’21, while the training set can be a mix with CrowdChecked.

Evaluation Measures

We adopt the ranking measures used in the CheckThat ’21 competition. In particular, we calculate the Mean Reciprocal Rank (MRR), Mean Average Precision (MAP@K), and Precision@K for K{1,3,5,10}K\in\{1,3,5,10\}. We optimize our models for MAP@5, as was in the CLEF-2021 CheckThat! lab subtask 2A.

Dataset Data Split Threshold Tweet-Article Pairs
CrowdChecked (Our Dataset) Train - 332,660
Train Jaccard 0.30 27,387
0.40 12,555
0.50 4,953
Train Cosine 0.50 48,845
0.60 26,588
0.70 11,734
0.80 3,496
CheckThat ’21 Train - 999
Dev - 199
Test - 202
Table 5: Statistics about our collected datasets in terms of tweet–verifying article pairs.

4.2 Baselines and State-of-the-Art

Retrieval

Following Shaar et al. (2021), we use an information retrieval model based on BM25 (Robertson and Zaragoza, 2009) that ranks the fact-checking articles based on the relevance score between their {’claim’, ’title’} and the tweet.

Sentence-BERT

is a bi-encoder model based on Sentence-BERT fine-tuned for detecting previously fact-checked claims using MNR loss. The details are in Section 3, General Scheme.

Team DIPS

Mihaylova et al. (2021) adopts a Sentence-BERT model that computes the cosine similarity for each pair of an input tweet and a verified claim (article). The final ranking is made by passing a sorted list of cosine similarities to a fully-connected neural network.

Team NLytics

Pritzkau (2021) uses a RoBERTa-based model optimized as a regression function obtaining a direct ranking for each tweet-article pair.

Team Aschern

Chernyavskiy et al. (2021) combines TF.IDF with a Sentence-BERT (ensemble with three models of each type). The final ranking is obtained from a re-ranking LambdaMART model.

4.3 Experimental Results

Below, we present experiments that (i) aim to analyze the impact of training with the distantly supervised data from CrowdChecked, and (ii) to further improve the state-of-the-art (SOTA) results using modeling techniques to better leverage the noisy examples (see Section 3). In all our experiments, we evaluate the model on the development and on the testing sets from CheckThat ’21 (see Table 5), and we train on a mix with CrowdChecked. The reported results for each experiment (for each metric) are averaged over three runs using different seeds.

Model MRR P@1 MAP@5
Baselines (CheckThat ’21)
Retrieval (Shaar et al., 2021) 76.1 70.3 74.9
SBERT (CheckThat ’21) 79.96 74.59 79.20
CrowdChecked (Our Dataset)
SBERT (jac > 0.30) 81.50 76.40 80.84
SBERT (cos > 0.50) 81.58 75.91 81.05
(Pre-train) CrowdChecked, (Fine-tune) CheckThat ’21
SBERT (jac > 0.30, Seq) 83.76 78.88 83.11
SBERT (cos > 0.50, Seq) 82.26 77.06 81.41
(Mix) CrowdChecked and CheckThat ’21
SBERT (jac > 0.30, Mix) 83.04 78.55 82.30
SBERT (cos > 0.50, Mix) 82.12 76.57 81.38
Table 6: Evaluation on the CheckThat ’21 test set. In parenthesis is the name of the training split, i.e., Jaccard or Cosine selection strategy, (Seq) first training on CrowdChecked and then on CheckThat ’21, (Mix) mixing the data from the two. The best results are in bold.

Threshold Selection Analysis

Our goal here is to evaluate the impact of using distantly supervised data from CrowdChecked. In particular, we fine-tune an SBERT baseline, as described in Section 3, using four different strategies: (i) fine-tune on the training data from CheckThat ’21, (ii) fine-tune on CrowdChecked, (iii) pre-train on CrowdChecked and then fine-tune on the training data from CheckThat ’21, (iv) mixing the data from both datasets.

Table 6 shows the results grouped based on training data used. In each group, we include the two best-performing models. We see that all SBERT models outperform the Retrieval baseline by 4–8 MAP@5 points absolute. Interestingly, training only on distantly supervised data is enough to outperform the SBERT model trained on the CheckThat ’21 by more than 1.5 MAP@5 points absolute. Moreover, the performance of both data labeling strategies (i.e., Jaccard and Cosine) is close, suggesting comparable amount of noise in them.

Next, we train on combined data from the two datasets. Unsurprisingly, both mixing the data and training on the two datasets sequentially (CrowdChecked \xrightarrow{} CheckThat ’21) yields additional improvement compared to training on a single dataset. We achieve the best result when the model is first pre-trained on the (jac > 0.3) subset of CrowdChecked, and then fine-tuned on CheckThat ’21: it improves by two points absolute in all measures compared to SBERT (CrowdChecked), and by four points compared to SBERT (CheckThat ’21).

Nevertheless, we must note that pre-training with the Cosine similarly (cos > 0.50) did not yield such sizable improvements as the ones when using Jaccard. We attribute this, on one hand, to the higher expected noise in the data according to our manual annotations (see Section 2.5), and on the other hand, to these examples being annotated by a similar model, and thus presumably easy for it.

Refer to caption
Figure 3: MAP@5 for different thresholds and distant supervision approaches. The Jaccard and the Cosine models are trained only on CrowdChecked, while (Seq) and (Mix) were trained also on CheckThat ’21.

We further analyze the impact of choosing different thresholds for the distant supervision approaches. Figure 3 shows the change of MAP@5 for each data labeling strategy. On the left, in the interval [0.3–0.5], are shown the results of the Jaccard-based data labeling strategy, and on the right ([0.5–0.8]) are for the Cosine strategy. Once again, the models trained on the data selected using Jaccard similarity perform similarly or better than the SBERT (CheckThat ’21) model (blue solid line). On the other hand, the Cosine-based selection outperforms the baseline only in small thresholds \leq 0.6. These observations are in favor of the hypothesis that the highly ranked pairs from the fine-tuned SBERT model are easy examples, and do not bring much signal to the model over the CheckThat ’21 data, whereas the Jaccard ranked ones significantly improve the model’s performance. We further see similar performance when training with data from the lowest two thresholds for the two similarities (without data mixing), which suggests that these subsets have similar characteristics.

Adding more distantly supervised data is beneficial for the model, regardless of the strategy. The only exception is the drop in performance when we decrease the Jaccard threshold from 0.5 to 0.4.

We attribute this to the quality of the data in that bracket, as the examples with lower similarity are expected to add more noise. However, the results improve drastically at the next threshold (which also doubles the number of examples), i.e., the model can generalize better from the new data. There is no such drop in the Cosine strategy. We explain this with expectation that noise increases proportionally to the decrease in model confidence.

Finally, we report the performance of each model both on the development and on the test sets in Appendix 11, Tables 11 and 12.

Model MAP@5
Dev Test
DIPS (Mihaylova et al., 2021) 93.6 78.7
NLytics (Pritzkau, 2021) - 79.9
Aschern (Chernyavskiy et al., 2021) 94.2 88.2
SBERT (jac > 0.30, Mix) 90.0 82.3
+ shuffling & trainable temp. 92.4 82.6
+ self-adaptive training (Eq. 1) 92.6 83.6
      + loss weights 92.7 84.3
       + TF.IDF + Re-ranking 93.1 89.7
         + TF.IDF + Re-ranking (ens.) 94.8 90.3
Table 7: Results on CheckThat ’21 (dev and test). We compare our model and its components (added sequentially) to the state of the art. The best results are in bold.

Modeling Noisy Data

We explore the impact of the proposed changes to the SBERT training approach: (i) shuffling and training temperature, (ii) data-related modification of the MNR loss for self-adaptive training with weights. We use the (jac > 0.30, mix) approach in our experiments, as the baseline SBERT models achieved the highest scores on the dev set (Table 11). In Table 7, we ablate each of these modifications by adding them iteratively to the baseline SBERT model.

First, we can see that adding a special shuffling procedure and a trainable temperature (τ\tau) improves the MAP@5 by 2 points on the dev set and by 0.3 points on the test set. Next, we see a sizable improvement of 1 MAP@5 point on the test set, when using the self-adaptive training with MNR loss. Moreover, an additional 0.7 points come from adding weights to the loss, arriving at MAP@5 of 84.3. These weights allow the model to give higher importance to the less noisy data during training.

Note that for these two ablations the improvements on the development set are diminishing. We attribute this to its small size (199 examples) and to the high values of MAP@5. Finally, note that our model without re-ranking outperforms almost all state-of-the-art models (except for that of team Aschern) by more than 4.5 points on the test dataset.

The last two rows of Table 7 show the results of our model that includes all proposed components, in combination with TF.IDF features and the LambdaMART re-ranking, described in Section 3. Here, we must note that our model is trained on part of the CheckThat ’21 training pool (80%) – the other part is used to train the re-ranking model. The full setup boosts the model’s MAP@5 to 89.7 when using a single model of the TF.IDF and SBERT (using the title/subtitle/claim as inputs, same as SBERT). With the ensemble architecture (re-ranking based on the scores of three TF.IDF and three SBERT models), we achieve our best results of 90.3 on the test set (adding 1.7 MAP@5 on dev, and 0.6 on test), outperforming the previous state-of-the-art approach (Aschern, 88.2) by 2 MAP@5 points, and by more than 11 compared to the second best model (NLytics, 79.9). This improvement corresponds to the observed gain over the SBERT model without re-ranking. Nevertheless, the change in the strength of the factors in LambdaMART is less. The TF-IDF models still have high importance for re-ranking – a total of 41% compared to 42.8% reported in Chernyavskiy et al. (2021). Here, we have a decrease mainly due to an increase of the importance of the reciprocal rank factor from 18.8% to 20.2% of the SBERT model that selects candidates.

5 Discussion

Our proposed distant supervision data selection strategies show promising results, achieving SOTA results on the CheckThat ’21. Nonetheless, we are not able to identify all matching pairs in the list of candidates in CrowdChecked. Hereby, we try to estimate their number using statistics from our manual annotations,222Due to the small number of annotated examples the variance in the estimates is large. as shown in Tables 3 and 4.

In particular, we estimate it by multiplying the fraction of correct pairs in each similarity bin by the number of examples in this bin. Based on cosine similarity, we estimate that out of the 332,600 pairs, the matching pairs are approximately 90,170 (27.11%).

Based on the Jaccard distribution, we estimate that 14.79% of all tweet-conversations (root of the conversation), and 22.23% of the tweet–reply (the tweet before the current one in the conversation) pairs are good, or nearly 61,500 examples.

Our experiments show that the models can effectively account for the noise in the training data. The self-adaptive training and the additional weighing in the loss (described in Section 3) yield 1 additional MAP@5 point each. This suggests that learning from noisy labels Han et al. (2018); Wang et al. (2019); Song et al. (2022); Zhou and Chen (2021) and using all examples in CrowdChecked can improve the results even further. Moreover, incorporating the negative examples (non-matching pairs) from CrowdChecked in the training could also help Lu et al. (2021); Thakur et al. (2021).

6 Related Work

Previously Fact-Checked Claims

While fake news and mis/disinformation detection have been studied extensively Li et al. (2016); Zubiaga et al. (2018); Martino et al. (2020); Alam et al. (2022); Guo et al. (2022); Hardalov et al. (2022), the problem of detecting previously fact-checked claims remains under-explored. Hassan et al. (2017) mentioned the task as a component of an end-to-end fact-checking pipeline, but did not evaluate it nor studied its contribution. Hossain et al. (2020) retrieved evidence from a list of known misconceptions and evaluated the claim’s veracity based on its stance towards the hits; while this task is similar, it is not about whether a given claim was fact-checked or not.

Recently, the task received more attention. Shaar et al. (2020) collected two datasets, from PolitiFact (political debates) and Snopes (tweets), of claims and corresponding fact-checking articles. The CLEF CheckThat! lab (Barrón-Cedeño et al., 2020a, b, c; Nakov et al., 2021b, c; Shaar et al., 2021; Nakov et al., 2022a, b, c) extended these datasets with more data in English and Arabic. The best systems (Pritzkau, 2021; Mihaylova et al., 2021; Chernyavskiy et al., 2022) used a combination of BM25 retrieval, semantic similarity using embeddings Reimers and Gurevych (2019), and reranking. Bouziane et al. (2020) used extra data from fact-checking datasets Wang (2017); Thorne et al. (2018); Wadden et al. (2020).

Finally, Shaar et al. (2022a) and Shaar et al. (2022b) explored the role of the context in detecting previously fact-checked claims in political debates.

Our work is most similar to that of Vo and Lee (2020), who mined 19K tweets and corresponding fact-checked articles. Unlike them, we focus on textual claims (they were interested in multimodal tweets with images), we collect an order of magnitude more examples, and we propose a novel approach to learn from such noisy data directly (while they manually checked each example).

Training with Noisy Data

Leveraging large collections of unlabeled data has been at the core of large-scale language models using Transformers (Vaswani et al., 2017), such as GPT Radford et al. (2018, 2019), BERT Devlin et al. (2019), and RoBERTa Liu et al. (2019). Recently, such models used noisy retrieved data Lewis et al. (2020); Guu et al. (2020) or active relabeling and data augmentation Thakur et al. (2021). Distant supervision is also a crucial part of recent breakthroughs in few-shot learning Schick and Schütze (2021a, b).

Yet, there has been little work of using noisy data for fact-checking tasks. Vo and Lee (2019) collected tweets containing a link to a fact-checking website, based on which they tried to learn a fact-checking language and to generate automatic answers. You et al. (2019) used similar data from tweets for fact-checking URL recommendations.

Unlike the above work, here we propose an automatic procedure for labeling and self-training specifically designed for the task of detecting previously fact-checked claims.

7 Conclusion and Future Work

We presented CrowdChecked, a large dataset for detecting previously fact-checked claims, with more than 330,000 pairs of tweets and corresponding fact-checking articles posted by crowd fact-checkers. We further investigated two techniques for labeling the data using distance supervision, resulting in training sets of 3.5K–50K examples. We also proposed an approach for training from noisy data using self-adaptive learning and additional weights in the loss function. Furthermore, we demonstrated that our data yields sizable performance gains of four points in terms MRR, P@1, and MAP@5 over strong baselines. Finally, we demonstrated improvements over the state of the art on the CheckThat ’21 test set by two points, when using our proposed dataset and pipeline.

In future work, we plan to experiment with more languages and more distant supervision techniques such as predictions from an ensemble model.

Acknowledgments

We want to thank Ivan Bozhilov, Martin Vrachev, and Lilyana Videva for the useful discussions and for their help with additional data analysis and manual annotations.

This work is partially supported by Project UNITe BG05M2OP001-1.001-0004 funded by the Bulgarian OP “Science and Education for Smart Growth”, co-funded by the EU via the ESI Funds.

The work was prepared within the framework of the HSE University Basic Research Program.

Ethics and Broader Impact

Dataset Collection

We collected the dataset using the Twitter API.333We use the Twitter API v2 with academic research access, http://developer.twitter.com/en/docs, following the terms of use outlined by Twitter.444http://developer.twitter.com/en/developer-terms/agreement-and-policy Specifically, we only downloaded public tweets, and we only distribute dehydrated Twitter IDs.

Biases

We note that some of the annotations are subjective, and we have clearly indicated in the text which these are. Thus, it is inevitable that there would be biases in our dataset. Yet, we have a very clear annotation schema and instructions, which should reduce the biases.

Misuse Potential

Most datasets compiled from social media present some risk of misuse. We, therefore, ask researchers to be aware that our dataset can be maliciously used to unfairly moderate text (e.g., a tweet) that may not be malicious based on biases that may or may not be related to demographics and other information within the text. Intervention with human moderation would be required in order to ensure this does not occur.

Intended Use

Our dataset can enable automatic systems for analysis of social media content, which could be of interest to practitioners, professional fact-checker, journalists, social media platforms, and policymakers. Such systems can be used to alleviate the burden of moderators, but human supervision would be required for more intricate cases and in order to ensure that no harm is caused.

Our models can help fight the COVID-19 infodemic, and they could support analysis and decision-making for the public good. However, the models can also be misused by malicious actors. Therefore, we ask the users to be aware of potential misuse. With the possible ramifications of a highly subjective dataset, we distribute it for research purposes only, without a license for commercial use. Any biases found in the dataset are unintentional, and we do not intend to do harm to any group or individual.

Environmental Impact

We would like to warn that the use of large-scale Transformers requires a lot of computations and the use of GPUs/TPUs for training, which contributes to global warming Strubell et al. (2019). This is a bit less of an issue in our case, as we do not train such models from scratch; rather, we fine-tune them on relatively small datasets. Moreover, running on a CPU for inference, once the model has been fine-tuned, is perfectly feasible, and CPUs contribute much less to global warming.

References

  • Alam et al. (2022) Firoj Alam, Stefano Cresci, Tanmoy Chakraborty, Fabrizio Silvestri, Dimiter Dimitrov, Giovanni Da San Martino, Shaden Shaar, Hamed Firooz, and Preslav Nakov. 2022. A survey on multimodal disinformation detection. In Proceedings of the 29th International Conference on Computational Linguistics, COLING ’22, Gyeongju, Republic of Korea.
  • Augenstein et al. (2019) Isabelle Augenstein, Christina Lioma, Dongsheng Wang, Lucas Chaves Lima, Casper Hansen, Christian Hansen, and Jakob Grue Simonsen. 2019. MultiFC: A real-world multi-domain dataset for evidence-based fact checking of claims. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP ’19, pages 4685–4697, Hong Kong, China.
  • Baly et al. (2020) Ramy Baly, Georgi Karadzhov, Jisun An, Haewoon Kwak, Yoan Dinkov, Ahmed Ali, James Glass, and Preslav Nakov. 2020. What was written vs. who read it: News media profiling using text analysis and social media context. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL ’20, pages 3364–3374, Online.
  • Barrón-Cedeño et al. (2020a) Alberto Barrón-Cedeño, Tamer Elsayed, Preslav Nakov, Giovanni Da San Martino, Maram Hasanain, Reem Suwaileh, Fatima Haouari, Nikolay Babulkov, Bayan Hamdan, Alex Nikolov, Shaden Shaar, and Zien Sheikh Ali. 2020a. Overview of CheckThat! 2020 — automatic identification and verification of claims in social media. In Proceedings of the 11th International Conference of the CLEF Association: Experimental IR Meets Multilinguality, Multimodality, and Interaction, CLEF ’2020, pages 215–236, Thessaloniki, Greece.
  • Barrón-Cedeño et al. (2020b) Alberto Barrón-Cedeño, Tamer Elsayed, Preslav Nakov, Giovanni Da San Martino, Maram Hasanain, Reem Suwaileh, Fatima Haouari, Nikolay Babulkov, Bayan Hamdan, Alex Nikolov, et al. 2020b. Overview of CheckThat! 2020: Automatic identification and verification of claims in social media. In International Conference of the Cross-Language Evaluation Forum for European Languages, CLEF ’20, pages 215–236, Thessaloniki, Greece. Springer.
  • Barrón-Cedeño et al. (2020c) Alberto Barrón-Cedeño, Tamer Elsayed, Preslav Nakov, Giovanni Da San Martino, Maram Hasanain, Reem Suwaileh, and Fatima Haouari. 2020c. CheckThat! at CLEF 2020: Enabling the automatic identification and verification of claims in social media. In Proceedings of the 42nd European Conference on Information Retrieval, ECIR ’20, pages 499–507, Lisbon, Portugal.
  • Bouziane et al. (2020) Mostafa Bouziane, Hugo Perrin, Aurélien Cluzeau, Julien Mardas, and Amine Sadeq. 2020. Team Buster.ai at CheckThat! 2020: Insights and recommendations to improve fact-checking. In CLEF (Working Notes), CLEF ’20, Thessaloniki, Greece.
  • Chernyavskiy et al. (2022) Anton Chernyavskiy, Dmitry Ilvovsky, Pavel Kalinin, and Preslav Nakov. 2022. Batch-softmax contrastive loss for pairwise sentence scoring tasks. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT ’22, pages 116–126, Seattle, United States.
  • Chernyavskiy et al. (2021) Anton Chernyavskiy, Dmitry Ilvovsky, and Preslav Nakov. 2021. Aschern at CLEF CheckThat! 2021: Lambda-Calculus of Fact-Checked Claims. In Proceedings of the Working Notes of CLEF 2021 - Conference and Labs of the Evaluation Forum, volume 2936 of CEUR Workshop Proceedings, pages 484–493, Bucharest, Romania.
  • Derczynski et al. (2017) Leon Derczynski, Kalina Bontcheva, Maria Liakata, Rob Procter, Geraldine Wong Sak Hoi, and Arkaitz Zubiaga. 2017. SemEval-2017 task 8: RumourEval: Determining rumour veracity and support for rumours. In Proceedings of the 11th International Workshop on Semantic Evaluation, SemEval ’17, pages 69–76, Vancouver, Canada.
  • Devlin et al. (2019) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT ’19, pages 4171–4186, Minneapolis, Minnesota.
  • Elsayed et al. (2019) Tamer Elsayed, Preslav Nakov, Alberto Barrón-Cedeno, Maram Hasanain, Reem Suwaileh, Giovanni Da San Martino, and Pepa Atanasova. 2019. Overview of the CLEF-2019 CheckThat! lab: Automatic identification and verification of claims. In International Conference of the Cross-Language Evaluation Forum for European Languages, CLEF ’20, pages 301–321, Virtual. Springer.
  • Ferreira and Vlachos (2016) William Ferreira and Andreas Vlachos. 2016. Emergent: A novel data-set for stance classification. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT ’16, pages 1163–1168, San Diego, California, USA.
  • Fleiss (1971) Joseph L Fleiss. 1971. Measuring nominal scale agreement among many raters. Psychological bulletin, 76(5):378.
  • Gu et al. (2022) Zihui Gu, Ju Fan, Nan Tang, Preslav Nakov, Xiaoman Zhao, and Xiaoyong Du. 2022. PASTA: Table-operations aware fact verification via sentence-table cloze pre-training. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP ’22, Abu Dhabi, UAE.
  • Guo et al. (2022) Zhijiang Guo, Michael Schlichtkrull, and Andreas Vlachos. 2022. A survey on automated fact-checking. Transactions of the Association for Computational Linguistics (TACL), 10:178–206.
  • Guu et al. (2020) Kelvin Guu, Kenton Lee, Zora Tung, Panupong Pasupat, and Ming-Wei Chang. 2020. Retrieval augmented language model pre-training. In Proceedings of the 37th International Conference on Machine Learning, volume 119 of ICML ’20, pages 3929–3938, Virtual.
  • Han et al. (2018) Bo Han, Quanming Yao, Xingrui Yu, Gang Niu, Miao Xu, Weihua Hu, Ivor W. Tsang, and Masashi Sugiyama. 2018. Co-teaching: Robust training of deep neural networks with extremely noisy labels. In Advances in Neural Information Processing Systems 31, NeurIPS ’18, pages 8536–8546, Montréal, Canada.
  • Hanselowski et al. (2019) Andreas Hanselowski, Christian Stab, Claudia Schulz, Zile Li, and Iryna Gurevych. 2019. A richly annotated corpus for different tasks in automated fact-checking. In Proceedings of the 23rd Conference on Computational Natural Language Learning, CoNLL ’19, pages 493–503, Hong Kong, China.
  • Hardalov et al. (2022) Momchil Hardalov, Arnav Arora, Preslav Nakov, and Isabelle Augenstein. 2022. A survey on stance detection for mis- and disinformation identification. In Findings of the Association for Computational Linguistics: NAACL 2022, Findings of NAACL ’22, pages 1259–1277, Seattle, Washington, USA.
  • Hassan et al. (2017) Naeemul Hassan, Gensheng Zhang, Fatma Arslan, Josue Caraballo, Damian Jimenez, Siddhant Gawsane, Shohedul Hasan, Minumol Joseph, Aaditya Kulkarni, Anil Kumar Nayak, Vikas Sable, Chengkai Li, and Mark Tremayne. 2017. ClaimBuster: The first-ever end-to-end fact-checking system. Proceedings of the International Conference on Very Large Data Bases, 10(12):1945–1948.
  • Henderson et al. (2017) Matthew Henderson, Rami Al-Rfou, B. Strope, Yun-Hsuan Sung, L. Lukács, R. Guo, S. Kumar, B. Miklos, and R. Kurzweil. 2017. Efficient natural language response suggestion for smart reply. ArXiv 1705.00652.
  • Hossain et al. (2020) Tamanna Hossain, Robert L. Logan IV, Arjuna Ugarte, Yoshitomo Matsubara, Sean Young, and Sameer Singh. 2020. COVIDLies: Detecting COVID-19 misinformation on social media. In Proceedings of the 1st Workshop on NLP for COVID-19 (Part 2) at EMNLP 2020, NLP-COVID19 ’20, Online.
  • Huang et al. (2020) Lang Huang, Chao Zhang, and Hongyang Zhang. 2020. Self-adaptive training: Beyond empirical risk minimization. In Advances in Neural Information Processing Systems 33, NeurIPS ’20, Virtual.
  • Lewis et al. (2020) Patrick S. H. Lewis, Ethan Perez, Aleksandra Piktus, Fabio Petroni, Vladimir Karpukhin, Naman Goyal, Heinrich Küttler, Mike Lewis, Wen-tau Yih, Tim Rocktäschel, Sebastian Riedel, and Douwe Kiela. 2020. Retrieval-augmented generation for knowledge-intensive NLP tasks. In Advances in Neural Information Processing Systems 33, NeurIPS ’20, Virtual.
  • Li et al. (2016) Yaliang Li, Jing Gao, Chuishi Meng, Qi Li, Lu Su, Bo Zhao, Wei Fan, and Jiawei Han. 2016. A survey on truth discovery. SIGKDD Explor. Newsl., 17(2):1–16.
  • Liu et al. (2019) Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. RoBERTa: A robustly optimized BERT pretraining approach. arXiv:1907.11692.
  • Loper and Bird (2002) Edward Loper and Steven Bird. 2002. NLTK: The natural language toolkit. In Proceedings of the ACL-02 Workshop on Effective Tools and Methodologies for Teaching Natural Language Processing and Computational Linguistics, pages 63–70, Philadelphia, Pennsylvania, USA.
  • Loshchilov and Hutter (2019) Ilya Loshchilov and Frank Hutter. 2019. Decoupled weight decay regularization. In Proceedings of the 7th International Conference on Learning Representations, ICLR ’19, New Orleans, Louisiana, USA.
  • Lu et al. (2021) Jing Lu, Gustavo Hernandez Abrego, Ji Ma, Jianmo Ni, and Yinfei Yang. 2021. Multi-stage training with improved negative contrast for neural passage retrieval. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP ’21, pages 6091–6103, Online and Punta Cana, Dominican Republic.
  • Martino et al. (2020) Giovanni Da San Martino, Stefano Cresci, Alberto Barrón-Cedeño, Seunghak Yu, Roberto Di Pietro, and Preslav Nakov. 2020. A survey on computational propaganda detection. In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI ’20, pages 4826–4832.
  • Mihaylova et al. (2021) Simona Mihaylova, Iva Borisova, Dzhovani Chemishanov, Preslav Hadzhitsanev, Momchil Hardalov, and Preslav Nakov. 2021. DIPS at CheckThat! 2021: Verified claim retrieval. In Proceedings of the Working Notes of CLEF 2021 - Conference and Labs of the Evaluation Forum, volume 2936 of CEUR Workshop Proceedings, pages 558–571, Bucharest, Romania.
  • Nakov et al. (2022a) Preslav Nakov, Alberto Barrón-Cedeño, Giovanni Da San Martino, Firoj Alam, Julia Maria Struß, Thomas Mandl, Rubén Míguez, Tommaso Caselli, Mucahid Kutlu, Wajdi Zaghouani, Chengkai Li, Shaden Shaar, Gautam Kishore Shahi, Hamdy Mubarak, Alex Nikolov, Nikolay Babulkov, Yavuz Selim Kartal, and Javier Beltrán. 2022a. The CLEF-2022 CheckThat! lab on fighting the COVID-19 infodemic and fake news detection. In Proceedings of the 44th European Conference on IR Research: Advances in Information Retrieval, ECIR ’22, pages 416–428, Stavanger, Norway.
  • Nakov et al. (2022b) Preslav Nakov, Alberto Barrón-Cedeño, Giovanni Da San Martino, Firoj Alam, Julia Maria Struß, Thomas Mandl, Rubén Míguez, Tommaso Caselli, Mucahid Kutlu, Wajdi Zaghouani, Chengkai Li, Shaden Shaar, Gautam Kishore Shahi, Hamdy Mubarak, Alex Nikolov, Nikolay Babulkov, Yavuz Selim Kartal, and Javier Beltrán. 2022b. Overview of the CLEF-2022 CheckThat! lab on fighting the COVID-19 infodemic and fake news detection. In Proceedings of the 13th International Conference of the CLEF Association: Information Access Evaluation meets Multilinguality, Multimodality, and Visualization, CLEF ’2022, Bologna, Italy.
  • Nakov et al. (2021a) Preslav Nakov, David Corney, Maram Hasanain, Firoj Alam, Tamer Elsayed, Alberto Barrón-Cedeño, Paolo Papotti, Shaden Shaar, and Giovanni Da San Martino. 2021a. Automated fact-checking for assisting human fact-checkers. In Proceedings of the 30th International Joint Conference on Artificial Intelligence, IJCAI ’21, pages 4551–4558.
  • Nakov et al. (2022c) Preslav Nakov, Giovanni Da San Martino, Firoj Alam, Shaden Shaar, Hamdy Mubarak, and Nikolay Babulkov. 2022c. Overview of the CLEF-2022 CheckThat! lab task 2 on detecting previously fact-checked claims. In Working Notes of CLEF 2022—Conference and Labs of the Evaluation Forum, CLEF ’2022, Bologna, Italy.
  • Nakov et al. (2021b) Preslav Nakov, Giovanni Da San Martino, Tamer Elsayed, Alberto Barrón-Cedeño, Rubén Míguez, Shaden Shaar, Firoj Alam, Fatima Haouari, Maram Hasanain, Nikolay Babulkov, Alex Nikolov, Gautam Kishore Shahi, Julia Maria Struß, and Thomas Mandl. 2021b. The CLEF-2021 CheckThat! lab on detecting check-worthy claims, previously fact-checked claims, and fake news. In Proceedings of the 43rd European Conference on Information Retrieval, ECIR ’21, pages 639–649, Lucca, Italy.
  • Nakov et al. (2021c) Preslav Nakov, Giovanni Da San Martino, Tamer Elsayed, Alberto Barrón-Cedeño, Rubén Míguez, Shaden Shaar, Firoj Alam, Fatima Haouari, Maram Hasanain, Watheq Mansour, Bayan Hamdan, Zien Sheikh Ali, Nikolay Babulkov, Alex Nikolov, Gautam Kishore Shahi, Julia Maria Struß, Thomas Mandl, Mucahid Kutlu, and Yavuz Selim Kartal. 2021c. Overview of the CLEF-2021 CheckThat! lab on detecting check-worthy claims, previously fact-checked claims, and fake news. In Proceedings of the 12th International Conference of the CLEF Association: Information Access Evaluation Meets Multiliguality, Multimodality, and Visualization, CLEF ’2021, Bucharest, Romania (online).
  • Nguyen et al. (2022) Van-Hoang Nguyen, Kazunari Sugiyama, Preslav Nakov, and Min-Yen Kan. 2022. FANG: Leveraging social context for fake news detection using graph representation. Commun. ACM, 65(4):124–132.
  • Panayotov et al. (2022) Panayot Panayotov, Utsav Shukla, Husrev Taha Sencar, Mohamed Nabeel, and Preslav Nakov. 2022. GREENER: Graph neural networks for news media profiling. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP ’22, Abu Dhabi, UAE.
  • Paszke et al. (2019) Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Köpf, Edward Yang, Zachary DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. 2019. PyTorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems 32, NeurIPS ’19, pages 8024–8035, Vancouver, Canada.
  • Popat et al. (2016) Kashyap Popat, Subhabrata Mukherjee, Jannik Strötgen, and Gerhard Weikum. 2016. Credibility assessment of textual claims on the web. In Proceedings of the 25th ACM International Conference on Information and Knowledge Management, CIKM ’16, pages 2173–2178, Indianapolis, Indiana, USA.
  • Porter (1980) Martin F Porter. 1980. An algorithm for suffix stripping. Program.
  • Pritzkau (2021) Albert Pritzkau. 2021. NLytics at CheckThat! 2021: Multi-class fake news detection of news articles and domain identification with RoBERTa - a baseline model. In Proceedings of the Working Notes of CLEF 2021 - Conference and Labs of the Evaluation Forum, volume 2936 of CEUR Workshop Proceedings, pages 572–581, Bucharest, Romania.
  • Radford et al. (2018) Alec Radford, Karthik Narasimhan, Tim Salimans, and Ilya Sutskever. 2018. Improving language understanding by generative pre-training. OpenAI.
  • Radford et al. (2019) Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. 2019. Language models are unsupervised multitask learners. OpenAI Blog.
  • Reimers and Gurevych (2019) Nils Reimers and Iryna Gurevych. 2019. Sentence-BERT: Sentence embeddings using Siamese BERT-networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP ’22, pages 3982–3992, Hong Kong, China.
  • Robertson and Zaragoza (2009) Stephen Robertson and Hugo Zaragoza. 2009. The probabilistic relevance framework: BM25 and beyond. Found. Trends Inf. Retr., 3(4):333–389.
  • Schick and Schütze (2021a) Timo Schick and Hinrich Schütze. 2021a. Exploiting cloze-questions for few-shot text classification and natural language inference. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics, EACL ’21, pages 255–269, Online.
  • Schick and Schütze (2021b) Timo Schick and Hinrich Schütze. 2021b. It’s not just size that matters: Small language models are also few-shot learners. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT ’21, pages 2339–2352, Online.
  • Shaar et al. (2022a) Shaden Shaar, Firoj Alam, Giovanni Da San Martino, and Preslav Nakov. 2022a. The role of context in detecting previously fact-checked claims. In Findings of the Association for Computational Linguistics: NAACL-HLT 2022, NAACL-HLT ’22, pages 1619–1631, Seattle, Washington, USA.
  • Shaar et al. (2020) Shaden Shaar, Nikolay Babulkov, Giovanni Da San Martino, and Preslav Nakov. 2020. That is a known lie: Detecting previously fact-checked claims. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL ’20, pages 3607–3618, Online.
  • Shaar et al. (2022b) Shaden Shaar, Nikola Georgiev, Firoj Alam, Giovanni Da San Martino, Aisha Mohamed, and Preslav Nakov. 2022b. Assisting the human fact-checkers: Detecting all previously fact-checked claims in a document. In Findings of the Conference on Empirical Methods in Natural Language Processing, EMNLP ’22, Abu Dhabi, UAE.
  • Shaar et al. (2021) Shaden Shaar, Fatima Haouari, Watheq Mansour, Maram Hasanain, Nikolay Babulkov, Firoj Alam, Giovanni Da San Martino, Tamer Elsayed, and Preslav Nakov. 2021. Overview of the CLEF-2021 CheckThat! lab task 2 on detecting previously fact-checked claims in tweets and political debates. In CLEF (Working Notes), pages 393–405.
  • Song et al. (2022) Hwanjun Song, Minseok Kim, Dongmin Park, Yooju Shin, and Jae-Gil Lee. 2022. Learning from noisy labels with deep neural networks: A survey. IEEE Transactions on Neural Networks and Learning Systems, pages 1–19.
  • Stefanov et al. (2020) Peter Stefanov, Kareem Darwish, Atanas Atanasov, and Preslav Nakov. 2020. Predicting the topical stance and political leaning of media using tweets. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL ’20, pages 527–537, Online.
  • Strubell et al. (2019) Emma Strubell, Ananya Ganesh, and Andrew McCallum. 2019. Energy and policy considerations for deep learning in NLP. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, ACL ’19, pages 3645–3650, Florence, Italy.
  • Tchechmedjiev et al. (2019) Andon Tchechmedjiev, Pavlos Fafalios, Katarina Boland, Malo Gasquet, Matthäus Zloch, Benjamin Zapilko, Stefan Dietze, and Konstantin Todorov. 2019. ClaimsKG: A knowledge graph of fact-checked claims. In International Semantic Web Conference, ISWC ’19, pages 309–324. Springer.
  • Thakur et al. (2021) Nandan Thakur, Nils Reimers, Johannes Daxenberger, and Iryna Gurevych. 2021. Augmented SBERT: Data augmentation method for improving bi-encoders for pairwise sentence scoring tasks. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT ’21, pages 296–310, Online.
  • Thorne et al. (2018) James Thorne, Andreas Vlachos, Christos Christodoulopoulos, and Arpit Mittal. 2018. FEVER: A large-scale dataset for fact extraction and VERification. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT ’18, pages 809–819, New Orleans, Louisiana, USA.
  • Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Advances in Neural Information Processing Systems 30, NeurIPS ’17, pages 5998–6008, Long Beach, California, USA.
  • Vo and Lee (2019) Nguyen Vo and Kyumin Lee. 2019. Learning from fact-checkers: Analysis and generation of fact-checking language. In Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR ’19, pages 335–344, Paris, France.
  • Vo and Lee (2020) Nguyen Vo and Kyumin Lee. 2020. Where are the facts? searching for fact-checked information to alleviate the spread of fake news. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP ’20, pages 7717–7731.
  • Wadden et al. (2020) David Wadden, Shanchuan Lin, Kyle Lo, Lucy Lu Wang, Madeleine van Zuylen, Arman Cohan, and Hannaneh Hajishirzi. 2020. Fact or fiction: Verifying scientific claims. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP ’20, pages 7534–7550, Online.
  • Wang et al. (2019) Hao Wang, Bing Liu, Chaozhuo Li, Yan Yang, and Tianrui Li. 2019. Learning with noisy labels for sentence-level sentiment classification. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP ’19, pages 6286–6292, Hong Kong, China.
  • Wang (2017) William Yang Wang. 2017. “Liar, liar pants on fire”: A new benchmark dataset for fake news detection. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, ACL ’17, pages 422–426, Vancouver, Canada.
  • Wolf et al. (2020) Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, Remi Louf, Morgan Funtowicz, Joe Davison, Sam Shleifer, Patrick von Platen, Clara Ma, Yacine Jernite, Julien Plu, Canwen Xu, Teven Le Scao, Sylvain Gugger, Mariama Drame, Quentin Lhoest, and Alexander Rush. 2020. Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, EMNLP (Demonstrations) ’20, pages 38–45, Online.
  • Wu et al. (2010) Qiang Wu, Christopher JC Burges, Krysta M Svore, and Jianfeng Gao. 2010. Adapting boosting for information retrieval measures. Information Retrieval, 13(3):254–270.
  • You et al. (2019) Di You, Nguyen Vo, Kyumin Lee, and Qiang Liu. 2019. Attributed multi-relational attention network for fact-checking URL recommendation. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, CIKM ’19, pages 1471–1480, Beijing, China.
  • Zhou and Chen (2021) Wenxuan Zhou and Muhao Chen. 2021. Learning from noisy labels for entity-centric information extraction. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP ’21, pages 5381–5392, Online and Punta Cana, Dominican Republic.
  • Zubiaga (2018) Arkaitz Zubiaga. 2018. A longitudinal assessment of the persistence of Twitter datasets. Journal of the Association for Information Science and Technology, 69(8):974–984.
  • Zubiaga et al. (2018) Arkaitz Zubiaga, Ahmet Aker, Kalina Bontcheva, Maria Liakata, and Rob Procter. 2018. Detection and resolution of rumours in social media: A survey. ACM Comput. Surv., 51(2).
  • Zubiaga et al. (2016) Arkaitz Zubiaga, Maria Liakata, Rob Procter, Geraldine Wong Sak Hoi, and Peter Tolmie. 2016. Analysing how people orient to and spread rumours in social media by looking at conversational threads. PLOS ONE, 11(3):1–29.

Appendix A Hyperparameters and Fine-Tuning

Below, we first describe the common parameters we use, and then we give the values of model-specific parameters.

Common Parameters

  • We develop our models in Python using PyTorch (Paszke et al., 2019), the Transformers library (Wolf et al., 2020), and the Sentence Transformers library. (Reimers and Gurevych, 2019)555http://github.com/UKPLab/sentence-transformers

  • We used NLTK (Loper and Bird, 2002) to filter out English stop words, the Twitter Tokenizer to split the tweets and to strip the handles, and the Porter stemmer (Porter, 1980) to stem the tokens.

  • For optimization, we use AdamW (Loshchilov and Hutter, 2019) with weight decay of 1e-8, β1\beta_{1} = 0.9, β2\beta_{2} = 0.999, ϵ\epsilon = 1e-08, for 10 epochs, and maximum sequence length of 128 tokens (per encoder).666When needed, we truncated the sequences token by token, starting from the longest sequence in the pair.

  • All Sentence BERT (SBERT) models are initialized from the stsb-bert-base777huggingface.co/sentence-transformers/stsb-bert-base checkpoint.

  • The SBERT models use cosine similarity both during training inside the MNR loss and during inference for ranking.

  • We selectd the values of the hyper-parameters on the development set of CheckThat ’21,888https://gitlab.com/checkthat_lab/clef2021-checkthat-lab/-/tree/master/task2 and we chose the best model checkpoint based on the performance on the development set (MAP@5).

  • We ran each experiment three times with different seeds and averaged the result scores.

  • The models were evaluated on each epoch or every 250 steps, whichever is less.

  • The evaluation measures are calculated using the official code from the CheckThat ’21 competition Shaar et al. (2021)999https://gitlab.com/checkthat_lab/clef2021-checkthat-lab/-/tree/master/task2/scorer and the SentenceTransformer’s library.

  • In our work, we list 199 examples for the development set of CheckThat ’21, while Shaar et al. (2021) lists 200. The difference comes from one duplicate row in the development set, which we found and filtered out.

  • We trained our models on 5x Tesla T4 GPUs and 1x GeForce GTX 1080Ti, depending on the dataset size, the experiments took between 10 minutes and 5 hours.

Baseline SBERT

  • Our baseline Sentence BERT is trained with LR of 2e-05, warmup of 0.1, and batch size of 32.

  • We set the temperature (τ\tau) in the MNR loss to 1.0, i.e., using unmodified MNR.

  • The model consists of 110M params, same as the bert-base Devlin et al. (2019), as it uses a bi-encoder scheme.

Proposed Pipeline

  • The model is trained with LR of 1e-05, warmup of 0.1, batch size of 8, ad group size of 4 during the dataset shuffling.

  • We tuned the settings of the self-adaptive training, and ended up with the folowing values: momentum α\alpha of 0.9, refurbishment process starting at the second epoch.

  • We set the learning rate for the temperature (τ\tau) in the MNR loss to 0.4.

  • In the re-ranking, we used 800 training examples to train SBERT and the remaining 199 examples to train LambdaMART.

  • We re-ranked the top-100 results from the best SBERT model with LambdaMART.

  • All other training details we kept from Chernyavskiy et al. (2021).

  • The model has 330M params, 3x as the size of the Baseline SBERT, as it trains three separate models.

  • In our preliminary experiments, SBERT-base and SBERT-large yielded the same results in terms of MAP@5, ad thus we experiment with the base versions.

Appendix B Dataset

Below, we first give some detail about the process of article collection, and then we discuss the overlap between our CrowdChecked dataset and CheckThat ’21.

B.1 Fact-checking Articles Collection

In order to obtain a collection of fact-checking articles for each tweet, we first formed a list of unique URLs shared in the fact-checking tweets from the crowd fact-checkers. Next, from each URL we downloaded the HTML of the whole page and extracted the meta information using CSS selectors and RegEx rules. In particular, we followed previous work Barrón-Cedeño et al. (2020b); Shaar et al. (2021) and collected: title (the title of the page), subtitle (short description of the fact-check), claim (the claim of interest), subtitle (short description of the fact-check), date (the date the article was published), and author (the author of the article). We do not parse the content of the article and the factual label, as the credibility of the claim is not related to the objective of this task, i.e., the goal is to find a fact-checking article, but not to verify it.

As a result, we collected 10,340 articles that were published in the period between 1995–2021. The per-year distribution is shown in Table 2 (in brown). The majority of the articles are from the period after 2015, with a peak at the ones from 2020/2021. We attribute this on the increased media literacy and on the nature of the Twitter dynamics Zubiaga (2018).

B.2 CheckThat ’21 Word Overlaps

Next, we analyzed the distribution of the Jaccard scores in the CheckThat ’21, shown in Figure 4. The distribution is different compared to the one observed in our newly collected dataset, as it peaks at around 0.4, and is slightly shifted towards lower similarity values, suggesting that the examples included are not easily solvable with basic lexical features Shaar et al. (2021), which we also observe in our experiments (see Section 4).

Refer to caption
Figure 4: Distribution of the Jaccard similarity scores. The score is an average of the sim(tweet, title) and sim(tweet, subtitle).

Appendix C Annotations

Setup and Guidelines

Each annotator was provided with the guidelines and briefed by one of the authors of this paper. For annotation, we used a Google Sheets document, where none of the annotators had access to the annotations by the others.

The annotation sheet contained the following fields:

  • tweet_text: the text of the fact-checking tweet;

  • text_conversation: the text of the root of the conversation;

  • text_reply: the text of the last tweet before the fact-checking one;

  • title: the title of the Snopes article;

  • subtitle: the subtitle of the Snopes article.

The annotation task was to mark whether the conversation matches and also whether the reply matches using check-boxes. We also allowed the annotators to add comments as a free-form text.

Demographics

We recruited three annotators: two male and one female, between 25 and 30 years old, with higher education (at least a bachelors degree), and currently enrolled in a MSc or PhD programs in Computer Science. Each annotator was proficient in English, but they were not native speakers.

Inter-Annotator Agreement

Here, we present the inter-annotator agreement. We measure the overall agreement using Fleiss kappa (Fleiss, 1971) (shown in Figure 8) and also the agreement between each two annotators using Cohen’s Kappa (shown in Table 9). The overall level of agreement between the annotators is good. Moreover, we can see that between annotator A and C the agreement is almost perfect both for the replies and for the conversations. The lowest agreement is between A and B, but it is still substantial.

Replay Conversation
Fleiss Kappa 0.745 0.750
Table 8: Fleiss Kappa inter-annotator agreement between all three of our annotators: A, B, and C.
Annotators Replay Conversation
Cohen’s Kappa
A \leftrightarrow B 0.650 0.655
A \leftrightarrow C 0.885 0.922
B \leftrightarrow C 0.698 0.673
Table 9: Cohen’s Kappa pairwise inter-annotator agreement between all pairs of our annotators.

Disagreement Analysis

After the annotations procedure was finished, we analyzed the examples for which the annotators disagreed, which fell in the following categories:

  1. (i)

    Claims depending on information from external sources, e.g., ‘Blame Russia again? [URL]’.

  2. (ii)

    Tweets containing multiple claims, for which the referenced article does not target the main claim, e.g., “‘It sounds like someone who is scared as heck that they will not win,” Shermichael Singleton says of Pres. Trump’s remarks encouraging his supporters to vote twice.’ Here, the corresponding crowd fact-check is ‘Did Trump Tell People To Vote Twice?’, i.e., the main claim is in the quote itself, while the remark about voting twice is secondary.

  3. (iii)

    The claim is ambiguous, e.g., ‘Fanta (soft drink) was created so that the Nazi’s could replace Coca-Cola during WWII [URL]’, and the fact-check is about ‘Was Fanta invented by the Nazis?’. Here, it is not clear who created Fanta.

  4. (iv)

    The claim is a partial match, e.g., ‘did President Trump have a great economy and job creation for 1st 3 years???’, and the fact-check is ‘Did Obama’s Last 3 Years See More Jobs Created Than Trump’s First 3?’, which only covers part of the claim in the tweet.

Tweet w/ Claim Snopes Verified Claim and Article
Correct Matches
(1) “Mussolini may have done many brutal and tyrannical things; he may have destroyed human freedom in Italy; he may have murdered and tortured citizens whose only crime was to oppose Mussolini; but ‘one had to admit’ one thing about the Dictator: he ‘made the trains run on time.’” [URL] Italian dictator Benito Mussolini made the trains run on time snopes.com/fact-check/loco-motive/
(2) "Full list of songs Clear Channel banned following the 911 attacks. Some of these don’t make any sense at all. 12 [URL]" Clear Channel Communications banned their American radio stations from playing specified songs in order to avoid offending listeners. snopes.com/fact-check/radio-radio/
(3) @user @user OMG! Were you on this planet when Obama did nothing during H1N1 crisis? Only difference was H1N1 caused more than 10000 deaths and Obama was golfing. Took 6 mos for him to even have a press conference! U.S. President Barack Obama waited until millions were infected and thousands were dead before declaring a public health emergency concerning swine flu. snopes.com/fact-check/obama-wait-swine-flu-n1h1/
Incorrect Matches
(4) Dick Van Dyke? What’s next? Penis Van Lesbian? What. Is. NEXT??? Dick Van Dyke’s real name is Penis Van Lesbian. snopes.com/fact-check/dick-van-dyke/
(5) "I’ve just found a 2012 report on how well informed TV viewers are NPR was top, of course. That’s the one the Republicans want to defund, as it’s contrary to their interests Also Fox viewers were less well informed than people who did not watch TV news at all" A four-year study has found that Fox News viewers have IQs 20 points lower than average. snopes.com/fact-check/news-of-the-weak/
(6) Trump just said he has seen gas prices at $.89-$.99 per gallon. Where I am it is currently $1.70. Anyone see prices Trump is talking about? The COVID-19 coronavirus disease is "spreading quickly from gas pumps." snopes.com/fact-check/covid19-gas-pump-handles/
Table 10: Examples from CrowdChecked, showing correct () and incorrect matches (). The examples in each group are sorted by their overlap with the claim made in the tweet.

Tweet-Article Pairs Analysis

In Table 10, we show examples of correct () and incorrect () matching pairs. We sorted the examples within each group based on the word overlap between the claim and the verified claim, e.g., (1) and (2) have more words in common between the two texts compared to the overlaps in (3), and similarly for (4)–(6).

First, we can see that high overlap does not guarantee a correct matching tweet–article pair, just like low overlap does not mean an incorrect pair, which is also visible from the analysis of the Jaccard similarity in Table 3. These two phenomena can be seen in (3), which contains a correct pair with low overlap, and in (4), where there is an incorrect match with high overlap. Next, some tweets may not contain a claim such as (4), as the user only asks questions, rather than stating something that can be fact-checked. In contrast, (6) contains a verifiable claim about gas prices, but the linked Snopes article fact-checks whether COVID spreads through gas pumps, which is irrelevant in this case. Row (5) is a partial match, and the tweet contains a check-worthy claim, but the article by the crowd fact-checker focuses on the IQ of the Fox News viewers, rather than on how well informed they are, and thus again the match is incorrect. Finally, in row (1), we can see that the verified claim is almost exactly included in the tweet, which is an easy case to match. In contrast, for the example in row (3), the model should do a semantic match based on some prior knowledge that the other name for influenza A virus subtype H1N1 is swine flu, and moreover, 10,000 should be associated with the word thousands.

Appendix D Experimental Results

Model MRR P@1 MAP@5
Baselines (CheckThat ’21)
Retrieval (Shaar et al., 2021) 76.1 70.3 74.9
SBERT (CheckThat ’21) 87.97 84.92 87.45
CrowdChecked (Our Dataset)
SBERT (cos > 0.50) 88.20 85.76 87.80
SBERT (cos > 0.60) 87.21 84.25 86.69
SBERT (cos > 0.70) 86.18 83.08 85.76
SBERT (cos > 0.80) 83.57 80.40 82.93
SBERT (jac > 0.30) 88.01 85.09 87.61
SBERT (jac > 0.40) 87.26 84.76 86.80
SBERT (jac > 0.50) 86.53 83.42 86.13
(Pre-train) CrowdChecked, (Fine-tune) CheckThat ’21
SBERT (cos > 0.50, Seq) 89.92 87.60 89.49
SBERT (cos > 0.60, Seq) 89.56 87.27 89.20
SBERT (cos > 0.70, Seq) 88.70 85.59 88.36
SBERT (cos > 0.80, Seq) 88.42 85.26 88.03
SBERT (jac > 0.30, Seq) 90.21 87.44 89.69
SBERT (jac > 0.40, Seq) 89.64 86.77 89.25
SBERT (jac > 0.50, Seq) 89.44 86.26 89.03
(Mix) CrowdChecked and CheckThat ’21
SBERT (cos > 0.50, Mix) 89.47 86.77 88.99
SBERT (cos > 0.60, Mix) 88.54 85.76 87.98
SBERT (cos > 0.70, Mix) 87.71 84.92 87.18
SBERT (cos > 0.80, Mix) 88.40 85.26 87.97
SBERT (jac > 0.30, Mix) 90.41 87.94 90.00
SBERT (jac > 0.40, Mix) 89.82 86.60 89.48
SBERT (jac > 0.50, Mix) 88.71 85.26 88.31
Table 11: Evaluation on the CheckThat ’21 development set. In parenthesis is shown the name of the training split, i.e., Jaccard (jac) or Cosine (cos) for data selection strategy, (Seq) for first training on CrowdChecked and then on CheckThat ’21, and (Mix) for mixing the data from the two datasets.

Here, we present the expanded results for our experiments described in Section 4. Tables 11 and 12 include the results for the threshold selection analysis experiments on the development dataset, and on the testing dataset, respectively. Here, Table 12 corresponds to Table 6 in the main text of the paper, and includes all metrics and all thresholds (shown in Figure 3). Next, the results from our Modeling Noisy Data experiments are in Table 13, which corresponds to Table 7 in the main paper. In all tables, we use the same notation and grouping as in the corresponding table in the main paper.

Precision MAP
Model MRR @1 @3 @5 @10 @20 @1 @3 @5 @10 @20
Baselines (CheckThat ’21)
Retrieval (Shaar et al., 2021) 76.1 70.3 26.2 16.4 8.8 4.6 70.3 74.1 74.9 75.7 75.9
SBERT (CheckThat ’21) 79.96 74.59 27.89 17.19 8.96 4.61 74.59 78.66 79.20 79.66 79.83
CrowdChecked (Our Dataset)
SBERT (cos > 0.50) 81.58 75.91 28.60 17.76 9.04 4.67 75.91 80.36 81.05 81.27 81.48
SBERT (cos > 0.60) 79.71 74.75 27.39 16.96 8.86 4.59 74.75 78.25 78.84 79.38 79.61
SBERT (cos > 0.70) 78.27 72.28 27.61 17.10 8.89 4.53 72.28 76.95 77.54 78.01 78.12
SBERT (cos > 0.80) 78.39 72.94 27.34 16.83 8.81 4.55 72.94 77.04 77.52 78.08 78.28
SBERT (jac > 30) 81.50 76.40 28.49 17.43 8.94 4.65 76.40 80.45 80.84 81.14 81.38
SBERT (jac > 40) 79.45 74.42 27.34 16.93 8.89 4.65 74.42 77.92 78.52 79.08 79.33
SBERT (jac > 50) 79.96 74.75 27.89 17.29 8.94 4.60 74.75 78.63 79.26 79.63 79.81
(Pre-train) CrowdChecked, (Fine-tune) CheckThat ’21
SBERT (cos > 0.50, Seq) 82.26 77.06 28.27 17.62 9.26 4.76 77.06 80.64 81.41 81.99 82.18
SBERT (cos > 0.60, Seq) 80.13 75.41 27.45 17.00 8.94 4.65 75.41 78.55 79.13 79.76 79.99
SBERT (cos > 0.70, Seq) 79.27 73.43 27.72 17.33 8.94 4.58 73.43 77.78 78.56 78.94 79.09
SBERT (cos > 0.80, Seq) 78.32 72.77 27.17 16.93 8.89 4.58 72.77 76.71 77.41 77.98 78.15
SBERT (jac > 0.30, Seq) 83.76 78.88 28.93 17.82 9.21 4.71 78.88 82.59 83.11 83.49 83.63
SBERT (jac > 0.40, Seq) 80.69 75.25 27.83 17.33 9.09 4.69 75.25 79.04 79.76 80.34 80.57
SBERT (jac > 0.50, Seq) 81.99 76.90 28.16 17.76 9.13 4.69 76.90 80.34 81.33 81.70 81.88
(Mix) CrowdChecked and CheckThat ’21
SBERT (cos > 0.50, Mix) 82.12 76.57 28.55 17.59 9.13 4.68 76.57 80.86 81.38 81.82 82.00
SBERT (cos > 0.60, Mix) 81.45 76.40 28.27 17.43 8.96 4.61 76.40 80.25 80.79 81.14 81.31
SBERT (cos > 0.70, Mix) 79.08 73.10 27.83 17.33 8.89 4.57 73.10 77.72 78.46 78.77 78.95
SBERT (cos > 0.80, Mix) 79.73 74.75 27.56 17.00 9.06 4.62 74.75 78.22 78.73 79.46 79.59
SBERT (jac > 0.30, Mix) 83.04 78.55 28.66 17.52 9.11 4.69 78.55 81.93 82.30 82.75 82.94
SBERT (jac > 0.40, Mix) 81.18 74.59 28.55 17.72 9.14 4.74 74.59 79.79 80.46 80.85 81.10
SBERT (jac > 0.50, Mix) 81.56 76.73 28.22 17.36 9.03 4.71 76.73 80.23 80.71 81.19 81.45
Table 12: Evaluation on the CheckThat ’21 test dataset. In parenthesis is shown the name of the training split: Jaccard (jac) or Cosine (cos) for data selection strategy, (Seq) for first training on CrowdChecked and then on CheckThat ’21, and (Mix) for mixing the data from the two datasets.
Precision MAP
Model MRR @1 @3 @5 @10 @1 @3 @5 @10
DIPS (Mihaylova et al., 2021) 79.5 72.8 28.2 17.7 9.2 72.8 77.8 78.7 79.1
NLytics (Pritzkau, 2021) 80.7 73.8 28.9 17.9 9.3 73.8 79.2 79.9 80.4
Aschern (Chernyavskiy et al., 2021) 88.4 86.1 30.0 18.2 9.2 86.1 88.0 88.3 88.4
SBERT (jac > 0.30, Mix) 83.0 78.6 28.7 17.5 9.1 78.6 81.9 82.3 82.8
+ shuffling & trainable temp. 83.2 77.7 29.1 17.8 9.1 77.7 82.2 82.6 82.9
+ self-adaptive training (Eq. 1) 84.2 78.7 29.3 18.1 9.3 78.7 83.0 83.6 83.9
       + loss weights 84.8 79.7 29.5 18.2 9.3 79.7 83.7 84.3 84.6
        + TF.IDF + Re-ranking 89.9 86.1 30.9 18.9 9.6 86.1 89.2 89.7 89.8
          + TF.IDF + Re-ranking (ens.) 90.6 87.6 30.7 18.8 9.5 87.6 89.9 90.3 90.4
Table 13: Results on the CheckThat ’21 test dataset. We compare our model and its components (added sequentially) to three state-of-the-art approaches.