ELF22: A Context-based Counter Trolling Dataset to Combat Internet Trolls
Abstract
Online trolls increase social costs and cause psychological damage to individuals. With the proliferation of automated accounts making use of bots for trolling, it is difficult for targeted individual users to handle the situation both quantitatively and qualitatively. To address this issue, we focus on automating the method to counter trolls, as counter responses to combat trolls encourage community users to maintain ongoing discussion without compromising freedom of expression. For this purpose, we propose a novel dataset for automatic counter response generation. In particular, we constructed a pair-wise dataset that includes troll comments and counter responses with labeled response strategies, which enables models fine-tuned on our dataset to generate responses by varying counter responses according to the specified strategy. We conducted three tasks to assess the effectiveness of our dataset and evaluated the results through both automatic and human evaluation. In human evaluation, we demonstrate that the model fine-tuned on our dataset shows a significantly improved performance in strategy-controlled sentence generation.
Keywords: Natural language, countering trolls, conditional text generation
languageresourceLanguage Resources
ELF22: A Context-based Counter Trolling Dataset to Combat Internet Trolls
Huije Lee∗, Young Ju NA, Hoyun Song, Jisu Shin, Jong C. Park |
KAIST |
Daehak-ro 291, Yuseong-gu, Daejeon, Republic of Korea |
{jae4258,hysong,jsshin,park}@nlp.kaist.ac.kr |
[email protected] |
Abstract content
1. Introduction
**footnotetext: equal contribution$\dagger$$\dagger$footnotetext: Present Address: Young Ju NA, Université Sorbonne Nouvelle, 17 rue de Santeuil 75005 ParisThe International Telecommunication Union (ITU) stated that the number of internet users increased from 4.1 billion in 2019 to 4.9 billion in 2021, as 782 million people started to participate in cyberspace activities during the COVID-19 pandemic. The Pew Research Center in 2021 reported that 93% of American adults participate in internet communities to interact with others. Together with the positive side of online communities that promote open discussion and facilitate social movements, the number of users experiencing online trolling or abuse has also increased. The Australia Institute [Australia Institute, 2019] reported the estimated social cost of online trolling and abuse at $3.7 billion in 2019.
Online trolling is a malicious attempt to provoke strong reactions from an interlocutor. Trolls, it would seem, do not care what they are saying as long as they have an impact on the chosen target [Fichman and Sanfilippo, 2016, Mihaylov and Nakov, 2016, Golf-Papez and Veer, 2017]. Regarding the issue of online trolls, there have been attempts by community managers such as Twitter, Facebook, and Reddit to delete troll comments or by users to ignore them [Sibai et al., 2015]. However, compared to the troll attackers, the voice of the victim group is overall weak. For example, Asian hate speech related to COVID-19 showed a large difference in the number of tweets, as there were 3 times more users displaying hate speech than those displaying counter speech, and hateful bots were more active and more successful in attracting followers than counter-hate bots [Ziems et al., 2020].
To counteract the problem posed by online trolling, we propose to adopt a method of actively countering trolls so as to regulate abusive language online. This allows the online discussion to keep going, rather than to see certain abusive messages deleted and the flow of conversation interrupted. As argued also by Golf-Papez and Veer [Golf-Papez and Veer, 2017], counter responses help support the users and protect them from trolls by neutralizing the negative impact of trolling. In countering trolls, we also propose to follow the line of making more utterances against trolls [Richards and Calvert, 2000], with controlled strategies [Hardaker, 2015].
Compared to counter speech studies [Qian et al., 2019, Chung et al., 2019, Tekiroğlu et al., 2020, Laugier et al., 2021], research on automatic sentence generation to counter trolls is still in its infancy. Despite the abundance of troll detection datasets [Mihaylov and Nakov, 2016, Wulczyn et al., 2017, Atanasov et al., 2019], there are few datasets with the rich contextual information needed to improve the quality of counter responses against trolls. Addressing this situation, we present a novel dataset, which can be used for automatically generating responses to combat trolls, acting as an “Elf bot”. For this purpose, we first collected pairs of troll comments and their counter responses in the Reddit community. We then formulated a scheme according to the seven counter response strategies [Hardaker, 2015]: engage, ignore, expose, challenge, critique, mock, and reciprocate. In order to help annotators understand counter response strategies more clearly, we categorized trolls into two types [Hardaker, 2013]: overt and covert trolls. We then examined the effectiveness of our annotated dataset. Finally, we assessed the quality of counter responses generated by strong baselines for a given response strategy through automatic and human evaluation.
We highlight three major contributions of our paper: 1) We provide the first dataset for countering trolls labeled with counter strategies and rich contextual text; 2) we show the use case of the dataset with robust baselines for three tasks: binary classification for troll strategies, multi-class classification for counter response strategies, and counter response generation; and 3) we demonstrate the effectiveness of the dataset through automatic and human evaluation on counter responses generated by models fine-tuned over this dataset.
The rest of the paper is organized as follows: In Section 2, we introduce related work on troll definition, troll detection, and counter speech. We explain the detailed labels of troll comments and counter responses for data annotation in Section 3. We describe the process of collecting and labeling for the dataset with statistics in Section 4. In Section 5, we describe three models trained over our dataset with their performances and demonstrate the effectiveness of our dataset through automatic and human evaluation. Finally, we show conclusion and future work in Section 6.
2. Related Work
Trolls can be criminalized and prosecuted for unlawful behavior and illegal activities that include intentional provocation of anguish [Golf-Papez and Veer, 2017]. Trolling behaviors encompass the intention that causes emotional distress to the targets111The 2021 Florida Statutes 784.048, retrieved from http://www.leg.state.fl.us/Statutes/index.cfm, and are grossly offensive, false, indecent, menacing, provocative or disturbing222Communications Act 2003, retrieved from https://www.legislation.gov.uk/ukpga/2003/21/section/127 that can amass to a criminal offense. Trolls communicate to manipulate people’s opinions and make mischievous attempts to elicit reactions from their targets, or, interlocutors. Trolls, or attention-seekers, aim to provoke responses from their targets through a variety of communication tactics such as hyper-critism, shaming, sarcasm, blackmailing, and publishing of private information [Fichman and Sanfilippo, 2016, Mihaylov and Nakov, 2016, Golf-Papez and Veer, 2017].
Taking into consideration these diverse tactical aspects of trolls, several studies constructed datasets with labeled troll types in order to detect trolls in online communities [Mihaylov and Nakov, 2016, Wulczyn et al., 2017, Atanasov et al., 2019]. Other studies have proposed a troll detection model by using an annotated dataset of troll comments [Kumar et al., 2014, Georgakopoulos et al., 2018, Chun et al., 2019, Atanasov et al., 2019]. Jigsaw and Google ran the Perspective project, providing APIs333https://www.perspectiveapi.com to score the text’s toxicity, together with a troll dataset444https://www.kaggle.com/c/jigsaw-toxic-comment-classification-challenge. While these studies focused only on the detection of trolls, we categorized trolls by their overtness and identified the best counter strategies by their types.
Other than such numerical values of toxicity of contents, there are some studies that focused on counter speeches to combat abusive language online. Mathew et al. [Mathew et al., 2018, Mathew et al., 2019] and Chung et al. [Chung et al., 2019] proposed datasets for counter speech with a taxonomy proposed by Benesch et al. [Benesch et al., 2016]. Subsequent studies have shown that these datasets have significantly improved the performance of counter speech generation models [Chung et al., 2020, Tekiroğlu et al., 2020, Laugier et al., 2021]. Unlike previous studies, however, we focused on trolling to identify counter strategies against each troll type for diversified counter response generation.
3. Data Annotation Scheme
Aggress is directly cursing or swearing others without any justification | |
Overt strategies | Shock is throwing an ill-disposed or prohibited topic that is avoided for political or religious reasons. |
Endanger is providing disinformation with the intent to harm others, and discovering this purpose by others. | |
Antipathize is creating a sensitive discussion that evokes an emotional and proactive response in others. | |
Covert strategies | Hypocriticize is excessively expressing disapproval of others or pointing out faults to the extent that it feels intimidating to others. |
Digress is making a discussion to be derailed into irrelevant or toxic subjects. |
Title | Post | Troll comment | Response | TL | RL |
If I’m not going to vaccinate myself, why? | Just heard that NC is considering giving portions of doses on-hand back to feds. If you’ve decided to not get jabbed, what’s your reasoning? | I am glad you and a bunch of dumbs live in a nation that lauds your ignorance. covid is going to kill some of you idiots moving forward. | I got my shots, TYVM. I asked in general to attempt an antagonizing dialog with folks. Please try better, and remember you catch far more flies with honey than vinegar. | 1 | 5 |
I think you guys complain too much | Everday I see posts like “there’s too much damage”, “too much mobility”, … I don’t know. LoL has 140 champions and they all sit between 45-55% winrate, Riot got the one of the most popular games out there for 10+ years. | cringe post you can still delete this | Cringe comment You can still delete this | 1 | 7 |
3.1. Troll Strategy
A troll teases people to make them angry, offends people, wants to dominate any single discussion, or tries to manipulate people’s opinions [Mihaylov and Nakov, 2016]. Trolling shows a specific type of aggressive online behavior involving antagonism for the sake of amusement such as a deliberate, deceptive and mischievous attempt to provoke a reaction from other online users [Herring, 2002, Shin, 2008, Binns, 2012, Hardaker, 2013, Golf-Papez and Veer, 2017]. In particular, Hardaker [Hardaker, 2013] introduced the notion of overt and covert trolls with a detailed scale of trolling bahaviour in six categories: aggress, shock, endanger, antipathize, hypocriticize, and digress. Since we focus on counter responses to combat trolls, we used two super-categories, overt and covert strategies, to classify trolls by their detailed tendencies (see Table 1).
3.2. Response Strategy
The function of countering trolls is to be a “virtually capable guardian”, who can neutralize the impact of trolling on targets. Note that when devising a counter response, it is not always possible to identify who triggered the action of trolling through comments [Hardaker, 2013]. Therefore, we followed the taxonomy of using seven response strategies for counter responses, introduced by Hardaker [Hardaker, 2015], as follows.
Engage strategy indicates acting in accordance with the troll’s intentions. This strategy involves accepting the troll’s opinions positively and responding sincerely to the troll. From the third party’s point of view, however, the users employing this strategy could be considered to have fallen prey to the troll.
Ignore strategy has the goal of not giving what the troll wants. Users attempt to warn others about doubtful behaviors of the troll. The strategy is to deter trolls from their malicious intention so that the trolls leave the discussion. For this strategy to work, it is necessary to be proactive in discovering the troll’s intentions.
Expose strategy takes a stance either going against the troll’s opinion or doubting the authenticity of the troll’s information. This strategy has the risk of being trolled by responding to the message of trolls; however, it could keep the discussion consistent and healthy.
Challenge strategy provides direct opposition to the troll with an aggressive response. Users often use emotional language to express their hostility. This strategy has a similar purpose to expose in terms of confronting trolls, but challenge tries go more strongly against what trolls have said.
Critique strategy goes beyond discovering the true intention of the troll, and overtly judges the troll’s behavior as degrading and uncreative. This strategy weakens the troll’s effectiveness by evaluating its actions rather than the content of the troll comment.
Mock strategy involves making the troll scorned and ridiculed. With mocking, users attempt to isolate the troll while strengthening the cohesion of discussion participants.
Reciprocate strategy attempts to attack the troll by pretending to be another troll. This strategy takes a more radical stance than the challenge strategy in that the purpose is to attack the troll rather than to self-defend against the troll.
4. ELF22 Dataset
We present a troll-response pairwise dataset, ELF22, along with contextual text from posts in the Reddit community. Our dataset contains posts with troll comments and counter responses. Each troll comment and counter response pair is annotated with a proposed strategy according to the data annotation scheme (see Section 3). In this section, we present a detailed description of our data labels, the data collection process, data statistics, and ethical considerations of the given dataset.
4.1. Data Collection
For the data collection, we crawled the posts that include troll comments from Reddit using Pushshift API [Baumgartner et al., 2020]. We describe how we limited our conditions as well as some of the significant parts of our data collection process. First, we extracted our data from diverse subreddit communities, taking into account the fact that trolls are “light speakers”, who are not, or do not have the capacity to be, deeply involved in a conversation, as opposed to “heavy speakers”, who present logical arguments to discuss with others [Choi et al., 2015]. Second, we focused on the importance of the fast interactions and interchange of messages amongst online community users, which aggravates the propagation of troll comments, since virality, responsiveness and volume are the three important factors in terms of conversation characteristics of Reddit [Choi et al., 2015]. Third, we considered the title and body text of the post as context. In order to exclude the possibility of disconnection in the context history, we only extracted root comments if they were described to be by trolls.
To collect unified data types, we limited our range of extracting data of online trolls. Concretely, we set the range of thumbs down of troll comments from -2 to -15. The upper limit of a downvoted comment was set to -15, as it excludes comments heavily biased against community tendencies. Next, we extracted the highest-scored comment of thumbs up among the following comments from the root comment as a counter response reference.
We discarded posts if they contained URLs, included pictures, had more than 512 characters, or used language other than English. We then selected the posts whose number of characters is at least 12. Samples of the collected data are shown in Table 2.
4.2. Data Annotation
For this annotation task, we recruited 12 annotators in total who are familiar with the Reddit website. Native English speakers or those whose education took place in a country where English is used as an official language participated in this experiment since our collected dataset is in English.
The annotation phase of the dataset is organized in two sessions, each of which took 7 days to complete. We asked all the annotators to follow the guidelines which include the definition of trolls, the importance of combating them, and an explanation of each strategy with examples (see Section 3). Since these sessions were conducted online, we devised a strictly controlled environment in order to obtain a qualified dataset. Each annotator had his or her own personalized Google sheet link to annotate a scheduled amount of the dataset every day. Each day, annotators submitted their labeled dataset in order to receive a confirmation from our team. In addition to this, we regularly opened online QA sessions to help annotators, at the same time ensuring that they were faithfully engaged and fully familiar with the task.
For the first annotation session, annotators were given Reddit posts with subreddit names, titles, and body text, including two target sentences to annotate: a ‘troll comment’ and its ‘response’. Annotators were asked to read the troll comment and label it according to whether they considered it to be a ‘covert’ or ‘overt’ type of troll. They were then asked to read the response, which was considered to be a combating sentence against the troll, before giving it a label according to the 7 strategies from the guidelines. We gave the annotators an option to propose their own choice of a counter response strategy against the troll if they believed that the given ‘response’ was considered ineffective. Since the entries of the dataset were systematically distributed so that three annotators could work on the same entries of the dataset, we found disagreements among the annotators after the first session. Therefore, for the second annotation session, if all the three annotators gave a different label to the text, we asked them to re-consider their choices by discussing among them and re-labeling the data. We chose the majority labels otherwise.
Table 2 shows examples of the five elements in each entry: subreddit (omitted in Table 2), title, post, troll (troll comment), and response (counter response). The first entry is a post about vaccination and asking the opinion of others regarding the decision of refusing to be vaccinated. The comment is identified as a troll by the subreddit users, which we interpret as a troll that indirectly aggresses other locutors but does not directly name any reddit user.
4.3. Data Statistics
Avg. # words | Max. # words | |
Title | 12.2 | 82 |
Body | 57 | 205 |
Troll comment | 38.4 | 126 |
Response | 25.6 | 128 |
Total | 133.2 | - |
Strategy | Train | Validation | Test | Total |
Overt | 2,331 | 166 | 340 | 2,837 |
Covert | 3,020 | 407 | 422 | 3,849 |
Engage | 2,134 | 340 | 343 | 2,817 |
Ignore | 216 | 10 | 35 | 261 |
Expose | 1,131 | 75 | 153 | 1,359 |
Challenge | 818 | 48 | 71 | 937 |
Critique | 340 | 31 | 52 | 423 |
Mock | 592 | 51 | 96 | 739 |
Reciprocate | 120 | 18 | 12 | 150 |
Total | 5,351 | 573 | 762 | 6,686 |
First, we collected from Reddit 5,700 posts that include 2,198 unique subreddits. The top 10 subreddits of the collected posts were: unpopularopinion, wallstreetbets, teenagers, nba, relationship_advice, DestinyTheGame, NoStupidQuestions, Genshin_Impact, fo76, and NonewNormal.
We removed posts with incomplete composition during the annotation process. If a sentence was labeled differently such as expose, challenge, and reciprocate by three annotators even after the second annotation session, we have considered it as outliers and discarded it. We only extracted the labeled posts that obtained the majority agreement among three annotators. After filtering, we finalized our dataset with a total 5,535 pairs of troll comments and counter responses, together with 1,151 ‘your response’ sentences that were made by the annotators.

Table 3 shows information about the average and maximum numbers of words of our dataset, where the average numbers of words of contextual texts, troll comments, and responses are 69.2, 38.4, and 25.6, respectively.
Figure 1 shows the distribution of response strategy types in our dataset. We see that Reddit users often used the engage strategy, followed by the exposure and challenge strategies. On the other hand, the ignore and reciprocate strategies occupied only a small portion of the dataset.
Table 4 presents the statistics of our dataset with respect to each label. The labeled posts are randomly split into training (80%), validation (10%), and test (10%). We note that each post has a unique post number in accordance with trolling comments. This means that we assigned the same post number for the comments of countering trolls and ‘your response’ sentences. The consequent distribution of entries of the ELF22 dataset was training (80%), validation (8.6%), and test (11.4%).
To check the validity and consistency of the collected dataset, we performed the Inter-Annotator Agreement (IAA). The Fleiss’s Kappa () value of the troll type label of our dataset was 0.539, and the value of the counter strategy label was 0.465, achieving a ‘moderate’ agreement between the two labels [McHugh, 2012].
4.4. Ethical Consideration
Our annotation task was approved by the Institutional Review Board (IRB)555Approval number: KH2021-126. All participants of annotation tasks indicated their understanding of the procedure for the annotation and acknowledged their agreement to participate. We report that all data are anonymized, and we plan to ensure that future researchers follow ethical guidelines, with the necessary information collected as the occasion demands.
The goal of our work is to categorize responses against trolls in online conversations and support the development of generation bots for countering trolls according to seven strategies proposed in this paper. As shown in Tables 2, 8, and 9, our dataset may contain sarcastic and aggressive language. We tried to observe how they communicate as-is, even though it could include socially biased content or hate speech. We expect that our dataset will be helpful for future work to study effective methodologies for responding to troll language online.
When deploying language models for response generation on our dataset, the models could generate unintended results, such as social and racial biases. There could be solutions including list-based token filtering, prompt conditioning [Keskar et al., 2019], domain-adaptive pre-training [Gururangan et al., 2020], or any detoxification techniques [Gehman et al., 2020] to avoid unintended response generation in accordance with the research purpose.
5. Experiments
Task A | Task B | |||||||
P | R | wF1 | MF1 | P | R | wF1 | MF1 | |
SVM | 0.58 | 0.58 | 0.58 | 0.57 | 0.32 | 0.34 | 0.33 | 0.19 |
RF | 0.60 | 0.59 | 0.54 | 0.52 | 0.30 | 0.45 | 0.28 | 0.09 |
BERT | 0.64 | 0.63 | 0.63 | 0.63 | 0.48 | 0.47 | 0.47 | 0.27 |
RoBERTa | 0.64 | 0.64 | 0.64 | 0.64 | 0.46 | 0.42 | 0.42 | 0.28 |
Task C | ||||
R-L | B-1 | M | BS | |
GPT-2 | 0.04 | 0.04 | 0.08 | 0.29 |
BART | 0.08 | 0.08 | 0.16 | 0.41 |
DailoGPT-ELF22 | 0.06 | 0.14 | 0.04 | 0.35 |
GPT-2-ELF22 | 0.06 | 0.15 | 0.04 | 0.34 |
BART-ELF22 | 0.10 | 0.15 | 0.09 | 0.40 |
5.1. Experimental Setup
In this section, we describe three tasks that are used to exploit our dataset: troll strategy classification as binary classification, response strategy classification as multi-label classification, and generating counter responses as text generation.
Task A. troll strategy classification This task involves classifying troll comments with titles and body texts and stating whether the comments are overt trolls or covert trolls. We implemented support vector machine (SVM) and random forest (RF) as dictionary-based classifiers, and BERT [Devlin et al., 2019] and RoBERTa [Liu et al., 2019] as pre-trained transformer-based classifiers. We utilized the pre-trained transformer models from the Huggingface library [Wolf et al., 2019]. The hyperparameter setting for each model is as follows:
SVM We used a linear kennel with a hyperparameter value of 0.01 out of [0.01, 0.1, 1, 10, 100, 1000, 10000] and a gamma of ‘auto’.
RF We used a hyperparameter value of 100 out of [0.01, 0.1, 1, 10, 100, 500, 1000].
BERT We used ‘bert-base-cased’ pre-trained checkpoint and fine-tuned the model on our train dataset with 10 epochs. The size of the batch was 32, the learning rate was -, the weight decay was -, the warm-up steps were 100, the drop-out rate was set to 0.1, and the max sequence length was 512, which are the same as those that Devlin et al. [Devlin et al., 2019] used to fine-tune models.
RoBERTa We used a ‘roberta-base’ pre-trained checkpoint with a similar number of model parameters to BERT. We followed hyperparameters of the same values as BERT.
For dictionary-based classifiers, the words with the top-512 frequencies from the title, body text, and troll comment were included in a bag of words as input. We prepared a list of candidate hyperparameters for a grid search and selected the one with the best weighted F1 score in the validation dataset. Training times of BERT and RoBERTa were about 10 minutes on one NVIDIA A100-PCIE-40GB. We reported the average results under 5 fine-tuning runs from our test dataset.
Task B. response strategy classification This task involves classifying responses with titles, body texts, and troll comments, stating whether the responses are used in one out of 7 counter strategies. We experimented with the same baselines used in Task A. Aside from 0.1 being selected for SVM , we used the same hyperparameter settings for each model as in Task A.
Task C. counter response generation This task involves generating responses with titles, body texts, troll comments, and labeled strategies. We implemented the following competitive baseline models:
BART [Lewis et al., 2020] is a transformer encoder-decoder model that is trained on Common Crawl News, Wikipedia, book corpus, and stories. We employed a ‘bart-base’ model that contained 6 encoder layers and 6 decoder layers. The size of the batch was 32.
GPT-2 [Radford et al., 2019] is a transformer decoder model that is trained on Common Crawl Webtext. We employed a ‘gpt2’ model that contained 12 decoder layers. The size of the batch was 16.
DialoGPT [Budzianowski and Vulić, 2019] is an additionally pre-trained GPT-2 model on a dialog corpus. Fine-tuning configuration is the same as that of the GPT-2 model.
We fine-tuned the models over 20 epochs with a learning rate of -, a weight decay of -, a drop-out rate of 0.1, a max input sequence length of 768, and a max output response length of 128.
In the decoding phase, we employed a beam search with the number of beams set to 5 and with blocking trigram repeats, which is widely used to avoid low-quality generation results [Klein et al., 2017, Paulus et al., 2017]. Based on the GPT2 tokenizer, the average token lengths of train, validation, and test input were 124.2, 122.9, 121, respectively. The average token lengths of train, validation, and test ground-truth responses were 28.1, 29.1, 25.5, respectively. Training times of GPT2, DialoGPT, and BART were each about 30 minutes on one NVIDIA A100-PCIE-40GB.
5.2. Automatic Evaluation
Tasks A and B. For the classification tasks, we evaluated the weighted Precision (P), Recall (R), F1 score (wF1), and macro F1 score (MF1).
Task C. For the generation task, we performed automatic evaluation and human evaluation. On Task C., we used ROUGE-L F1 score [Lin, 2004], BLEU-1 [Papineni et al., 2002], METEOR [Banerjee and Lavie, 2005], and BERTScore [Zhang et al., 2020] for the automatic evaluation. ROUGE-L F1, BLEU-1, and METEOR are based on n-gram overlaps that evaluate the similarity between the ground-truth response and the predicted one. BERTscore is a machine learning based automatic metric that captures semantic similarities by comparing their contextual embeddings.
5.3. Results
Table 5 shows the performance of the models on Tasks A and B. The models are fine-tuned on our train dataset and evaluated against the ground-truth labels for 762 test samples. The pre-trained transformer models BERT and RoBERTa outperformed the machine learning-based models SVM and RF. It implies that labeled strategies in our dataset are independent of simple lexical features, while sufficiently considering contextual text. For Task A, the wF1 scores of SVM and RF were close to 0.5, slightly higher than the random selection performance of binary classification. For Task B, we observed that the transformer-based models had about 50% higher wF1 performance than the machine learning-based models. We speculate that Task B needs to consider semantic features to improve the performance. We also see that SVM and RF show low F1 performance on reciprocate and ignore strategies that contain a small number of train samples, resulting in poor MF1 performance. On the other hand, BERT and RoBERTa show robust performance with the unbalanced dataset.
Table 6 shows the performance of the generation models on Task C. To confirm the effectiveness of the proposed dataset, we experimented with two pre-trained models and three fine-tuned models. Three fine-tuned models showed higher BLEU scores than pre-trained models. We find that BART-ELF22 shows the highest ROUGE and BLEU scores, indicating that words in generated response have the most overlap with ground-truth references. BART showed the highest METEOR performance, which means that it achieved high n-gram precision and recall with the ground-truth references. Aside from BLEU, BART and BART-ELF22 showed high performance on ROUGE, METEOR, and BERTScore. The high BERTScores of BART and BART-ELF22 imply that the generated sentences of the two models derive outputs semantically similar to the references.
5.4. Human Evaluation
To verify that the proposed dataset is enough to train a generation model which produces counter sentences against trolls with a specified response strategy, we conducted human evaluation on counter response generation models. Automatic evaluation metrics are related to the quality of generated texts, but they underestimate texts with a smaller vocabulary and heavily rely on ground-truth references. In particular, this phenomenon is frequently observed due to the characteristics of the Reddit community, which has a lot of free discussions and dialogues. To demonstrate the effectiveness of our dataset while embracing this phenomenon, we evaluated the counter response texts based on two criteria: relevance and compatibility.
First, we experimented with how well our dataset could be used to train a generation model to generate a counter response of a certain strategy when there is a corresponding golden response strategy and text. We evaluated generated texts from the ‘relevance’ point of view [Zhu and Bhat, 2021]. Relevance includes an assessment of the following two factors: 1) how coherent or contextually matched the generated counter response is with the troll text, and 2) how well the generated text counters troll text. We ranked differently generated responses to assess relevance. The higher rank refers to more contextually coherent and effective responses to troll comments. On the other hand, the lower rank refers to responses that may be out of context and endanger the troll victims.
Second, we assessed how effectively our dataset could help train a generation model to create counter responses that correspond to strategies not in the dataset from the ‘compatibility’ perspective. The dataset consists of pairs of a troll comment and its response, with only one golden response strategy and response text to one troll comment. In this regard, we checked if it was necessary to evaluate whether the dataset would be sufficiently effective for the model to generate responses corresponding to various but untrained response strategies against a troll text. Compatibility captures how well the model generates responses to given strategies, scored on a Likert scale of 1 to 5. If a counter response is related to the given strategy and close to a human response, the response gets a higher score.
A total of 5 evaluators fluent in English and familiar with Reddit forums participated in this human evaluation task. These five evaluators were equally assigned with 100 randomly selected samples from our dataset. Before the evaluation process, all evaluators were introduced to the definition of trolls and the seven strategies of countering trolls.
All the five sentences in a single post were evaluated by five evaluators: three response sentences for relevance evaluation and two response sentences for compatibility evaluation. The three response sentences for relevance evaluation include: (1) the ground-truth response; (2) the GPT-2 generated sentence; and (3) the BART-ELF22 generated sentence666We also performed a test on BART, but the result is excluded from the evaluation because the output is easily discerned. We employed the BART-ELF22 model that performed better in the automatic evaluation of Task C.. Two response sentences include the generation of GPT-2 and that of BART-ELF22 when a randomly selected strategy was given.

Figure 2 shows the results of human evaluation. We confirm that BART-ELF22 (Ours) performed better than GPT-2 (baseline) in both relevance and compatibility.
For the relevance criterion, we compared three responses: ground-truth, a response by BART-ELF22 which was fine-tuned with our dataset, and GPT-2, which was not fine-tuned with our dataset. As shown in the left bar plot in Figure 2, the mean values of ground-truth, BART-ELF22, and GPT-2 were 1.37, 2.12, and 2.51, respectively. According to Friedman test, there was a significant difference in mean rank between three responses (, ). For the post hoc comparison test, we used Wilcoxon Signed Ranks test. The result was in the order of ground-truth, BART–ELF22, and GPT-2, with significance established at . We speculate that fine-tuning with our dataset does improve the quality of the model to generate context-relevant sentences.
For the compatibility criterion, we compared two responses generated by BART-ELF22 and GPT-2. As shown in the right bar plot in Figure 2, the mean values of BART-ELF22 and GPT-2 were 3.22 () and 2.10 (), respectively. According to one-way repeated measure ANOVA test, there was a significant difference in the mean values between two responses (). We see that fine-tuning with our dataset significantly improves the performance of strategy-controlled sentence generation.
5.5. Case Study
Text | |
Title | Smoking kill gain too much? |
Body | I have been working out like crazy and have been seeing good results but I have seen that smoking can kill testosterone cell. I dont smoke that much like 3 cigerattes a day will it affect my gains too much? |
Comment | Everything is good with moderation. Don’t go crazy and everything should be fine.You just…wont gain as much as you normally would if you didn’t smoke. |
RL | in an expose way |
Model | Text |
GT | LOL smoking is not good with moderation. |
GPT-2 | What is wrong with you? I don’t know why you are getting downvoted lol |
BART-ELF22 | Are you talking about the same thing? |
BART-ELF22 (engage) | Well yes and no. I am not saying I don’t see any results in moderation, but I am saying this is a general trend with smokers. If I smoke that much that will not affect my mental health. |
BART-ELF22 (challenge) | Tell that to alcoholics. We all know what we’re getting into here. |
Table 7 shows examples of five corresponding sentences generated by the model when given contextual information and a response strategy labeled in our dataset. In Table 7, the three counter responses above are generated by the ground-truth reference, GPT-2, and BART-ELF22 models, respectively. The two counter responses below are generated by BART-ELF22 with the other strategies.
We find that the counter sentences generated by GPT-2 contain frequently repetitive words and composed of fewer informative tokens regardless of strategies. In BART-ELF22, as in GPT-2, there were some cases where the variation was small even if the given strategy was changed, but the frequency was relatively small compared to GPT-2. We observe that BART-ELF22 generates data by varying counter responses according to the specified strategy. This result suggests that the way the BART-ELF22 appends the strategy phrase-wise to the context does work. We see that the proposed dataset supports improvement on controlled response generation according to the response strategy.
6. Conclusions
We presented a dataset in which types of trolls and response strategies are annotated against troll comments with the help of rich contextual information. We verified the effectiveness of our dataset from the experimental results, which showed that fine-tuning on our dataset improved the performances of all the tasks as well as the quality of counter responses. We also see that the fine-tuned text generation model generated data by varying counter responses according to the specified strategy, even when the specified strategy was not in our dataset. In future work, we will explore the shared features that are potentially obtained from counter speech datasets so as to improve the generation model. With this dataset, we hope to support the community in enabling healthy and open communication. The ELF22 dataset is publicly available at https://github.com/huijelee/ELF22.
7. Acknowledgements
This work was supported by the National Research Foundation of Korea (NRF) (No. 2020R1A2C1010759, A system for offensive language detection and automatic feedback with correction) grant funded by the Korea government (MSIT).
Title | Post | Troll comment | Response | TL | RL |
I’m so sorry guys, but I can’t HODL anymore | Had to sell my DOGE, lost my wife yesterday to a heart attack, and I’m going to need the money to handle the affairs that normally come from the death of a spouse. No, I’m not ok, pretty far from it. Sorry shibes. | Is that the rule : Rule no 1 never invest money you dont have? Sorry not trolling | I started with $20, and her death was very sudden. Two incomes down to one is harsh, and I needed to cash out. | 2 | 1 |
I have Xbox game pass and I’m still boredh | Any recommendations | You are what doctors call “fucking lazy” | Thank you for your input | 2 | 2 |
Women can’t have anything nice on Reddit how come when a guy does literally the bare minimum for a female individual they’re called a “simp” | I post in aother sub about how Deanna Price became the second woman ever too throw over 80 meters in the hammer throw. There are so many comments about her gender and appearance, and all I was trying to do was celebrate this woman’s hard earned athletic achievement. Two weeks ago, someone else posted in the same sub about Ryan Crouser setting a new shot put word record, and no one said shit about his appearance I hate it here sometimes | What? Get the fuck out of here. I literally just looked at your post that you are talking about and like 95% of all the comments are positive. | I never said they were all negative, there’s a lot of nice comments, but there are a lot of really bad ones too. | 1 | 3 |
HS cheerleader carwash fundraisers are creepy | I had to explain to my 16 year old daughter, why her in a bikini washing cars for a public fundraiser was not gonna be in our plans. Come to find out today after school my daughter was glad she didn’t go as in the police were called on someone that was stalking the girls from across the street and making gestures for about 40 minutes. The whole concept is weird why even expose the girls like that in the first place | It’s fine. What’s with this recent trend of making everything into the worst possible conclusion it could possibly be? It’s just a bunch of dads checking out the young budding daughters of the other dads in the town. You know just like “oh hey man, your daughter has grown up so well, she’s almost a woman now, has she had her first blood?” just dad stuff. | I hope you never have kids, you fucking weirdo. If someone ever said that to me I would punch their lights out. If you have a daughter I dare you to show this comment to her or try to explain your thinking behind it, and watch them physically gag at your creepy ass | 1 | 4 |
Title | Post | Troll comment | Response | TL | RL |
i am now actually afraid of solo queuing | valorant solo q is actually scary to me now. i’ve gone up against way too many smurfs and dealt with so much toxicity its actually scary to me now. ranked is something i actually want to grind, but all the problems with the game is actually worrying me. does anyone else have this problem or am i just being fucking stupid? | Man you all sounds so soft lmao | Imagine a world where gaming is only filled with little toxic shts who think it’s cool to just shoot out insults just because it’s the internet. Making the community better only benefits us. This isn’t the past anymore. | 2 | 5 |
How do I play against mages? | I always feel like I am being outranged and can never get any significant damage down on them except for W1. Furthermore, I am always pushing the lane with W, but I am not sure how to trade without it. How should I be looking to space against mages? | You don’t. Just dodge the game. | dodge every mage im the game ?XDDD | 2 | 6 |
I hate fireworks | Why are they even legal? It scares the shit out our pets and our neighbors. It leaves all kinds of crap on the ground and smoke is horrid. How could they legalize it??? | JFC a bunch of snowflakes here on r/Ohio. Go live in your basement by all means. Especially if you can’t comprehend basic English while living in Ohio. Your grammar and spelling sucks. | I see you sub to r/republican, r/conservative, as well as r/AskAnAmerican Are you then a proud patriot? | 1 | 7 |
8. Bibliographical References
References
- Atanasov et al., 2019 Atanasov, A., De Francisci Morales, G., and Nakov, P. (2019). Predicting the role of political trolls in social media. In Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL), pages 1023–1034.
- Australia Institute, 2019 Australia Institute. (2019). Trolls and polls–the economic costs of online harassment and cyberhate.
- Banerjee and Lavie, 2005 Banerjee, S. and Lavie, A. (2005). Meteor: An automatic metric for mt evaluation with improved correlation with human judgments. In Proceedings of the acl workshop on intrinsic and extrinsic evaluation measures for machine translation and/or summarization, pages 65–72.
- Baumgartner et al., 2020 Baumgartner, J., Zannettou, S., Keegan, B., Squire, M., and Blackburn, J. (2020). The pushshift reddit dataset. In Proceedings of the International AAAI Conference on Web and Social Media, volume 14, pages 830–839.
- Benesch et al., 2016 Benesch, S., Ruths, D., Dillon, K. P., Saleem, H. M., and Wright, L. (2016). Counterspeech on twitter: A field study. In Dangerous Speech Project. Available at: https://dangerousspeech.org/counterspeech-ontwitter-a-field- study/.
- Binns, 2012 Binns, A. (2012). Don’t feed the trolls! Journalism Practice, 6:547–562, 08.
- Budzianowski and Vulić, 2019 Budzianowski, P. and Vulić, I. (2019). Hello, it’s GPT-2 - how can I help you? towards the use of pretrained language models for task-oriented dialogue systems. In Proceedings of the 3rd Workshop on Neural Generation and Translation.
- Choi et al., 2015 Choi, D., Han, J., Chung, T., Ahn, Y.-Y., Chun, B.-G., and Kwon, T. T. (2015). Characterizing conversation patterns in reddit: From the perspectives of content properties and user participation behaviors. In Proceedings of the 2015 ACM on Conference on Online Social Networks, page 233–243.
- Chun et al., 2019 Chun, S., Holowczak, R., Dharan, K., Wang, R., Basu, S., and Geller, J. (2019). Detecting political bias trolls in twitter data. In Proceedings of the 15th International Conference on Web Information Systems and Technologies, page 334–342.
- Chung et al., 2019 Chung, Y.-L., Kuzmenko, E., Tekiroglu, S. S., and Guerini, M. (2019). CONAN - COunter NArratives through nichesourcing: a multilingual dataset of responses to fight online hate speech. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 2819–2829.
- Chung et al., 2020 Chung, Y.-L., Tekiroglu, S. S., and Guerini, M. (2020). Italian counter narrative generation to fight online hate speech. In CLiC-it.
- Devlin et al., 2019 Devlin, J., Chang, M.-W., Lee, K., and Toutanova, K. (2019). BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies.
- Fichman and Sanfilippo, 2016 Fichman, P. and Sanfilippo, M. (2016). Online Trolling and Its Perpetrators: Under the Cyberbridge. Rowman & Littlefield Publishers.
- Gehman et al., 2020 Gehman, S., Gururangan, S., Sap, M., Choi, Y., and Smith, N. A. (2020). Realtoxicityprompts: Evaluating neural toxic degeneration in language models. In Trevor Cohn, et al., editors, Findings of the Association for Computational Linguistics: EMNLP 2020, Online Event, 16-20 November 2020, volume EMNLP 2020, pages 3356–3369.
- Georgakopoulos et al., 2018 Georgakopoulos, S. V., Tasoulis, S. K., Vrahatis, A. G., and Plagianakos, V. P. (2018). Convolutional neural networks for toxic comment classification. In Proceedings of the 10th hellenic conference on artificial intelligence, pages 1–6.
- Golf-Papez and Veer, 2017 Golf-Papez, M. and Veer, E. (2017). Don’t feed the trolling: rethinking how online trolling is being defined and combated. Journal of Marketing Management, 33(15-16):1336–1354.
- Gururangan et al., 2020 Gururangan, S., Marasović, A., Swayamdipta, S., Lo, K., Beltagy, I., Downey, D., and Smith, N. A. (2020). Don’t stop pretraining: Adapt language models to domains and tasks. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 8342–8360.
- Hardaker, 2013 Hardaker, C. (2013). “uh…. not to be nitpicky, but… the past tense of drag is dragged, not drug.”: An overview of trolling strategies. Journal of Language Aggression and Conflict, 1(1):58–86.
- Hardaker, 2015 Hardaker, C. (2015). “i refuse to respond to this obvious troll”:an overview of responses to (perceived) trolling. Corpora, 10:201–229.
- Herring, 2002 Herring, S. (2002). Computer-mediated communication on the internet. Annual Review of Information Science and Technology, 36:109–168, 01.
- Keskar et al., 2019 Keskar, N. S., McCann, B., Varshney, L. R., Xiong, C., and Socher, R. (2019). Ctrl: A conditional transformer language model for controllable generation. arXiv preprint arXiv:1909.05858.
- Klein et al., 2017 Klein, G., Kim, Y., Deng, Y., Senellart, J., and Rush, A. (2017). OpenNMT: Open-source toolkit for neural machine translation. In Proceedings of ACL 2017, System Demonstrations, pages 67–72.
- Kumar et al., 2014 Kumar, S., Spezzano, F., and Subrahmanian, V. (2014). Accurately detecting trolls in slashdot zoo via decluttering. In 2014 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM 2014), pages 188–195.
- Laugier et al., 2021 Laugier, L., Pavlopoulos, J., Sorensen, J., and Dixon, L. (2021). Civil rephrases of toxic texts with self-supervised transformers. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, pages 1442–1461, Online.
- Lewis et al., 2020 Lewis, M., Liu, Y., Goyal, N., Ghazvininejad, M., Mohamed, A., Levy, O., Stoyanov, V., and Zettlemoyer, L. (2020). BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.
- Lin, 2004 Lin, C.-Y. (2004). ROUGE: A package for automatic evaluation of summaries. In Text Summarization Branches Out.
- Liu et al., 2019 Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer, L., and Stoyanov, V. (2019). Roberta: A robustly optimized BERT pretraining approach. arXiv:1907.11692.
- Mathew et al., 2018 Mathew, B., Kumar, N., Goyal, P., Mukherjee, A., et al. (2018). Analyzing the hate and counter speech accounts on twitter. arXiv preprint arXiv:1812.02712.
- Mathew et al., 2019 Mathew, B., Saha, P., Tharad, H., Rajgaria, S., Singhania, P., Maity, S. K., Goyal, P., and Mukherjee, A. (2019). Thou shalt not hate: Countering online hate speech. In Proceedings of the international AAAI conference on web and social media, volume 13, pages 369–380.
- McHugh, 2012 McHugh, M. L. (2012). Interrater reliability: the kappa statistic. Biochemia medica: Biochemia medica, 22(3):276–282.
- Mihaylov and Nakov, 2016 Mihaylov, T. and Nakov, P. (2016). Hunting for troll comments in news community forums. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 399–405.
- Papineni et al., 2002 Papineni, K., Roukos, S., Ward, T., and Zhu, W.-J. (2002). Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics, pages 311–318.
- Paulus et al., 2017 Paulus, R., Xiong, C., and Socher, R. (2017). A deep reinforced model for abstractive summarization.
- Qian et al., 2019 Qian, J., Bethke, A., Liu, Y., Belding, E., and Wang, W. Y. (2019). A benchmark dataset for learning to intervene in online hate speech. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 4755–4764.
- Radford et al., 2019 Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., and Sutskever, I. (2019). Language models are unsupervised multitask learners. OpenAI Blog, 1(8), 2019.
- Richards and Calvert, 2000 Richards, R. D. and Calvert, C. (2000). Counterspeech 2000: A new look at the old remedy for bad speech. BYU L. Rev., page 553.
- Shin, 2008 Shin, J. (2008). Morality and internet behavior: A study of the internet troll and its relation with morality on the internet. In Proceedings of Society for Information Technology & Teacher Education International Conference 2008, pages 2834–2840.
- Sibai et al., 2015 Sibai, O., De Valck, K., Farrell, A. M., and Rudd, J. M. (2015). Social control in online communities of consumption: A framework for community management. Psychology & Marketing, 32(3):250–264.
- Tekiroğlu et al., 2020 Tekiroğlu, S. S., Chung, Y.-L., and Guerini, M. (2020). Generating counter narratives against online hate speech: Data and strategies. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 1177–1190, Online.
- Wolf et al., 2019 Wolf, T., Debut, L., Sanh, V., Chaumond, J., Delangue, C., Moi, A., Cistac, P., Rault, T., Louf, R., Funtowicz, M., et al. (2019). HuggingFace’s Transformers: State-of-the-art Natural Language Processing. ArXiv, pages arXiv–1910.
- Wulczyn et al., 2017 Wulczyn, E., Thain, N., and Dixon, L. (2017). Ex machina: Personal attacks seen at scale. In Proceedings of the 26th international conference on world wide web, pages 1391–1399.
- Zhang et al., 2020 Zhang, T., Kishore, V., Wu, F., Weinberger, K. Q., and Artzi, Y. (2020). Bertscore: Evaluating text generation with bert. In International Conference on Learning Representations.
- Zhu and Bhat, 2021 Zhu, W. and Bhat, S. (2021). Generate, prune, select: A pipeline for counterspeech generation against online hate speech. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pages 134–149, Online.
- Ziems et al., 2020 Ziems, C., He, B., Soni, S., and Kumar, S. (2020). Racism is a virus: Anti-asian hate and counterhate in social media during the covid-19 crisis. arXiv preprint arXiv:2005.12423.