∎
University of Adelaide /
Defence Science and Technology Group
Adelaide, South Australia, Australia
ORCID ID: 0000-0003-3830-9014
22email: derek.weber@{adelaide.edu.au,dst.defence.gov.au} 33institutetext: F. Neumann 44institutetext: School of Computer Science,
University of Adelaide
Adelaide, South Australia, Australia
ORCID ID: 0000-0002-2721-3618
44email: [email protected]
A General Method to Find Highly Coordinating Communities in Social Media through Inferred Interaction Links
Abstract
Political misinformation, astroturfing and organised trolling are online malicious behaviours with significant real-world effects. Many previous approaches examining these phenomena have focused on broad campaigns rather than the small groups responsible for instigating or sustaining them. To reveal latent (i.e., hidden) networks of cooperating accounts, we propose a novel temporal window approach that relies on account interactions and metadata alone. It detects groups of accounts engaging in various behaviours that, in concert, come to execute different goal-based strategies, a number of which we describe. The approach relies upon a pipeline that extracts relevant elements from social media posts, infers connections between accounts based on criteria matching the coordination strategies to build an undirected weighted network of accounts, which is then mined for communities exhibiting high levels of evidence of coordination using a novel community extraction method. We address the temporal aspect of the data by using a windowing mechanism, which may be suitable for near real-time application. We further highlight consistent coordination with a sliding frame across multiple windows and application of a decay factor. Our approach is compared with other recent similar processing approaches and community detection methods and is validated against two relevant datasets with ground truth data, using content, temporal, and network analyses, as well as with the design, training and application of three one-class classifiers built using the ground truth; its utility is furthermore demonstrated in two case studies of contentious online discussions.
Keywords:
Coordinated Behaviour Online Social Networks Influence Operations1 Introduction
Online social networks (OSNs) have established themselves as flexible and accessible systems for activity coordination and information dissemination. This benefit was illustrated during the Arab Spring (Carvin, 2012) but inherent dangers are increasingly apparent in ongoing political interference and disinformation (Bessi and Ferrara, 2016; Howard and Kollanyi, 2016; Ferrara, 2017; Keller et al., 2017; Neudert, 2018; Morstatter et al., 2018). Modern information campaigns are participatory activities, which aim to use their audiences to amplify their desired narratives, not just receive it (Starbird et al., 2019). Through cyclical reporting (i.e., social media feeding stories and narratives to traditional news media, which then sparks more social media activity), social media users can unknowingly become “unwitting agents” as “sincere activists” of state-based operations (Starbird and Wilson, 2020). The use of political bots to influence the framing and discussion of issues in the mainstream media (MSM) remains prevalent (Bessi and Ferrara, 2016; Woolley, 2016; Woolley and Guilbeault, 2018; Rizoiu et al., 2018). The use of bots to amplify individual voices above the crowd, known as the megaphone effect, requires coordinated action and a degree of regularity that may leave traces in the digital record.
Relevant research has focused on high level analyses of campaign detection and classification (Lee et al., 2013; Varol et al., 2017; Alizadeh et al., 2020), the identification of botnets and other dissemination groups (Vo et al., 2017; Gupta et al., 2019; Woolley and Guilbeault, 2018), and coordination at the community level (Kumar et al., 2018; Hine et al., 2017; Cresci, 2020). Some have considered generalised approaches to social media analytics (e.g., Lim et al., 2019; Weber, 2019; Pacheco et al., 2020b; Graham et al., 2020a; Nizzoli et al., 2020), but unanswered questions regarding the clarification of coordination strategies remain.
Expanding on our work presented at ASONAM’20 (Weber and Neumann, 2020), we present a novel approach to detect groups engaging in potentially coordinated activities, revealed through anomalously high levels of coincidental behaviour. Links in the groups are inferred from behaviours that, when used intentionally, are used to execute a number of identifiable coordination strategies. We use a range of techniques to validate our new technique on two relevant datasets, as well as comparison with ground truth and a synthesized dataset, and show it successfully identifies coordinating communities.
Our approach infers ties between accounts to construct latent coordination networks (LCNs) of accounts, using criteria specific to different coordination strategies. These are sometimes referred to as user similarity networks (e.g., Nizzoli et al., 2020). The accounts may not be directly connected, thus we use the term ‘latent’ to mean ‘hidden’ when describing these connections. The inference of connections is performed solely on the accounts’ activity, i.e., not their content, only metadata and temporal information, though it could be expanded to make use of these, similar to Pacheco et al. (2020b) and Graham et al. (2020a). Highly coordinating communities (HCCs) are then detected and extracted from the LCN. We propose a variant of focal structures analysis (FSA, Şen et al., 2016) to do this, in order to take advantage of FSA’s focus on finding influential sets of nodes in a network, while also reducing the computational complexity of the algorithm. A window-based approach is used to enforce temporal constraints.
The following research questions guided our evaluation:
- RQ1
-
How can HCCs be found in an LCN?
- RQ2
-
How do the discovered communities differ?
- RQ3
-
Are the HCCs internally or externally focused?
- RQ4
-
How consistent is the HCC messaging?
- RQ5
-
What evidence is there of consistent coordination?
- RQ6
-
How well can HCCs in one dataset inform the discovery of HCCs in another?
This paper expands upon Weber and Neumann (2020) by providing further methodological detail and experimental validation, and case studies in which the technique is applied to new real world datasets relating to contentious political issues, as well as consideration of algorithmic complexity and comparison with several similar techniques. Prominent among the extra validation provided is the use of machine learning classifiers to show that our datasets contain similar coordination to our ground truth, and the application of a sliding frame across the time windows as a way to search for consistent coordination.
This paper provides an overview of relevant literature, followed by a discussion of online coordination strategies and their execution. Our approach is then explained, and its experimental validation is presented111See https://github.com/weberdc/find_hccs for code and data.. Following the validation, the algorithmic complexity and performance of the technique is presented, two case studies are explored, demonstrating the utility of the approach in two real-world politically-relevant datasets, and we compare our technique to those of Pacheco et al. (2020b), Graham et al. (2020a), Nizzoli et al. (2020) and Giglietto et al. (2020b).
1.1 A Motivating Example
In the aftermath of the 2020 US Presidential election, a data scientist noticed a pattern emerging on Twitter222Tweeted 2020-11-17: https://twitter.com/conspirator0/status/1328479128908132358. Figure 1a shows a tweet by someone who was so upset with their partner voting for Joe Biden in the election that they decided to divorce them immediately and move to Pakistan (in the midst of the COVID-19 pandemic). This might seem an extreme reaction, but the interesting thing was that the person was not alone. The researcher had identified dozens of similar, but not always identical, tweets by people leaving for other cities but for the same reason (Figure 1b). Analysis of these accounts also revealed they were not automated accounts. This kind of pattern of tweeting identical text is sometimes referred to as “copypasta” and can be used to give the appearance of a genuine grassroots movement on a particular issue. It had been previously used by ISIS terrorists as they approached the city of Mosul, Iraq, in 2014, which they took over and occupied for several years when the local forces believed a giant army was on its way based on the level of relevant online activity (Brooking and Singer, 2016).
It is unclear whether this “copypasta” campaign is a deliberate information operation, designed to damage trust in the electoral system and ability of Americans to accept the loss of a preferred political party in elections, or simply a group of like-minded jokers starting a viral gag or engaging in a kind of flashmob. At the very least, it is important to be able to identify which accounts are participating in the event, and how they are coordinating their actions.


1.2 Online Information Campaigns and Related Work
Social media has been increasingly used for communication in recent years (particularly political communication), and so the market has followed, with media organisations using it for cheap, wide dissemination and consumers increasingly looking to it for news (Shearer and Grieco, 2019). Over that same time period, people have begun to explore how to exploit the features of the internet and social media that bring us benefits: the ability to target marketing to specific audiences that connects businesses with the most receptive customers also enables highly targeted non-transparent political advertising (Woolley and Guilbeault, 2018) and the ability to expose people to propaganda and recruit them to extremist organisations (Badawy and Ferrara, 2018); the anonymity that supports the voiceless in society to express themselves also enables trolls to attack others without repercussions (Hine et al., 2017; Burgess and Matamoros-Fernández, 2016); and the automation that enables news aggregators also facilitates social and political bots (Ferrara et al., 2016; Woolley, 2016; Cresci, 2020). In summary, targeted marketing and automation coupled with anonymity provide the tools required for potentially significant influence in the online sphere, perhaps enough to swing an election.
Effective influence campaigns relying on these capabilities will somehow coordinate the actions of their participants. Early work on the concept of coordination by Malone and Crowston (1994) described it as the dependencies between the tasks and resources required to achieve a goal. One task may require the output of another task to complete. Two tasks may share, and require exclusive access to, a resource or they may both need to use the resource simultaneously.
At the other end of the spectrum, sociological studies of influence campaigns can reveal their intent and how they are conducted, but they consider coordination at a much higher level. Starbird et al. (2019) highlight three kinds of campaigns: orchestrated, centrally controlled campaigns that are run from the top down (e.g., paid teams, Chen, 2015; King et al., 2017); cultivated campaigns that infiltrate existing issue-based movements to drive them to particular extreme positions; and emergent campaigns arising from enthusiastic communities centred around a shared ideology (e.g., conspiracy groups and other fringe movements). Though their strategies differ, they use the same online interactions as normal users, but their patterns differ.
Disinformation campaigns effectively trigger human cognitive heuristics, such as individual and social biases to believe what we hear first (anchoring) and what we hear frequently and can remember easily (availability cascades) (Tversky and Kahneman, 1973; Kuran and Sunstein, 1999); thus the damage is already done by the time lies are exposed. Recent experiences of false information moving beyond social media during Australia’s 2019-2020 bushfires highlight that identifying these campaigns as they occur can aid OSN monitors and the media to better inform the public (Graham and Keller, 2020).
In between task level coordination and entire campaigns, at the level of social media interactions, as demonstrated by Graham and Keller (2020), we can directly observe the online actions and effects of such campaigns, and infer links between accounts based on pre-determined criteria. Relevant efforts in computer science has focused on detecting information operations on social media via automation (Ferrara et al., 2016; Grimme et al., 2017; Davis et al., 2016; Cresci, 2020), campaign detection via content (Lee et al., 2013; Assenmacher et al., 2020; Alizadeh et al., 2020), URLs (Ratkiewicz et al., 2011; Cao et al., 2015; Giglietto et al., 2020b) and hashtags (Ratkiewicz et al., 2011; Burgess and Matamoros-Fernández, 2016; Varol et al., 2017; Weber et al., 2020), temporal patterns (Chavoshi et al., 2017; Hine et al., 2017; Nasim et al., 2018; Mazza et al., 2019; Pacheco et al., 2020a), and community detection (Vo et al., 2017; Morstatter et al., 2018; Gupta et al., 2019). Other studies have explored how bots and humans interact in political settings (Bessi and Ferrara, 2016; Rizoiu et al., 2018; Woolley and Guilbeault, 2018), including exploring how deeply embedded bots appear in the network and their degree of organisation (Woolley and Guilbeault, 2018). There is, however, a research gap: the computer science study of the “orchestrated activities” of accounts in general, regardless of their degree of automation (Grimme et al., 2018; Alizadeh et al., 2020; Nizzoli et al., 2020; Vargas et al., 2020). It must be noted that bot activity, even coordinated activity, may be entirely benign and even useful (Ferrara et al., 2016; Graham and Ackland, 2017).
Though some studies have observed the existence of strategic behaviour in and between online groups (e.g., Keller et al., 2017; Kumar et al., 2018; Hine et al., 2017; Keller et al., 2019; Giglietto et al., 2020b), the challenge of identifying a broad range of strategies and their underpinning execution methods remains relatively unstudied.
Inferring social networks from OSN data requires attendance to the temporal aspect to understand information (and influence) flow and degrees of activity (Holme and Saramäki, 2012). Real time processing of OSN posts can enable tracking narratives via text clusters (Assenmacher et al., 2020), but to process networks requires graph streams (McGregor, 2014) or window-based pipelines (e.g., Weber, 2019), otherwise processing is limited to post-collection activities (Graham et al., 2020a; Pacheco et al., 2020b).
This work contributes to the identification of strategic coordination behaviours, along with a general technique to enable detection of groups using them.
2 Coordination Strategies
Online influence relies on two primary mechanisms: dissemination and engagement. For example, an investigation of social media activity following UK terrorist attacks in 2017333https://crestresearch.ac.uk/resources/russian-influence-uk-terrorist-attacks/ identified accounts promulgating contradictory narratives, inflaming racial tensions and simultaneously promoting tolerance to sow division. By engaging aggressively, the accounts drew in participants who then spread the message.
Dissemination aims to maximise audience, to convince through repeated exposure and, in the case of malicious use, to cause outrage, polarisation and confusion, or at least attract attention to distract from other content.
Engagement is a subset of dissemination that solicits a response. It relies on targeting individuals or communities through mentions, replies and the use of hashtags as well as rhetorical approaches that invite responses (e.g., inflammatory comments or, as present in the UK terrorist example above, pleas to highly popular accounts).

A number of online coordination strategies have been observed in the literature making use of both dissemination and engagement, including:
- 1.
- 2.
- 3.
- 4.
OSN | POST | REPOST | REPLY | MENTION | TAG | LIKE |
---|---|---|---|---|---|---|
tweet | retweet | reply tweet | @mention | #hashtags | favourite | |
post | share | comment | mention | #hashtag | reactions | |
Tumblr | post | repost | comment | @mention | #tag | heart |
post | crosspost | comment | /u/mention | subreddit | up/down vote |
Different behaviour primitives (e.g., Table 1) can be used to execute these strategies. Dissemination can be carried out by reposting, using hashtags, or mentioning highly connected individuals in the hope they spread a message further. Accounts doing this covertly will avoid direct connections, and thus inference is required for identification. Giglietto et al. (2020b) propose detecting anomalous levels of coincidental URL use as a way to do this; we expand this approach to other interactions.
Some strategies require more sophisticated detection: detecting bullying through dogpiling (e.g., as happened during the #GamerGate incident, studied by Burgess and Matamoros-Fernández (2016), or to those posing questions to public figures at political campaign rallies444https://www.bbc.co.uk/bbcthree/article/72686b6d-abd2-471b-ae1d-8426522b1a97) requires collection of (mostly) entire conversation trees, which, while trivial to obtain on forum-based sites (e.g., Facebook and Reddit), are difficult on stream-of-post sites (e.g., Twitter, Parler and Gab). Detecting metadata shuffling requires long term collection on broad issues to detect the same accounts being active in different contexts.
Figure 2 shows representations of several strategies, giving clues about how they might be identified. To detect Pollution, we match the authors of posts mentioning the same (hash)tag. This way we can reveal not just those who are using the same hashtags with significantly greater frequency than the average but also those who use more hashtags than is typical. To detect Boost, we match authors reposting the same original post, and can explore which sets of users not only repost more often than the average, but those who repost content from a relatively small pool of accounts. Alternatively, we can match authors who post identical, or near identical text, as seen in our motivating example (Section 1.1); Graham et al. (2020a) have recently developed open sourced methods for this kind of matching. Considering reposts like retweets, however, it is unclear whether platforms deprioritise them when responding to stream filter and search requests, so special consideration may be required when designing data collection plans. Finally, to detect Bully, we match authors whose replies are transitively rooted in the same original post, thus they are in the same conversation. This requires collection strategies that result in complete conversation trees, and also stipulates a somewhat strict definition of ‘conversation’. On forum-based OSNs, the edges of a ‘conversation’ may be relatively clear: by commenting on a post, one is ‘joining’ the ‘conversation’. Delineating smaller sets of interactions within all the comments on a post to find smaller conversations may be achieved by regarding each top-level comment and its replies as a conversation, but this may not be sufficient. Similarly, on stream-based OSNs, a conversation may be engaged in by a set of users if they all mention each other in their posts, as it is not possible to reply to more than one post at a time.
2.1 Problem Statement
A clarification of our challenge at this point is:
To identify groups of accounts whose behaviour, though typical in nature, is anomalous in degree.
There are two elements to this. The first is discovery. How can we identify not just behaviour that appears more than coincidental, but also the accounts responsible for it? That is the topic of the next section. The second element is validation. Once we identify a group of accounts via our method, what guarantee do we have that the group is a real, coordinating set of users? This is especially difficult given inauthentic behaviour is hard for humans to judge by eye (Cresci et al., 2017).
3 Methodology
The major OSNs share a number of features, primarily in how they permit users to interact. By focusing on these commonalities, it is possible to develop approaches that generalise across the OSNs that offer them.
Traditional social network analysis relies on long-standing relationships between actors (Wasserman and Faust, 1994; Borgatti et al., 2009). On OSNs these are typically friend/follower relations. These are expensive to collect and quickly degrade in meaning if not followed with frequent activity. By focusing on active interactions, it is possible to understand not just who is interacting with whom, but to what degree. This provides a basis for constructing (or inferring) social networks, acknowledging they may be transitory.
LCNs are built from inferred links between accounts. Supporting criteria relying on interactions alone, as observed in the literature (Ratkiewicz et al., 2011; Keller et al., 2019), include retweeting the same tweet (co-retweet), using the same hashtags (co-hashtag) or URLs (co-URL), or mentioning the same accounts (co-mention). To these we add joining the same ‘conversation’ (a tree of reply chains with a common root tweet) (co-conv). As mentioned earlier, other ways to link accounts rely on similar or identical content (e.g., Lee et al., 2013; Keller et al., 2019; Pacheco et al., 2020a; Graham et al., 2020a) or metadata (Mariconti et al., 2017; Ferrara, 2017).
3.1 The LCN / HCC Pipeline
The key steps to extract HCCs from raw social media data are shown in Figure 3 and documented in Algorithm 1. The example in Figure 3 is explained after the algorithm has been explained, in Section 3.1.2.

Input: : Social media posts, : Coordination criteria, : Extraction parameter
Output: : A list of HCCs
Step 1.
Convert social media posts to common interaction primitives, . This step removes extraneous data and provides an opportunity for the fusion of sources by standardising all interactions (thus including only the elements required for the coordination being sought).
Step 2.
From , filter the interactions, , relevant to the set = of criteria (e.g., co-mentions and co-hashtags).
Step 3.
Infer links between accounts given , ensuring links are typed by criterion. The result, , is a collection of inferred pairings. The count of inferred links between accounts and due to criterion is .
Step 4.
4 Construct an LCN, , from the pairings in . This network = is a set of vertices representing accounts connected by undirected weighted edges of inferred links. These edges represent evidence of different criteria linking the adjacent vertices. The weight of each edge between vertices representing accounts and for each criterion is , and is equal to .
Most community detection algorithms will require the multi-edges be collapsed to single edges. The edge weights are incomparable (e.g., retweeting the same tweet is not equivalent to using the same hashtag), however, for practical purposes, the inferred links can be collapsed and the weights combined for cluster detection using a simple summation, e.g., Equation (1), or a more complex process like varied criteria weighting.
(1) |
Some criteria may result in highly connected LCNs, even if its members never interact directly. Not all types of coordination will be meaningful – people will co-use the same hashtag repeatedly if that hashtag defines the topic of the discussion (e.g., #auspol for Australian politics), in which case it is those accounts who co-use it significantly more often than others which are of interest. If required, the final step filters out these coincidental connections.
Step 5.
Identify the highest coordinating communities, , in (Figure 3e), using a suitable community detection algorithm, such as Blondel et al. (2008)’s Louvain algorithm (used by Morstatter et al., 2018; Nasim et al., 2018), k nearest neighbour () (used by Cao et al., 2015), edge weight thresholding (used by Lee et al., 2013; Pacheco et al., 2020b), or FSA (Şen et al., 2016), an algorithm from the Social Network Analysis community that focuses on extracting sets of highly influential nodes from a network. Depending on the size of the dataset under consideration, algorithms suitable for very large networks may need to be considered Fang et al. (2019). Some algorithms may not require the LCN’s multi-edges to be merged (e.g., Bacco et al., 2017). We present a variant of FSA (Şen et al., 2016), FSA_V (Algorithm 2), because it is designed to take advantage of FSA’s benefits, while addressing some of its costs. FSA does not just divide a network into communities (so that every node belongs to a community), but extracts only sets of nodes that form influential communities within the overall network. FSA_V reduces the computational complexity introduced by FSA, which recursively applies Louvain to divide the network into smaller components and then, under certain circumstances, stitches them back together. The reason for this is to make it more suitable for application to a streaming scenario, in which execution speed is a priority.
Similar to FSA, FSA_V initially divides into communities using the Louvain algorithm but then builds candidate HCCs within each, starting with the ‘heaviest’ (i.e., highest weight) edge (representing the most evidence of coordination). It then attaches the next heaviest edge until the candidate’s mean edge weight (MEW) is no less than () of the previous candidate’s MEW, or is less than ’s overall MEW. In testing, edge weights appeared to follow a power law, so was introduced to identify the point at which the edge weight drops significantly; requires tuning. A final filter ensures no HCC with a MEW less than ’s is returned. Unlike in FSA (Şen et al., 2016), recursion is not used, nor stitching of candidates, resulting in a simpler algorithm.
Input: =: An LCN, : HCC threshold
Output: : Highly coordinating communities
This algorithm prioritises edge weights while maintaining an awareness of the network topology by examining adjacent edges, something ignored by simple edge weight filtering. Our goal is to find sets of strongly coordinating users, so it is appropriate to prioritise strongly tied communities while still acknowledging coordination can also be achieved with weak ties (e.g., accounts paid to retweet a single tweet).
The complexity of the entire pipeline is low order polynomial due primarily to the pairwise comparison of accounts to infer links in Step , which can be constrained by window size when addressing the temporal aspect. For large networks (meaning networks with many accounts), that may be too costly to be of practical use; the solution for this relies on the application domain inasmuch as it either requires a tighter temporal constraint (i.e., a smaller time window) or tighter stream filter criteria, causing a reduction in the number of accounts, potentially along with a reduction in posts. The complexity of the pipeline and HCC extraction is discussed in more detail in Section 3.3.
3.1.1 Addressing the Temporal Aspect
Temporal information is a key element of coordination, and thus is critical for effective coordination detection. Frequent posts within a short period may represent genuine discussion or deliberate attempts to game trend algorithms (Grimme et al., 2018; Varol et al., 2017; Assenmacher et al., 2020). We treat the post stream as a series of discrete windows to constrain detection periods. An LCN is constructed from each window (Step 4), and these are then aggregated and mined for HCCs (Step 5). As we assume posts arrive in order, their timestamp metadata can be used to sort and assign them to windows.
3.1.2 A Brief Example
Figure 3 gives an example of searching for co-hashtag coordination across Facebook, Twitter, and Tumblr posts. The posts are converted to interaction primitives in Step 1, shown in Figure 3a. The information required from each post is the identity of the post’s author555Linking identities across social media platforms is beyond the scope of this work, but the interested reader is referred to Adjali et al. (2020) for a recent contribution to the subject., the timestamp of the post for addressing the temporal aspect, and the hashtag mentioned (there may be many, resulting in separate records for each). This is done in Figure 3b, which shows the filtered mentions (in orange) and hashtag uses (in purple), ordered according to timestamp.
Step 3 in Figure 3c involves searching for evidence of coordination through searching for our target coordination strategies through pairwise examination of accounts and their interactions. Here, three accounts co-use a hashtag while only two of them co-mention another account.
By Step 4 in Figure 3d, the entire LCN has been constructed, and then Figure 3e shows its most highly coordinating communities.
As mentioned above, to account for the temporal aspect, the LCNs produced for each time window in Figure 3d can be aggregated and then mined for HCCs, or HCCs could be extracted from each window’s LCN and then they can be aggregated, or analysed in near real-time, as dictated by the application domain.
3.2 Validation Methods
As mentioned in Section 2.1, the second element of addressing our research challenge is that of validation. Once HCCs have been discovered, it is necessary to confirm that what has been found are examples of genuine coordinating groups. This step is required before the further question of whether the coordination is authentic (e.g., grassroots activism) or inauthentic (e.g., astroturfing).
3.2.1 Datasets
In addition to relevant datasets, we make use of a ground truth (GT), in which we expect to find coordination (cf., Keller et al., 2017; Vargas et al., 2020). By comparing the evidence of coordination (i.e., HCCs) we find within the ground truth with the coordination we find in the other datasets, we can develop confidence that: a) our method finds coordination where we expect to find it (in the ground truth); and b) our method also finds coordination of a similar type where it was not certain to exist. Furthermore, to represent the broader population (which is not expected to exhibit coordination), similar to Cao et al. (2015), we create a randomised HCC network from the non-HCC accounts in a given dataset, and then compare its HCCs with the HCCs that had been discovered by our method.
3.2.2 Membership Comparison Across Variables
Our primary factors include the HCC extraction method (using FSA_V, , or thresholds), the window size, , and the strategy being targeted (Boost, Pollution or Bully). Given our interest prioritises the sets of accounts over how they are connected, for each pair of variations we compare the membership of the HCCs discovered, as a whole, and the number of HCCs discovered, as well as the edge count. Also included are the exact numbers of HCC members common to each pair. These figures provide further context for the degree of overlap between the HCC members identified under different conditions (i.e., factor values). We use Jaccard and overlap similarity measures (Verma and Aggarwal, 2020) to compare the accounts appearing in each (ignoring their groupings) and render them as heatmaps. The Jaccard similarity coefficient, , is a value between and which represents the similarity between two sets of items, and :
(2) |
If there is significant imbalance in the sizes of and , then their similarity may be low, even if one is a subset of the other. An alternative measure, the Overlap or Szymkiewicz-Simpson coefficient (Verma and Aggarwal, 2020), takes this imbalance into account by using the size of the smaller of the two sets as the denominator:
(3) |
In a circumstance such as ours, it is unclear whether a longer time window will garner more results after HCC extraction is applied. The Jaccard and overlap coefficients can be used to quickly understand two facts about the sets of accounts identified as HCC members with different values of :
-
•
Is one set a subset of the other? If so, the overlap coefficient will reach , while the Jaccard coefficient may not, if the two sets differ in size. If they are disjoint, it will be , along with the Jaccard coefficient.
-
•
Do the sets differ in size? If the sets are different sizes, but one is a subset of the other, the overlap coefficient will hide this fact, while the Jaccard coefficient will expose it. If both coefficients have values close to , then the sets are clearly different in membership and potentially also in size. If the coefficient values are very close, then the sets are close in size, because the denominators are similar in size, meaning , but this will only occur if they share many members (i.e., is high).
The exact numbers of common accounts complement the degree of overlap visible in the heat maps to provide the reader a basis to easily compare how different variations compare with each other beyond examining them only as pairs and therefore illuminates the influence each particular factor has overall. For example, by being able to compare the results using each value for , it is possible to see the progression of results as the window size increases (both in raw numbers and with a colour scale in the heatmaps).
3.2.3 Network Visualisation
A second subjective method of analysis for networks is, of course, to visualise them. We use two visualisation tools, visone (https://visone.info) and Gephi (https://gephi.org), both of which make use of force directed layouts, which help to clarify clusters in the network structure. In particular, Gephi provides access to the Fruchterman-Reingold (FR) algorithm (Fruchterman and Reingold, 1991), which is force directed, but constrains the layout within a circle. This enables us to see to what degree individual clusters dominate the overall graph, in a way that standard force-directed layouts do not. Node colour is used to represent cluster membership. Clusters are identified with the Louvain method (Blondel et al., 2008). Node size can be used to represent degree or, for nodes representing accounts, the number of posts an account contributes to a corpus. Edge thickness and colour (darkness) are used to represent the weight of the edges. As HCCs are weighted undirected networks and each connected component is an HCC, node colour can be used to represent the number of posts, and edge weight can be represented by thickness and, depending on the density of the network, colour darkness. Node shape can be also employed to represent a variable; for analyses that involve multiple criteria (e.g., co-conv and co-mention), we use shape to represent which combination of criteria an HCC is bound by (e.g., just co-mention or a combination of co-mention and co-conv or just co-conv).
By extending the HCC account networks with nodes to represent the ‘reasons’ or instances of evidence that link each pair of nodes, e.g., the tweets they retweet in common, or accounts they both mention or reply to, creating a two-level account-reason network in doing so, we can investigate how HCCs relate to one another. In this case, the two-level network has two types of nodes (accounts and reasons) and two types of edges (‘coordinates with’ links between accounts and ‘caused by’666Or other appropriate phrasing to indicate that the account is linked to the reason, by which it is indirectly linked to other accounts. N.B. The reason itself may be an account, e.g., in the context of co-mentions, but for the purpose of the two-level network, the mentioned account is treated as a reason. links between ‘reasons’ and accounts). Visualising the two-level network by colouring nodes by their HCC and using a force-directed layout highlights how closely associated HCCs are with each other, not only revealing what reasons draw many HCCs together (i.e., many HCCs may be bound by a single reason, or an HCC may be entirely isolated from others in the broader community), but also how many reasons may bind them (i.e., many reasons may bind an HCC together). Deeper insights can be revealed from this point using multi-layer network analyses.
3.2.4 Consistency of Content
To help answer RQ2, it is helpful to look beyond network structures and consider how consistent the content within an HCC777I.e., the content produced by the HCC’s members. is relative to other HCCs and the population in general. This will be most applicable when the type of strategy the HCC is suspected to have engaged in requires repetition, e.g., co-retweeting or copypasta. If HCCs are boosting a message, it is reasonable to assume the content of HCCs members will be more similar internally that when compared externally, to the content of non-members. To analyse this internal consistency of content, we treat each HCC member’s tweets as a single document and create a doc-term matrix using character n-grams for terms. Comparing the members’ document vectors using cosine similarity in a pairwise fashion creates a matrix where is the number of accounts in the HCC network. This approach was chosen for its performance with non-English corpora (Damashek, 1995), and because using individual tweets as documents produced too sparse a matrix in a number of tests we conducted. The pairwise account similarity matrix can be visualised, using a range of colours to represent similarity. By ordering the accounts on both the and axes to ensure they are grouped within their HCCs, if our hypothesis is right that similarity within HCCs is higher than outside, then we should observe clear bright squares extending from the diagonal of the matrix. The diagonal itself will be the brightest because it represents each account’s similarity with itself.
If HCCs contribute few posts, which are similar or identical to other HCCs, then bright squares may appear off the diagonal, and this would be evidence similar to clusters of account nodes around a small number of reason nodes in the two-level account-reason networks mentioned in the previous section.
This method offers no indication of how active each HCC or HCC member is, so displays of high similarity may imply low levels of activity as well as high content similarity, just because of the lower likelihood that highly active accounts are highly similar in content (by contributing more posts, there are simply more opportunities for accounts’ content to diverge). The use of the 5-character n-gram approach is designed to offset this because each tweet in common between two accounts will yield a large number of points of similarity, as will the case when the same two tweets are posted in the same order (i.e., two accounts both post tweet and then ), because the overlap between the tweets will yield at least 4 points of similarity.
3.2.5 Variation of Content
Converse to the consistency of content within HCC is the question of content variation, and how does the variation observed in detected HCCs differ from that of RANDOM groupings. Highly coordinated behaviour such as co-retweeting involves reusing the same content frequently, resulting in low feature variation (e.g., hashtags, URLs, mentioned accounts), which can be measured as entropy (Cao et al., 2015). A frequency distribution of each HCC’s use of each feature type is used to calculate each entropy score. Low feature variation corresponds to low entropy values. As per Cao et al. (2015), we compare the entropy of features used by detected HCCs to RANDOM ones and visualise their cumulative frequency. Entries for HCCs which did not use a particular feature are omitted, as their scores would inflate the number of groups with entropy.
3.2.6 Hashtag Analysis
Hashtags can be used to define discussion groups or chat channels (Woolley, 2016), so hashtag analysis can be used to study those communities. It is another aspect to content analysis that relies upon social media users declaring the topic of their post through the hashtags they include. At the minimum, we can plot the frequency of the most frequently used hashtags as used by the most active HCCs. In doing so, we can quickly see which hashtags different HCCs make use of, and how they relate by how they overlap. Some hashtags will be unique to HCCs, while others will be used by many. This exposes the nature of HCC behaviour: they may focus their effort on a single hashtag, perhaps to get it trending, or they may use many hashtags together, perhaps to spread their message to different communities.
To further explore how hashtags are used together, we perform hashtag co-occurrence analysis, creating networks of hashtags linked when they are mentioned in the same tweet. These hashtag co-occurrence networks are sometimes referred to as semantic networks (Radicioni et al., 2020). When visualised it is possible to see themes in the groupings of hashtags, and to gain insights from how the theme clusters are connected (including when they are isolated from one another). To extend this approach, Louvain clustering (Blondel et al., 2008) can be applied and hashtag nodes can be coloured by cluster, providing a statistical measure to how related the hashtags are.
3.2.7 Temporal Patterns
Campaign types can exhibit different temporal patterns (Lee et al., 2013), so we use the same temporal averaging technique as Lee et al. (2013) (dynamic time warping barycenter averaging) to compare the daily activities of the HCCs in the GT and RANDOM datasets with those in the test datasets. The temporal averaging technique produces a single time series made from the time series of all activities of each member in a set of accounts. Using this technique avoids averaging out of time series that are off-phase from one another, by aligning them before averaging them.
Another aspect of temporal analysis is the comparison of HCCs detected in different time windows, including specifically observing whether any HCCs between time windows share members and what the implications are for the behaviour of those members. This is non-trivial for any moderately large dataset, but examination of the ground truth can provide some insight into the behaviours exhibited by accounts known to be coordinating.
3.2.8 Focus of Connectivity
Groups that retweet or mention themselves create direct connections between their members, meaning if one is discovered, it may be trivial to find its collaborators. To be covert, therefore, it would be sensible to have a low internal retweet and mention ratios (IRR and IMR, respectively). Formally, if and are the the sets of retweets and mentions of accounts within an HCC, respectively, and and are the corresponding sets of retweets and mentions of accounts outside the HCC, then, for a single HCC,
(4) |
and
(5) |
3.2.9 Consistency of Coordination
The method presented Section 3.1 highlights HCCs that coordinate their activity at a high level over an entire collection period. Further steps can be taken to determine which HCCs are coordinating their behaviour repeatedly and consistently across adjacent time windows. In this case, for each time window, we consider not just the nodes and edges from the current LCN, but additionally from previous windows, applying a degrading factor the contribution of their edge weights. To build an LCN from a sliding frame of time windows, the new LCN includes the union of the nodes and edges of the individual LCNs from the current and previous windows, but to calculate the edge weights, we apply a decay factor, , to the weights of edges appearing in windows before the current one. In this way, we apply a multiplier of to the edge weights, where is the number of windows into the past: the current window is windows into the past, so its edges are multiplied by ; the immediate previous window is window back, so its edge multiplier is ; the one before that uses , and so on until the farthest window back uses . Generalising from Step 4, the weight for an edge between accounts and for criterion at window and a sliding window windows wide is given by
(6) |
In this way, to create a baseline in which the sliding frame is only one window wide, one only need choose =, regardless of the value of . As , the contributions of previous windows’ are given more consideration.
3.2.10 Machine Learning with One-Class Classifiers
An approach that aids in the management of data with many features is classification through machine learning. This is an approach that has been used extensively in campaign detection, in which tweets are classified, rather than accounts (e.g., Lee et al., 2013; Chu et al., 2012; Wu et al., 2018). As our intent is to determine whether HCCs detected in datasets are similar to those detecting in ground truth, it is acceptable to rely on one-class classification (i.e., an HCC detected in a dataset is recognised as COORDINATING/positive, or NON-COORDINATING/unknown). The more common binary classification approach was used by Vargas et al. (2020), however our approach has two distinguishing features:
-
1.
We rely on one-class classification because we have positive examples of what we regard as COORDINATING from the ground truth, and everything else is regarded as unknown, rather than definitely ‘not coordinating’. A one-class classifier can, for example, suggest a new book from a wide range (such as a library) based on a person’s reading history. In such a circumstance, the classifier designer has access to positive examples (books the reader likes or has previously borrowed) but all other instances (books here) are either positive or negative. When our one-class classifier recognises HCC accounts as positive instances, it provides confidence that the HCC members are coordinating their behaviour in the same manner as the accounts in the ground truth.
-
2.
We rely on features from both the HCCs and the HCC members and use the HCC members as the instances for classification, given it is unclear how many members an HCC may have, and accounts that are members of HCCs may have traits in common that are distinct from ‘normal’ accounts. For this reason the feature vectors that our classifier is trained and tested on will comprise features drawn from the individual accounts and their behaviour as well as the behaviour of the HCC of which they are a member. Feature vectors for members of the same HCC will naturally share the feature values drawn from their grouping.
Regarding the construction of the feature vector, at a group level, we consider not just features from the HCC itself, which is a weighted undirected network of accounts, but of the activity network built from the interactions of the HCC members within the corpus. The activity network is a multi-network (i.e., supports multiple edges between nodes) with nodes and edges of different types. The three node types are account nodes, URL nodes, and hashtag nodes. Edges represent interactions and the following types are modelled: hashtag uses, URL uses, mentions, repost/retweets, quotes (cf. comments on a Facebook or Tumblr post), reply, and ‘in conversation’, meaning that one account replied to a post that was transitively connected via replies to an initial post by an account in the corpus. This activity network therefore represents not just the members of the HCC but also their degree of activity.
Classifier algorithms.
We use the GT to train three classifiers. Two were well-known classifiers drawn from the scikit-learn Python library (Pedregosa et al., 2011): Support Vector Machine (SVM) and Random Forest (RF). The SVM and RF classifiers were trained with two-fold validation and the RF was configured to use trees (estimators). The third classifier was a bagging PU classifier (BPU, Mordelet and Vert, 2014)888Thanks to Roy Wright for his implementation: https://github.com/roywright/pu_learning/blob/master/baggingPU.py also configured to use a RF classifier internally. Contrasting “unlabeled” training instances were created from the RANDOM dataset. Feature vector values were standardised prior to classification and upsampling was applied to create balanced training sets.
The classifiers predict whether instances provided to them are in the positive or unlabeled classes, which we refer to, to aid readability, as ‘COORDINATING’ and ‘NON-COORDINATING’, respectively.
Performance metrics.
The performance metrics used include the classifier’s accuracy, scores for each class, and the Precision and Recall measures that the scores are based upon. High precision implies the classifier is good at recognising samples correctly, and high Recall implies that a classifier does not miss instances of the class they are trained on in any testing data. For example, a good apple classifier will successfully recognise an apple when it is presented with one, and when presented with a bowl of fruit, it will successfully find all the apples in it. The score combines these two measures:
(7) |
and provides insight into to the balance between the classifiers precision and recall. The accuracy of a classifier is the proportion of instances in a test data set that the classifier labeled correctly. In this way, the accuracy is the most coarse of these measures, because it offers little understanding of whether the classifier is missing instances it should find (false negatives) or labeling non-matching instances incorrectly (false positives). The score begins to address this failing, but direct examination of the Precision and Recall provides the most insight into each classifier’s performance.
3.2.11 Bot Analysis
Although coordinated behaviour in online campaigns is often conducted without automation (Starbird et al., 2019), automation is still commonly present in campaigns, especially in the form of social bots, which aim to present themselves as typical human users (Ferrara et al., 2016; Cresci, 2020). For this reason, the technique presented here is a valid tool for exposing teams of cooperating bot and social bot accounts. We use the Botometer (Davis et al., 2016) service to evaluate selected accounts for bot-like behaviour, which provides a detailed assessment of an account’s botness based on over features in six categories. Each category is assigned a value from , meaning human, to , meaning automated. Although other studies have relied on as a threshold for labelling an account as a bot, there is a a significant overlapping area between humans that act in a very bot-like manner, and bots that are quite human-like, so we adopt the approach of Rizoiu et al. (2018) and regard scores below to be human and scores above to be bots.
3.3 Complexity Analysis
The steps in processing timeline presented in Section 3.1 are reliant on two primary factors: the size of the corpus of posts, , being processed, and the size of the set of accounts, , that posted them. Therefore Then the complexity of Step 1 is linear, , because it requires processing each post, one-by-one. The set of interactions, , it produces may be larger than , because a post may include many hashtags, mentions, or URLs, but given posts are not infinitely long (even long Facebook or Tumblr posts only include several thousand words), the number of interactions will also be linear, even if it , for some constant . Step 2 filters these interactions down to only the ones of interest, , based on the type of coordinated activity sought, , so , and again the complexity of this step is also linear, , as it requires each interaction to be considered. Step 3 seeks to find evidence of coordination between the accounts in the dataset, and so requires examining each filtered interaction and building up data structures to associated each account with their interactions (), then emitting pairs of accounts matching the coordination criteria, producing the set , which requires the pairwise processing of all accounts, and so is with a complexity of . This, however, also depends on the pairwise comparison of each account’s interactions, which is likely to be small, practically, but theoretically could be as large as if one user is responsible for every interaction in the corpus (but then would be ). On balance, as a result, we will regard the processing of each pair of users’ interactions as linear with a constant factor (i.e., = ). In Step 4, producing the LCN, , from the criteria is a matter of considering each match one-by-one, so is again linear (though potentially large, depending on ). The final step (5) is to extract the HCCs from the LCN, and its performance and complexity very much depend upon the algorithm employed, but significant research has been applied in this field (as considered in, e.g., Bedru et al., 2020). For FSA_V, which relies on the Louvain algorithm with complexity (Blondel et al., 2008), it considers edges within each community to build its HCC candidates, so has a complexity of less than , where is the number of edges in the LCN, meaning its complexity is linear. FSA_V’s complexity is therefore .
We regard the computation complexity of the entire pipeline as the highest complexity of its steps, which are:
-
1.
Extract interactions from posts:
-
2.
Filter interactions:
-
3.
Find evidence of coordination:
-
4.
Build LCN from the evidence:
-
5.
Extract HCCs from LCN using, e.g., FSA_V:
The maximum of these is Step 3, the search for evidence of coordination, . Though in theoretical terms the method is potentially highly costly, in practical terms, we are bound by the number of accounts in the collection (which is highly dependent on the manner in which the data was collected and the nature of online discussion it to which it pertains) and may be managed by constraining the time window, further reducing the number of posts (and therefore accounts) considered, as long as that suits the type of coordination being sought.
4 Evaluation
Our approach was evaluated in two phases:
-
•
The first was conducted as an experiment using the validation methods mentioned above and two datasets known to include coordinated behaviour, as well as a ground truth dataset.
-
•
The second phase involved two case studies applying the approach against datasets relating to politically contentious topics expected to include polarised groups.
The first stage of the evaluation involved searching for Boost by co-retweet and other strategies while varying window sizes (). FSA_V was compared against two other community detection algorithms when applied to the LCNs built in Step 4 (aggregated). We then validated the resulting HCCs through a variety of network, content, and temporal analyses and machine learning classification, guided by the research questions posed in Section 1. Discussion of further applications and performance metrics is also presented.
Tweets | Retweets | Accounts | Tweets per | Retweets per | ||
---|---|---|---|---|---|---|
Account / Day | Account / Day | |||||
DS1 | 115,913 | 63,164 | (54.5%) | 20,563 | 0.31 | 0.17 |
(GT) | 4,193 | 2,505 | (59.7%) | 134 | 1.74 | 1.04 |
DS2 | 1,571,245 | 729,937 | (56.5%) | 1,381 | 3.12 | 1.45 |
4.1 The Experiment Datasets
The two real-world datasets selected (Table 2) represent two primary collection techniques: filtering a stream of posts using keywords direct from the OSN (DS1) and collecting the posts of specific accounts (DS2):
- DS1
-
Tweets relating to a regional Australian election in March 2018, including a ground truth subset (GT); and
- DS2
-
A large subset of the Internet Research Agency (IRA) dataset published by Twitter in October 2018999https://about.twitter.com/en_us/values/elections-integrity.html.
The data were collected, held and analysed in accordance with an approved ethics protocol101010Protocol H-2018-045 was approved by the University of Adelaide’s human research ethics committee..
DS1 was collected using RAPID (Lim et al., 2019) over an 18 day period (the election was on day 15) in March 2018. The filter terms included nine hashtags and political handles (candidate and party accounts)111111Not included, but available on request, as per the ethics protocol.. The dataset was expanded by retrieving all replied to, quoted and political account tweets posted during the collection period. The political account tweets formed our ground truth.
The IRA tweets cover 2009 to 2018, but DS2 consists of all posted in 2016, the year of the US Presidential election. Because DS2 consists entirely of IRA accounts (Chen, 2015), it was expected to include evidence of cooperation.
4.2 Experimental Set Up
Window size was set at (in minutes) and the three community detection methods used on the aggregated LCNs were:
-
•
FSA_V (=);
-
•
with = (cf., Cao et al., 2015);
-
•
a simple threshold retaining the heaviest edges by removing those with a normalised value below .
Values for and the threshold were based on experimenting with values in , maximising the MEW to HCC size ratio, using the = DS1 and DS2 aggregated LCNs. Values for were based on Zhao et al. (2015) observation that 75% of retweets occur within six hours of posting. This implies that if attempts were made to boost a tweet, retweeting it in much shorter times would be required for it to stand out from typical traffic. Varol et al. (2017) checked Twitter’s trending hashtags every minutes, so values chosen for ranged from 15 minutes to a day, growing by a factor of approximately four at each increment. Coordinated retweeting was expected to occur in the smaller windows, but then replaced by coincidental co-retweeting as the window size increases.
4.3 Experimental Results
The research questions introduced in Section 1 guide our discussion, but we also present follow-up analyses.
4.3.1 HCC Detection (RQ1)
Detecting different strategies.
The three detection methods all detected HCCs when searching for Boost (co-retweets), Pollute (co-hashtags), and Bully (co-mentions), details of which are shown in Table 3. Notably, kNN consistently builds a single large HCC, highlighting the need to filter the network prior to applying it (cf., Cao et al., 2015). The kNN HCC is also consistently nearly as large as the original LCN for DS2, perhaps due to the low number of accounts and the fact that every edge of the retained vertices is retained, regardless of weight. It is not clear, then, that kNN is producing meaningful results, even if it can extract a community.
GT | DS1 | DS2 | |||||||||
Strategy | Nodes | Edges | Comp. | Nodes | Edges | Comp. | Nodes | Edges | Comp. | ||
LCN | Boost | 15 | 44 | 112 | 5 | 8,855 | 80,702 | 419 | 855 | 23,022 | 14 |
Pollute | 15 | 51 | 154 | 2 | 13,831 | 1,281,134 | 73 | 1,203 | 65,949 | 5 | |
Bully | 60 | 70 | 482 | 1 | 16,519 | 1,925,487 | 222 | 1,103 | 37,368 | 5 | |
FSA_V | Boost | 15 | 9 | 6 | 3 | 633 | 753 | 167 | 113 | 758 | 19 |
Pollute | 15 | 9 | 5 | 4 | 135 | 93 | 50 | 24 | 15 | 9 | |
Bully | 60 | 11 | 7 | 4 | 338 | 280 | 119 | 109 | 1,123 | 16 | |
kNN | Boost | 15 | 9 | 21 | 1 | 1,041 | 33,621 | 1 | 675 | 22,494 | 1 |
Pollute | 15 | 11 | 37 | 1 | 724 | 153,424 | 1 | 1,040 | 65,280 | 1 | |
Bully | 60 | 18 | 135 | 1 | 1,713 | 663,413 | 1 | 692 | 35,136 | 1 | |
Threshold | Boost | 15 | 11 | 16 | 3 | 85 | 68 | 31 | 8 | 10 | 2 |
Pollute | 15 | 24 | 26 | 3 | 44 | 37 | 10 | 6 | 13 | 1 | |
Bully | 60 | 15 | 19 | 3 | 25 | 23 | 8 | 10 | 10 | 3 |
Varying window size .
Different strategies may be executed over different time periods, based on their aims. Boosting a message to game trending algorithms requires the messages to appear close in time, whereas some forms of Bullying exhibit only consistency and low variation (mentioning the same account repeatedly). Polluting a user’s timeline on Twitter can also be achieved by frequently joining their conversations over a sustained period.
Graph Attributes | HCC Sizes | Nodes in common | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
Nodes | Edges | HCCs | Min. | Max. | Mean | Std. Dev. | =15 | =60 | =360 | =1440 | ||
GT | 15 | 9 | 6 | 3 | 3 | 3 | 3.00 | 0.00 | 9 | 9 | 8 | 8 |
60 | 14 | 9 | 5 | 2 | 3 | 2.80 | 0.40 | - | 14 | 10 | 12 | |
360 | 13 | 9 | 5 | 2 | 3 | 2.60 | 0.49 | - | - | 13 | 12 | |
1440 | 17 | 12 | 6 | 2 | 3 | 2.80 | 0.37 | - | - | - | 17 | |
DS1 | 15 | 633 | 753 | 167 | 2 | 18 | 3.79 | 2.21 | 633 | 218 | 93 | 100 |
60 | 619 | 1,293 | 151 | 2 | 13 | 4.10 | 2.30 | - | 619 | 208 | 193 | |
360 | 503 | 1,119 | 127 | 2 | 19 | 3.96 | 2.58 | - | - | 503 | 350 | |
1440 | 815 | 2,019 | 141 | 2 | 110 | 5.78 | 12.60 | - | - | - | 815 | |
DS2 | 15 | 113 | 758 | 19 | 2 | 65 | 5.95 | 13.94 | 113 | 34 | 29 | 25 |
60 | 77 | 394 | 18 | 2 | 27 | 4.28 | 5.64 | - | 77 | 62 | 54 | |
360 | 98 | 775 | 15 | 2 | 32 | 6.53 | 9.13 | - | - | 98 | 56 | |
1440 | 69 | 380 | 15 | 2 | 27 | 4.60 | 6.15 | - | - | - | 69 |






Varying searching for Boost, we found different accounts were prominent over different time frames (Table 4); the overlap in the accounts detected in each time frame differed considerably even though the number of HCCs stayed relatively similar. Figure 4 shows the Jaccard and overlap similarity between the sets of accounts appearing in each window size (agnostic of HCC membership). Examining these reveals a number of facts. The overlap results for shows very high levels of similarity, but lower levels of Jaccard similarity. For all datasets, as grows finds more and more HCC members, always finding the ones it found with smaller window sizes (overlap similarity values appear close to , shown as yellow). The highest Jaccard similarities for seem to group the shorter periods () and the longer periods (). FSA_V finds different sets of members in each time window, without significant overlap, though for DS2 it appears that the windows longer than minutes have some common members amongst themselves, but very few with the ones from . As might be expected, thresholding by LCN edge weight results in the identification of additional accounts as increases, and the Jaccard similarity of GT and DS1 (Figure 4c) reveals that accounts identified in the shorter time windows () are very different to those from the longer time windows, but they still overlap somewhat (Figure 4d). This suggests that although there are some accounts that coordinate in short periods, other accounts coordinate more over the longer time periods. These include media accounts that are consistently highly active over longer periods than the bursty kind of activity expected of active discussion participants who might log on to Twitter in the evening for a few hours.
Other than in GT, which revealed very few HCCs, the sizes of the HCCs found seemed to follow a rough power law; most were very small but one or a few were very large (see the HCC Sizes section in Table 4). The number of HCCs did not vary significantly nor consistently as increased. The number of edges retrieved tells us in DS1, as the window increased, more edges had high enough weights to be retained, whereas DS2 edge counts diminished correspondingly, implying that the LCNs were progressively dominated by a smaller number of very heavy edges, while other remained relatively light.
HCC detection methods.
Similarly, HCCs discovered by the three community extraction methods (Table 5) exhibit large discrepancies, suggesting that whichever method is used, tuning is required to produce interpretable results. This is evident in the literature: Cao et al. conducted significant pre-processing when identifying URL sharing campaigns 2015, and Pacheco et al. showed how specific strategies could identify groups in the online narrative surrounding the Syrian White Helmet organisation (Pacheco et al., 2020a). Here we present the variation in results while controlling methods and other variables and keeping the coordination strategy constant, as our focus is the effectiveness of the method.
Graph Attributes | HCC Sizes | Nodes in common | |||||||
---|---|---|---|---|---|---|---|---|---|
Nodes | Edges | HCCs | Min. | Max. | FSA_V | Threshold | |||
DS1 | FSA_V | 633 | 753 | 167 | 2 | 18 | 633 | 56 | 36 |
kNN | 1,041 | 33,621 | 1 | 1,041 | 1,041 | - | 1,041 | 44 | |
Threshold | 85 | 68 | 31 | 2 | 14 | - | - | 85 | |
DS2 | FSA_V | 113 | 758 | 19 | 2 | 65 | 113 | 88 | 4 |
kNN | 675 | 22,494 | 1 | 675 | 675 | - | 675 | 8 | |
Threshold | 8 | 10 | 2 | 2 | 6 | - | - | 8 |






The networks were visualised using the FR layout in Figure 5, revealing further structure within the networks, each of which consisted of a single connected component. To examine the structure of the single component more closely, we applied Louvain analysis (Blondel et al., 2008) and coloured the largest detected clusters. The clustering reveals distinct communities within both the lone HCC found in each of the datasets. It is possible the DS2 ones are more easily discernible either due to the smaller number of accounts ( compared with ) or because the accounts were, in fact, organised teams of malicious actors acting over a longer time frame.
4.3.2 HCC Differentiation (RQ2)
How similar are the discovered HCCs to each other and to the rest of the corpus? The HCC detection methods used relied on network information; in contrast we examine content, metadata and temporal information to validate the results. We contrast DS1 and DS2 results with GT and a RANDOM dataset, constructed to match the HCC distributions in DS1 (FSA_V, =). As DS2 consisted entirely of bad actors, and GT consisted entirely of political accounts, it was felt non-HCC accounts from DS1 would be more representative of non-coordinating ‘normal’ accounts.
Internal consistency.
Visualising the similarities between accounts using the method in Section 3.2.4 (Figure 6), the HCCs are discernible as being internally similar. The RANDOM groupings demonstrated little to no similarity, internal or external, as expected, while the DS2 HCCs demonstrated high internal similarity, as expected of organised accounts over an extended period. The internal consistency of the DS1 HCCs is not as clear as for DS2, possibly due to the greater number of HCCs. Where HCCs are highly similar to others (indicated by yellow cells off the diagonal), it is highly likely these are due to small HCCs, e.g., with two or three members, retweeting the same small number of tweets (fewer than ten) as each other. The use of filtering in conjunction with FSA_V may help remove potentially spurious HCCs, as could a final merge phase, joining HCC candidates whose evidence for coordination matches closely (e.g., two small HCCs retweeting of the same tweets, kept separate by FSA_V but clearly similar).




Temporal patterns.
We applied the temporal averaging technique described in Section 3.2.7 to compare the daily activities of the HCCs found in GT, DS1 and RANDOM (all of which occur over the same time period) in Figure 7a and weekly activities in DS2 in Figure 7b. The GT accounts were clearly most active at two points prior to the election (around day ), during the last leaders’ debate and just prior to the mandatory electoral advertising blackout. DS1 and RANDOM HCCs were only consistently active at different times: around the day leaders’ debate and on election day, respectively. Inter-HCC variation may have dragged the mean activity value down, as many small HCCs were inactive each day. Reintroducing FSA’s stitching element to FSA_V may avoid this. In DS2, HCC activity increased in the second half of , culminating in a peak around the election, inflated by two very active HCCs, both of which used many predominantly benign hashtags over the year.


Hashtag use.
The most frequent hashtags in the most active HCCs revealed the most in GT (Figure 8a). It is possible to assign some HCCs to political parties via the partisan hashtags (e.g., #voteliberals and #orangelibs), although the hashtags of contemporaneous cultural events are also prominent: e.g., #silentinvasion, #detours and #adlww all relate to a contemporaneous international writers’ festival. DS1 hashtags are all politically relevant, but are dominated by a single small HCC (rendered in pale green) which used many hashtags very often (Figure 8b). These accounts clearly attempted to disseminate their tweets through using hashtags in tweets. Furthermore, the hashtags they use relate to political discussions in many regions around the country (all listed hashtags that end in ‘pol’ relate to Australian states’ or the national political discussion communities). Their prominence in hashtag use effectively hampers our ability to analyse the hashtag use of other HCCs, however, but seeing the results in context is important, as it helps to confirm that the pale green HCC is likely engaging in inauthentic behaviour. We can still see that a large portion of hashtag use amongst the other listed HCCs relates to #savotes, #savotes2018, and #saparli, focussing on the South Australian election. If the hashtags had been irrelevant to the election, that could have provided evidence of accounts attempting to divert the discussion to other topics (because those tweets would still have needed to include the collection filter terms – i.e., ones relating to the election – to have been captured in the first place). Similarly, DS2 hashtags were dominated by a single HCC (using relatively general hashtags in tweets) and one issue-motivated HCC (Figure 8c). Given DS2 covers an entire year, it is unsurprising that the largest HCCs use such a variety of hashtags that their hashtags do not appear on the chart (very little evidence of most of the HCCs listed in the legend appear visible in the barchart, even though the axis is a log scale), but it is revealing that at least a small number of HCCs devoted much of their content to using hashtags, while the other most active HCCs did not, indicating that different HCCs detected by searching for one coordination strategy (co-retweet) are engaging (perhaps even more strongly) in other strategies. Perhaps these hashtag disseminator HCCs acted as general distraction or supporter groups, contributing messages sporadically but not consistently.



Analysing hashtag co-occurrences can help further explore the HCC discussions to determine if HCCs are truly single groups or merged ones. Applied to GT HCC activities (Figure 8a), it was possible to delineate subsets of hashtags in use: e.g., one HCC promoted a political narrative in some tweets with #orangelibs (which is an attack on another political party) and discussed cultural events such as the writers’ festival in others with #adlww (Figure 9), but was definitely one group.

Given the great number of hashtags used in even moderate sized datasets such as DS1, using hashtag co-occurrence analysis to examine the broader election discussion in DS1 requires filtering to reveal the underlying core structure of the hashtag co-occurrence network. We limited the minimum frequency of co-occurrences to and also removed the most frequently occurring hashtags (#savotes, #savotes2018, #saparli and #auspol) to produce Figure 10. Application of Louvain cluster detection (Blondel et al., 2008) exposes five clear clusters, though domain knowledge tells us that there is interesting conflation of topics within some of the clusters. The green cluster contains a subclusters of clusters relating to current affairs television programmes (#pmlive, #abc730, #insiders, #outsiders, #qanda and #thedrum), political party and advocacy groups and relevant issues (#onenation, #getup, #climatechange, #climatecrisis, and #stopadani). It also includes political hashtags (e.g., hashtags ending with pol and #votes) that might fit better in the yellow cluster, which is dominated by them and forms the core of the co-hashtag network by including the heaviest edges. The purple cluster consists primarily of location names, apart from #renewableenergy which hangs off #southaustralia, the focus of the election collection.
The other two clusters make apparent the fact that Twitter is an international network and clashes of hashtags can draw in content irrelevant to local issues. The hashtag #liberals in the blue cluster can refer either to the Liberal party in South Australia (the major party that ultimately won the election) but also is used as a focus in American politics, especially rightwing politics, as reflected by the links to #maga, #guncontrol and #2a (i.e., the 2nd Amendment of the United States’ Constitution, which refers to the right to bare arms), as well as #nationalwalkoutday. During the collection period, high school students in the United States staged a national day of protest against gun violence following a mass school shooting121212https://www.nytimes.com/2018/03/14/us/school-walkout.html. The red cluster also highlights content from outside the area of interest, with many terms relating to locations in other countries, possibly bound by sports, given the presence of #fulltime, #nrlstormtigers, #aflwdogsdees, and #sydvbri, the last three of which refer to Australian sporting matches between specific teams.


DS2 covers a longer period and seemed to consist of different teams of accounts driving different topics. As a consequence, its co-hashtag network reveals clearly delineated (but often connected) discussion topics, as shown in Figure 11. It is immediately notable that although the accounts in the dataset were flagged as trolls implicated in attempting to influence the US election, a lot of content is not in English and, in fact, appears to target other countries. Three examples are apparent:
-
•
The green cluster in the centre consists primarily of Russian news-related hashtags, perhaps aimed at a Russian audience to direct their attention to US election-related content.
-
•
The pale blue central cluster has many hashtags related to the Middle East, including the ISIS terrorist group, but also German politicians and German names for nearby countries, such as Turkey. Germany’s response to refugees from Syria escaping ISIS was politically contentious and may have been seen as an opportunity to foster divisions in the European Union and within Germany.
-
•
The green cluster on the lower left is aimed at discussions of the United Kingdom’s (UK) exit from the European Union (EU), otherwise referred to as Brexit. The UK held a referendum in 2016 on whether it should leave the EU and the campaigning caused significant division within the UK and European communities.
Other significant communities in the co-hashtag network are the pink Tea Party / Conservatives Online (#tcot and #ccot) cluster, tightly connected to the emerging #MAGA cluster supporting Donald Trump, the red cluster focused on American patriotism and the highly active brown cluster including the terms #news, #local, #business and #world. The activity of HCCs shown in Figure 8c above presents a different and complementary view into hashtag use in the dataset, as very little of it apparent in the co-hashtag network.
Examining the Ground Truth.
The importance of having ground truth in context is demonstrated by Keller et al. (2017, 2019). By analysing the actions of known bad actors in a broad dataset, they could identify not just different subteams within the actors and their strategies, but their effect on the broader discussion. Many datasets comprising only bad actors (e.g., DS2) miss this context.
Considering GT alone, the HCCs identified consist only of members within the same political party, across all values of . Some accounts appeared in each window size. HCCs of six major parties were identified. Figure 12 shows the HCCs for each value. Some accounts and parties appeared at each window size, (e.g., parties L, A, G, and nodes L2, A1, G2), while some only appear in a few (e.g., parties C and S). This shows that different parties exhibited different approaches to retweeting and different members were involved over different time frames. Although party S members co-retweeted enough to appear in two time windows, they were not consistently active enough to re-appear in the largest time window, where their activity was overtaken by other accounts. It is particularly noticeable that the L party had two core cooperating accounts, L2 and L4, who were active enough to appear in each time window, and then a large team active in the hour-long window, implying that a deliberate strategy of team-based co-retweeting was employed (rather than a coincidental one). Rather than the posting times being highly coordinated (so that retweets could appear nearly simultaneously), it appears as if the L accounts were simply relatively highly attentive to their colleagues’ tweets and retweets and retweeted them when they saw them appear (which often occurred within an hour), as could be expected of a modern social media-savvy party.




Examining the content of these HCCs confirmed that they were genuine communities using the co-retweet strategy (not necessarily deliberately). The top retweeted tweets of each HCC (=15) are:
-
•
“RT @alpsa: A message from former @AustralianLabor Prime Minister, REDACTED. https://t.co/URL131313URLs starting with ‘https://t.co/’ refer back to the original retweeted tweet’s URL, and are obscured here for readability and anonymity.”
-
•
“RT @REDACTED: Liberals promise $40m to tackle elective surgery waiting times in South Australian hospitals. #SAVotes2018… https://t.co/URL”
-
•
“RT @SALibMedia: Under Labor there aren’t enough job opportunities for young South Australians. Here’s what they are saying about @REDACTED and @alpsa #saparli https://t.co/URL”
-
•
“RT @REDACTED: The results of this state election are clear - celebrity candidates and pop up parties come and go, but the Greens…https://t.co/URL”
Using the tweets each HCC posted, it is possible to attribute each to a political affiliation, if not directly to a party, without resorting to inspecting the identities of the members.
4.3.3 Focus of Connectivity (RQ3)


The IRRs and IMRs for the HCCs in the DS1, DS2, GT and RANDOM datasets is shown in Figure 13. The larger the HCC size, the greater the likelihood of retweeting or mentioning internally, so it is notable that DS2’s largest HCC has IRR and IMR’s of around , though even the smaller HCCs have low ratios. Ratios for the smallest HCCs seem largest, possibly due to low numbers of posts, many of which may be retweets or include a mention, inflating the ratios. The hypothesis that political accounts would retweet and mention themselves frequently is not confirmed by these results, possibly because they are retweeting and mentioning official or party accounts outside the HCCs.
4.3.4 Content Variation (RQ4)



We compared the entropy of features used by DS1 and DS2 HCCs to RANDOM ones (Figure 14). Many of DS1’s small HCCs used only one of a particular feature, resulting in an entropy score of (Figure 14a). In contrast, DS2’s fewer HCCs have higher entropy values (Figure 14b), likely because they were active for longer (over , not , days) and had more opportunity to use more feature values. The majority of HCCs used few hashtags and URL domains, which is to be expected as the dominating domain is twitter.com, which is embedded in all retweets as links back to the original tweet. Compared to the RANDOM HCCs (Figure 14c), DS1 HCCs had lower variation in all features, while the longer activity period of DS2 resulted in distinctly different entropy distributions. Because DS1 HCC activity appears to have been more deliberate, and perhaps coordinated, it may be that the HCCs were more focused on their topic of conversation (especially when contrasted with RANDOM HCCs). Compared with RANDOM HCCs, DS1 HCCs retweeted fewer accounts, used fewer URLs (though they were from a similar distribution of domains), and many fewer mentions and hashtags. Many non-HCC accounts posted only a single retweet as their contribution to the discussion, and so it may be that a relatively high proportion the RANDOM HCC members only posted a single tweet, causing the distributions observed. As noted previously, the RANDOM HCC members posted tweets compared with DS1 HCCs’ tweets, despite having the same number of members, so DS1 HCC members posted more than times as often. DS1 accounts clearly posted more tweets per individual than the RANDOM accounts, yet their distribution appears similar, and notably different to those of both DS2 and GT (Figure 15).

4.3.5 Consistent Coordination (RQ5)


=15 | =60 | =360 | =1440 | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
N | E | C | N | E | C | N | E | C | N | E | C | ||
DS1 | Baseline | 633 | 753 | 167 | 619 | 1,293 | 151 | 503 | 1,119 | 127 | 815 | 2,019 | 141 |
=0.5 | 604 | 711 | 168 | 1,178 | 2,121 | 149 | 519 | 1,183 | 129 | 800 | 2,037 | 137 | |
=0.7 | 578 | 697 | 160 | 847 | 1,569 | 149 | 518 | 1,155 | 130 | 792 | 1,997 | 136 | |
=0.9 | 596 | 706 | 165 | 585 | 1,223 | 145 | 530 | 1,188 | 134 | 796 | 1,995 | 141 | |
DS2 | Baseline | 113 | 758 | 19 | 77 | 394 | 18 | 98 | 775 | 15 | 69 | 380 | 15 |
=0.5 | 116 | 760 | 20 | 79 | 396 | 18 | 100 | 776 | 16 | 69 | 380 | 15 | |
=0.7 | 110 | 756 | 18 | 79 | 395 | 18 | 102 | 777 | 17 | 69 | 381 | 15 | |
=0.9 | 113 | 758 | 19 | 79 | 396 | 18 | 100 | 776 | 16 | 69 | 381 | 15 |
The sliding frame technique from Section 3.2.9 was applied to DS1 and DS2 to reveal HCCs engaging in coordination consistently in adjacent time windows. The baseline used = (i.e., a sliding frame a single time window wide) and =. For the three other conditions, was set to as increases approximately times each time, and =. In this way, the choice of = would most strongly consider the contribution of LCNs from preceding time windows. Once applied for each time window, the aggregated LCNs were mined for HCCs and then the membership of these were compared in the same manner as in Section 4.3.1 using Jaccard similarity (Equation 2), which, as noted earlier, is stricter about set matching than the Overlap method (Equation 3). Even so, as can be seen in Figure 16, changes introduced by using the decaying sliding frame with different values were insignificant in all cases, except for DS1 and =. The implication, which is borne out when the exact network sizes (in nodes) are compared in Table 6, is that the previous windows did not add significant numbers of nodes, but instead increased the weight of existing edges, so the HCCs that were detected consisted of the same members working together over time, rather than splitting into subsets. To hide a team’s coordination, one might expect that its members would associate separately in different time window, but that does not appear to have happened significantly in these datasets, except in the shorter time windows in DS1.
The greatest variation in node and edge count occurs in the shorter windows in DS1 (=), probably because of the greater number of accounts active in DS1 (compared to DS2): accounts have more alters to form HCCs with in DS1, which has k accounts, whereas choice in DS2 is limited to k accounts. The near doubling of accounts in DS1’s HCCs when = implies accounts co-retweeted often just within a single hour, and then not again (at least not for = hours). This effect is swamped by the much more active consistent co-retweeting of a smaller set of users when is increased to and above. Given the membership varies so little in the other conditions, an analysis of how these HCCs form and change over time is required. It is clear, however, that this approach would be best suited to filter-based collections, as they are likely to capture more accounts.
4.3.6 Validation via HCC Classification (RQ6)
Our final validation method relies on the HCCs in GT as positive examples of coordinating sets of accounts, given it is reasonable to assume that they ought to be coordinating their activities during an election campaign.
Feature selection.
As mentioned in Section 3.2.10, features are drawn from individual accounts and their groupings, specifically based on their individual and collective behaviour and homophily. For this reason, we select account-level features as well as group-level features to make up each account’s feature vector, meaning that some of the values for HCC co-members will be identical. The following are the selected account-level features (all are regarding posts made by the account within the collection period):
-
•
Post count: Number of posts they made.
-
•
Repost count: Number of reposts (retweets) included in the account’s posts.
-
•
Reply count: Number of replies included in the account’s posts.
-
•
Posting rate: Number of posts in the corpus divided by the length of the collection period in minutes.
-
•
Unique mentions: The number of unique accounts the account in question mentioned. N.B. A post may include multiple mentions, so the number of mentioned accounts and mention uses may exceed the number of posts.
-
•
Mention count: The number of mentions of accounts made by this account.
-
•
Unique hashtags: The number of unique hashtags (in lower case) used by the account. N.B. A post can include multiple hashtags, so the number of unique hashtags and their uses may exceed the number of posts.
-
•
Hashtag uses: The number of times hashtags were used by this account.
-
•
Unique URLs: The number of unique URLs used by this account. N.B. A post can include multiple URLs, so the number of unique URLs and their uses may exceed the number of posts.
-
•
URL uses: The number of times URLs were used by this account.
-
•
Default profile image: A Boolean representing whether or not the account is using the default profile image provided by the social media platform (i.e., whether the account has customised their profile image).
-
•
Length of profile description: The length, in characters, of the account’s profile description in their first post in the corpus.
-
•
Profile URL length: The length, in characters, of the account’s profile URL in their first post in the corpus.
The following features are drawn from the HCC’s activity network (see Section 3.2.10) and included in the feature vector of each member of the HCC:
-
•
Aggregate post count: The total number of posts by the members of the HCC in the corpus.
-
•
Member count: The number of accounts in the HCC.
-
•
Interaction count: The number of edges in the HCC’s activity network.
-
•
User count: The number of account nodes in the HCC’s activity network (includes accounts outside the HCC with whom the members interacted).
-
•
Author count: The number of the accounts in the HCC.
-
•
Unique hashtag count: The number of unique hashtags used in the HCC’s members’ posts.
-
•
Hashtag uses: The number of hashtag use interactions in the HCC’s activity network.
-
•
Unique URL count: The number of unique URLs used in the HCC’s members’ posts.
-
•
hashtag uses: The number of URL use interactions in the HCC’s activity network.
-
•
Repost count: The number of HCC’s members’ posts that were reposts (retweets).
-
•
Quote count: The number of HCC’s members’ posts that were quotes (i.e., reposts with comments).
-
•
Mention count: The number of mention interactions in the HCC’s interaction network.
-
•
Reply count: The number of reply interactions in the HCC’s interaction network.
-
•
‘In conversation’ counts: The number of times HCC members participated in ‘conversations’ (trees of reply posts rooted in a single post) in the corpus.
-
•
Proportion of HCC members reposted: The frequency of reposts of HCC members divided by the overall number of reposts within the corpus (cf. IRR, see Section 4.3.3).
-
•
Proportion of HCC members mentioned: The frequency of mentions of HCC members divided by the overall number of mentions within the corpus (cf. IMR, see Section 4.3.3).
-
•
Proportion of HCC members replied to: The frequency of replies to HCC members divided by the overall number of replies within the corpus.
Classification results.
After being trained on the GT HCCs, the classifiers were then applied to the HCCs in DS1 and DS2. Unfortunately, the SVM classifier entirely failed to identify the positive class (i.e., coordinating accounts) in DS2, labelling every instance as non-coordinating. The SVM results for DS2 are therefore not considered beyond this point.
= | = | = | = | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Classifier | Acc. | Acc. | Acc. | Acc. | |||||||||
DS1 | SVM | 0.91 | 0.92 | 0.91 | 0.76 | 0.73 | 0.79 | 0.58 | 0.37 | 0.68 | 0.87 | 0.86 | 0.88 |
RF | 0.65 | 0.70 | 0.59 | 0.59 | 0.60 | 0.58 | 0.59 | 0.38 | 0.70 | 0.92 | 0.92 | 0.93 | |
BPU | 0.69 | 0.72 | 0.66 | 0.65 | 0.63 | 0.66 | 0.60 | 0.37 | 0.71 | 0.88 | 0.86 | 0.89 | |
DS2 | SVM | 0.50* | 0.00* | 0.67 | 0.50* | 0.00* | 0.67 | 0.50* | 0.00* | 0.67 | 0.50* | 0.00* | 0.67 |
RF | 0.89 | 0.88 | 0.90 | 0.94 | 0.94 | 0.94 | 0.97 | 0.97 | 0.97 | 0.91 | 0.90 | 0.91 | |
BPU | 0.93 | 0.93 | 0.94 | 0.92 | 0.91 | 0.92 | 0.95 | 0.95 | 0.96 | 0.54 | 0.18 | 0.68 |
The accuracy of the best classifier for each dataset and time window ranged from to (shown in Table 7), with better performance identifying HCC members in DS2 than in DS1. scores (outside =) for the COORDINATING () and NON-COORDINATING () instances ranged from to and to , respectively. Each classifier performed best for DS1 in different time windows, and RF and BPU performed similarly well for DS2 in all time windows, except for BPU struggling in the =1440 time window. All classifiers also performed the least well in the six hour time window for DS1, possibly because the GT HCCs’ activity coordination was most prominent over the short time frames of an hour or less, and otherwise at the day level. Even so, scores consistently hover around , which is significantly better than random, though the scores around indicate difficulty identifying all COORDINATING HCC members, a detail which is discussed in more detail below. The accuracy and results show that the the classifiers could all be successfully trained to recognise GT HCCs in most time windows and that the GT HCCs represented most of the HCCs in DS1 and DS2, despite the different levels of activity (DS2 HCC members interacted more than DS1 or GT HCC members in their corpus, primarily because the collection period was so much greater).
= | = | = | = | ||||||
---|---|---|---|---|---|---|---|---|---|
Classifier | Precision | Recall | Precision | Recall | Precision | Recall | Precision | Recall | |
DS1 | SVM | 0.86 | 0.99 | 0.85 | 0.63 | 0.73 | 0.25 | 0.91 | 0.82 |
RF | 0.61 | 0.80 | 0.59 | 0.62 | 0.80 | 0.25 | 1.00 | 0.85 | |
BPU | 0.66 | 0.78 | 0.66 | 0.60 | 0.86 | 0.23 | 1.00 | 0.76 | |
DS2 | SVM | 0.00* | 0.00 | 0.00* | 0.00 | 0.00* | 0.00 | 0.00* | 0.00 |
RF | 1.00 | 0.79 | 1.00 | 0.88 | 0.96 | 0.98 | 1.00 | 0.81 | |
BPU | 0.97 | 0.89 | 0.94 | 0.88 | 1.00 | 0.91 | 0.78 | 0.10 |
Table 8 shows Precision and Recall across all classifiers and datasets for the COORDINATING class. (Given our emphasis on recognising COORDINATING instances, we do not present the corresponding results for the NON-COORDINATING class here.) For all time windows, Precision is high for the classifiers against DS1 (ranging from to ) and is for all DS2, meaning that the HCCs are clearly discernible from the NON-COORDINATING instances. Recall varies significantly for DS1 (between and ), but less so for DS2 (between and ), meaning that some DS1 HCCs were misclassified but most were correctly identified in DS2. The Recall scores for = explain why the scores were so low in Table 7, because the corresponding Precision scores are still relatively high. As mentioned above, there is something particular about the six hour time window (=), as the GT HCC members (via their account features and group behavioural features) were less easily distinguishable from the randomised NON-COORDINATING accounts, resulting in poorer performing classifiers. The reason for this is possibly the choice of window boundaries. The time window boundaries rested at 0000, 0600, 1200, and 1800 hours, while boundaries defined more by work activity (e.g., 0200, 0800, 1400, 2000 hours) may better match human activity patterns. For other, less geographically bound datasets (i.e., ones where the activity comes from around the world, rather than from a single or small group of adjacent timezones), other ground truth may be required.
SVM was the best performing classifier for COORDINATING accounts in DS1 in the shorter time windows (=), but RF performed best for the day long time window. Although Precision is high for =, it is hard to argue the BPU classifier performed significantly better than the others. For DS2, both RF and BPU performed well, and similarly, except for BPU’s Recall in the = time window. For that reason, we can argue that RF performed best overall, but no particular consistently can be observed.
Consequently, by accepting the assumption that the ground truth HCCs exhibited at least one type of coordination, these classifiers provide confidence that the other HCCs appear similar to the GT ones and thus may have behaved in similar ways. The question of intent remains, however. Examining the content subjected to coordination will likely provide clues, but deeper examination of behaviour to identify, e.g., Princpal-Actor patterns (Giglietto et al., 2020b), may also be enlightening. More examples of similar coordination activities as well as other coordination types would bolster the positive training and testing sets, as well as expand knowledge regarding coordination strategies in use online.
4.3.7 Multiple Criteria: Bullying
Some strategies can involve a combination of actions. Behaviours that contribute to Bullying by dogpiling, for example, include joining conversations started by the target’s posts and mentioning the target repeatedly, within a confined time frame. As DS1 included all replied to tweets, we investigated it inferring links via co-mentions and co-convs with a window size of minutes, and FSA_V with =, having maximised the ratio of MEW to HCC size. Of HCCs discovered, the largest had five accounts and most only had two. Only had more than ten inferred connections, but five have more than . These heavily connected accounts, after deep analysis, were simply very active Twitter users who engaged others in conversation via mentions, which outweighed the more strict co-conv criterion of participants replying into the same conversation reply tree.
A larger window size was considered (=) in case co-conv interactions were more prevalent. FSA_V (=) exposed little further evidence of co-conv (Figure 17), finding small HCCs again dominated by co-mentions, not many of which had more than one inferred connection, implying most links were incidental; FSA_V did not filter these out.

This provides an argument for a more sophisticated approach to combining LCN edge weights for analysis, instead of Equation (1), and that FSA_V could be modified to better balance HCC size and edge weight. Furthermore, it is likely that bullying accounts will not just co-mention accounts frequently, but have low diversity in the accounts they co-mention, i.e., they repeatedly co-mention a small set of accounts, and spend a disproportional number of their tweets doing so. A further consideration is that participants in long discussions (reply trees) often include the author of the original tweet that sparked the discussion, and it would be misleading to include their account in results, implying that they bullied themselves. Finally, patterns of behaviour that would clearly qualify as conversations were observed in the datasets that did not fit the strict ‘conversation tree’ model: accounts would mention several collocutors at the start of every tweet, but only potentially reply a tweet of one of them to continue the conversation. Importantly, sometimes the mentioned accounts included in tweets were prominent individuals whose names were included not because they were active participants in the conversation, but because the tweeter wanted to draw their attention to the conversation (regardless of the likelihood that the attempt would succeed; e.g., some tweets included references to prominent and busy politicians who would be unlikely to wade into arbitrary online discussions). These nuances are not explored here.
4.3.8 HCC inter-relationships
To study the relationships between HCCs, we create two-level networks starting with the HCC network and then adding nodes representing the elements of evidence linking them, known as reason nodes (e.g., the tweets they co-retweet or the hashtags they use in common). Figure 18 shows the largest component after such expansion was conducted on the HCCs in Figure 17. HCC accounts (circles) share colours and the distribution of the reasons for their connection (diamonds) show which other accounts are uniquely mentioned by an HCC and which are mentioned by more than one HCC. Heavy links between HCC accounts with few adjacent reason vertices imply these accounts are mentioning a small set of other accounts on many occasions.

4.3.9 Boosting Accounts, not just Posts
It is possible to Boost an account rather than just a post. Returning to DS2, we sought HCCs from accounts retweeting the same account (FSA_V, =), and found that the hashtag use revealed further insights (Figure 19). No longer does one HCC dominate the hashtags. Instead clear themes are exhibited by different HCCs, but again, they are not the largest HCCs. The red HCC uses #blacklivesmatter and other Black rights-related hashtags (including #blm, #blacktolive, #blackskinisnotacrime, #policebrutality and #btp141414BTP refers to the British Transport Police, the conduct of which was discussed in accounts of the arrest of a Black man at a London train station in mid-2016, e.g., https://www.theguardian.com/uk-news/2016/jul/28/man-complains-after-police-place-spit-hood-over-head-during-arrest-london-bridge.), while the purple HCC uses pro-Republican ones (#maga and #tcot), and the green HCC is more general. Given the number of tweets these HCCs posted over 2016 (at least ), it is clear they concentrated their messaging on particular topics, some politically charged. It is arguable that their contributions helped inflame tensions and stoke divisions in socially sensitive topics, not just in the United States, but in the UK as well, and at the very least sought to draw the attention of others.
The green HCC may be acting in a distractor role, as previously mentioned, given their contribution of tweets over the year.

4.3.10 Validation of Inauthentic Behaviour Detection
The approach presented can be used to perform analytics similar to the Rapid Retweet Network used by Pacheco et al. (2020a), who used it to expose tight clusters of bot-like accounts, which retweeted the same tweet within seconds of it appearing. We varied this for the DS1 dataset (due to its small nature) and searched for accounts which retweeted the same tweet within seconds, regardless of the age of the original tweet. We discovered a tight cluster of accounts, most with relatively high Botometer ratings151515The English “all features” score was used as both our datasets are primarily either English speaking or are predominantly aimed at English speaking audiences. (Davis et al., 2016), shown in Figure 20. The bot ratings were as follows: node 26: ; node 22: ; node 2: ; and node 17: . All were high relative to the other accounts in the corpus, most of which had scores well below ; all four were had scores well above , but the scores of two were also well above the ‘bot’ threshold of . On further inspection, they appeared to support vocational training and left-wing issues and posted retweets almost exclusively, but the content all related to the election. This finding enhances the bot ratings by making it clear which bots (or bot-like accounts) appear to work together. It also raises further questions regarding bot detection systems, however, as some of the accounts appeared to be genuine human, though abnormally active on Twitter. These accounts appeared to work together to actively disseminate messages aligned with their preferred narrative, though with a very low IRR (just shy of ) despite most of their activity being retweets (), so to a certain degree it matters not whether they are automated or genuinely human-driven, but whether they are engaging in astroturfing or other inauthentic behaviour. In this circumstance, they may be genuine agenda-driven users, but they were definitely highly attentive to the same sources. Alternatively, when we consider their bot ratings more closely, it is possible that there is a mixture of account types, with node 26, in particular, acting as an automated ‘cheerleader’ for nodes 22 and 17. Relative timings of their posts (to answer whether node 26 consistently was the second co-retweeter when paired with nodes 22 and 17) could reveal support for this hypothesis.

4.3.11 Performance
In Table 9 we present the timings observed for the stages of processing for DS1 and DS2 conducted on a Dell Precision 5520 laptop equipped with an Intel Core i7-7820HQ CPU (2.9GHz), 32Gb RAM, and an NVMe PC300 480Gb SSD, running Windows 10. Parsing raw data is relatively cheap, with DS2’s 1.5m tweets processed in just over a minute, and LCN construction dependent on the degree of activity within the time windows and the number of accounts. DS1’s larger account pool increased the (node) size of the networks generated, and all associated post-processing. The size of DS1 LCNs were an order of magnitude greater than DS2’s (in vertices and edges), resulting in correspondingly increasing execution times for aggregation and HCC extraction.
DS1 | DS2 | |||||||
Tweets | 115,913 | 1,571,245 | ||||||
Parse raw (Step 1) | 19.0 (from JSON) | 74.0 (from CSV) | ||||||
Window size (minutes) | 15 | 60 | 360 | 1440 | 15 | 60 | 360 | 1440 |
Find evidence and build LCNs | 15.0 | 28.0 | 123.0 | 427.0 | 121.0 | 106.0 | 246.0 | 567.0 |
Aggregate LCNs | 27.0 | 65.6 | 168.5 | 170.7 | 70.4 | 55.2 | 35.6 | 22.7 |
HCCs: FSA_V | 28.3 | 58.2 | 126.1 | 209.3 | 6.3 | 4.2 | 5.8 | 5.0 |
HCCs: kNN | 9.0 | 22.7 | 97.5 | 206.4 | 4.3 | 4.3 | 4.7 | 4.6 |
HCCs: Threshold | 5.2 | 11.9 | 34.6 | 64.0 | 2.2 | 2.3 | 2.7 | 2.7 |
4.4 Applications
Complementing the detailed validation presented above, in this section, we offer two case studies in which our method has been used in order to demonstrate its utility. Extending a study of polarised online communities in a discussion of bushfires in Australia (Weber et al., 2020), the co-URL and co-URL domain analysis we conducted revealed how sources of were used by discussion participants, and how that use differed between the polarised communities. A second study of Twitter activity during the Democratic and Republican Conventions in the United States in August, 2020, making use of co-retweeting analysis in order to reveal influence attempts with social bots and co-hashtag analysis to discover discussion groups and their relations.
4.4.1 #ArsonEmergency and Australia’s “Black Summer”
During the Australian summer of 2019-2020, evidence of inauthentic Twitter behaviour emerged on the hashtag #ArsonEmergency (Weber et al., 2020) over an day period, including a week prior to and more than a week after a study on the matter (Graham and Keller, 2020) was reported in the online tech and then mainstream media (Stilgherrian, 2020). Analysis of the tweets including the hashtag over the period revealed two clearly polarised retweeting communities, one supporting the narrative that arson was the cause of the bushfires and that eco-activism had prevented forest fuel load management (Supporters) and one that countered the narrative, providing evidence that the fires were mostly started by natural or unintended causes (e.g., lightning and sparks from machinery) and the bushfires’ ferocity was exacerbated by climate change (Opposers). The remaining accounts in the dataset were referred to as Unaffiliated.
Differences between the communities’ interaction and information sharing behaviour was apparent. Supporters interacted with other accounts more by using replies, quotes, and mentions more than Opposers or the Unaffilated, as well as more hashtags and external URLs (i.e., referring to domains other than twitter.com), but they retweeted less. Notably, an analysis of the most shared URLs revealed that Opposers shared articles debunking the narrative exclusively, while Supporters shared a mixture of articles, mostly ones supporting or actively discussing their preferred narrative as well as some conspiratorial content. Unaffiliated accounts shared narrative-focused URLs initially, but in the latter phase of the collection they shared debunking articles times more often.
Analysis of the co-use of URLs revealed further behavioural differences between the communities (Figure 21). Opposers were more focused that Supporters in the URLs they shared, relying on a small set of debunking articles, also heavily shared by Unaffiliated accounts. Supporters were less tightly clustered around particular articles, and did share some debunking material as well as a variety of narrative-aligned articles.

Co-domain analysis identified not just distinct URLs but distinct URL domains favoured by the different communities. Figure 22 shows two clusters of domains: the red one contains domains from a number of conservative and right wing media organisations, while the blue one contains academic and centre and left wing media organisations. Although Supporters mostly referred to the narrative-supporting domains while the Opposers mostly referred to the debunking domains, it is notable that members of both communities referred heavily the ABC and the Guardian, which both published articles debunking the arson theory, often with reports from local fire fighting and law enforcement organisations. What is lost at this level of analysis is the way in which the articles were discussed when mentioned in tweets, including whether the tweets were agreeing or attacking the article content.

These co-analyses reveal how focused the polarised communities in their information sharing activities, contributing to the argument that the targeted efforts of the Opposer community may have helped influence the broader Unaffiliated community into sharing debunking articles.
4.4.2 Twitter Discussion Groupings during the 2020 US Political Conventions
A second case study making use of these techniques relates to the search for social bots attempting to influence the online discussion surrounding the Democratic and Republican National Conventions in August 2020, at which the parties formally nominate their candidates for the Presidential Election, later in the year. For a hour period over each -day convention, tweets were filtered using RAPID, starting with #demconvention and #rnc2020 for the Democratic National Convention (DNC) and the Republican National Convention (RNC), respectively. For the three hours prior to the formal collection, RAPID’s topic tracking feature was enabled, which added hashtags used frequently in the discussions observed, resulting in larger sets of filter terms for each convention:
-
•
DNC: #demconvention, #bidenharris, #bidenharris2020, #khive,
#signsacrossamerica, #unitedforbiden, and #wewantjoe; -
•
RNC: #rnc2020, #rncconvention, and #nevertrump.
Despite the disparity in hashtags used (i.e., found to occur sufficiently frequently alongside the seed terms), each dataset ultimately comprised approximately million tweets by over thousand unique users at each convention. Bots are often used to boost tweets, reaching other accounts that follow them, or by flooding hashtag communities or gaming trending algorithms (Woolley, 2016; Hegelich and Janetzko, 2016; Keller et al., 2019; Graham and Keller, 2020; Graham et al., 2020a). Social bots are specifically designed to mimic genuine human users, hiding the fact they are automated (Ferrara et al., 2016; Grimme et al., 2018). They do this to avoid detection, and in doing so can contribute to astroturfing campaigns, artificially boosting narratives while making them appear as simply popular grass roots movements.
By searching for accounts that co-retweeted within a ten second window of each other and then using a minimum normalised edge weight threshold of , several HCCs were identified in each convention (see Figure 23). Analysis of the HCC members using Botometer (Davis et al., 2016) found the majority had a Complete Automation Probability (CAP) ratings above (the relatively high threshold used by Rizoiu et al., 2018, as noted earlier). Further analysis of the HCCs’ content provided some indication of their agendas, and examination of their account age and posting rates enabled categorisation into official accounts (verified by Twitter), unofficial reposters (topic-focused aggregators), and accounts that gave the appearance of typical human users. These ‘normal people’, however, posted at very high average daily rates for years, often at far greater rates than previous automation detection methods have used (e.g., 50 tweets a day, Neudert, 2018).
The largest HCC (the large blue HCC in Figure 23b) consisted of a cluster of potential social bot accounts supporting an official political campaign account, @TrumpWarRoom, responsible for tweets during the Republican Convention. For each pair of members in each HCC, we considered the proportion of time that one account retweeted a tweet before the other, to determine if both accounts were potentially working together (in which case, they would be equally likely to retweet a tweet first), or if one was a ‘cheerleader’ for the other (in which case the cheered account would always retweet first, quickly followed by the other account). We found strong evidence that at least three of the accounts were cheerleaders for @TrumpWarRoom, retweeting the same tweet within ten seconds on , , and occasions over the four day collection period. These particular accounts had daily tweeting rates of , and tweets per day for , and years, respectively. Given the age of these accounts, it is clear that they have successfully avoided Twitter’s bot scanning processes for some considerable time.


We also applied co-hashtag analysis (FSA_V, =, = seconds) to the two datasets and plotted two-level networks of the resulting HCCs with the hashtags they used (Figure 24). Regardless of the content, a number of structures are immediately apparent. These include:
-
•
clusters that are bound around a few yellow diamond hashtag nodes (e.g., DNC clusters 5, 6 and 8) or lie between hashtags (e.g., DNC clusters 2 and 4);
-
•
fan shapes that consist of a small number of accounts using a wide variety of hashtags (e.g., DNC clusters 1 and 7);
-
•
island clusters that are bound by the hashtags they use but are isolated from the broader community which has ignored the hashtags they are using (e.g., DNC clusters 7 and 8).


The fact that the clusters are coloured according to their HCC in Figure 24 highlights how what FSA_V is extracting what it regards as distinct clusters are, in fact, bound by the topics they are discussing (by the hashtags they are co-using). This indicates that there may be benefit in re-introducing the re-stitching step in FSA (Şen et al., 2016) that FSA_V avoids, or also experimenting further with FSA itself. Using conductance cutting (Brandes et al., 2007) for cluster detection aligned better with the visible clusters, but these clustering may be misleading to varying degrees, as it may combine polarised HCCs, as can be seen on closer inspection below.
Several co-hashtag clusters in Figure 24a provide insight into the nature of parts of the online discussion.
-
•
Cluster 5 is closely centred on two hashtags (#goodyear and #ohio) that relate to then US President Donald Trump’s call for a boycott of Goodyear tires161616The Goodyear factory in Ohio banned clothing with political messaging, including the Trump campaign’s MAGA caps, during the election campaign. Source: https://www.abc.net.au/news/2020-08-20/donald-trump-calls-for-goodyear-boycott-over-alleged-maga-ban/12577372 though it is unclear whether the surrounding accounts are for or against the boycott. Several hashtags linked on the left edge of the cluster indicate that some are against, as they refer to support for the then Democratic candidate Vice President Joe Biden.
-
•
The fan cluster at the top consists of two accounts that are attempting to disseminate their message across America, as each hashtag is a US state code (e.g., #ga for Georgia) or a minority group (e.g., #latinos). These hashtags are all apparently unique, apart from the one highlighted just below cluster 1 surrounded by a blue diamond, which is #blm, linking cluster 1 to cluster 2, and the one to the left, which is another state code (#nc for North Carolina).
-
•
Cluster 2 binds a number of HCCs spanning two relatively disjoint hashtags, one being #vote (below the cluster) and the other being the name of a musician who had recently encouraged his fans to vote.
-
•
Cluster 3 is more diffuse than the others and appears to relate to a discussion of data science and big data in the context of the election campaign.
-
•
Cluster 4 appears to join a number of potentially opposed HCCs, as they refer to #trump2020landslide and #snowflakes as well as #epstein171717Jeffrey Epstein was a billionaire arrested for sex crimes before dying in custody, however he was known to Donald Trump, and therefore this hashtag’s use can be seen as an attack on his political campaign. https://www.forbes.com/sites/lisettevoytko/2020/10/18/spider-book-excerpt-how-trumps-presidency-helped-expose-jeffrey-epstein/ and #trumpvirus (a condemnation of the Trump administration’s handling of the response to the COVID-19 pandemic), the final hashtag which links the cluster into the broader community.
-
•
The island clusters 7 and 8 are focused on groups of particular politicians, which were not picked up by the broader community: Republicans who had pledged to vote for the Democratic candidate and US Congress members known to campaign for social equality, respectively.
The links between the clusters are sometimes deceptive. Already, we observed that some single clusters include polarised HCCs, however it is also possible to see internally (politically) consistent clusters that are linked but also contrary in their views. Cluster 1 in Figure 24a is linked to the left by #nc to another left-leaning cluster (calling for gun control), which itself is linked to the left by #america to another small cluster, which is clearly right-leaning (one of its hashtags is #voteredtosaveamerica). These visualisations may highlight how HCCs can be merged, but care must be taken regarding the conclusions to draw from such an action.
Analysis of the co-hashtag HCCs and their hashtags in Figure 24b offers further examples of these observations and offers new insights. Clusters 1 and 2 are joined by the blue diamond-highlighted hashtag, #blacklivesmatter, but cluster 1 is a detractor group (using #alllivesmatter) while cluster 2 is a supporter group using several Black rights-related hashtags. Cluster 4 discusses riots following Black Lives Matter protests in Kenosha, Wisconsin, however while the two sets of hashtags highlighted at the top of the cluster relate mostly to current events (e.g., #kenosha and #covid19 on the left, and #kenoshariots and #thursdaythoughts, plus #walkaway, which links to a small fan, as it is a pro-Republican statement to avoid conflict), the hashtags at the bottom of the cluster are more clearly right wing or conservative in nature, referring to a relevant media organisation, #kag2020 (Keep America Great, a pro-Trump slogan) and #ccot (Christian Conservative on Twitter). Whereas cluster 4 in Figure 24a includes polarised HCCs, the placement of the hashtag nodes they are linked to offers no guidance on how they might be separated, cluster 4 in Figure 24b indicates that an alternative layout algorithm may aid analysis. Cluster 6 represents a concerted anti-Trump effort with many attacking hashtags, but the isolation of the HCC at the cluster’s centre makes it clear that not many of the others tweeting during the RNC took its lead. Cluster 5 is an effort to draw attention to an instance of police brutality, which also did not gain traction with the broader co-hashtag community.
5 Conceptual Comparison and Critique
Methods to discover coordinated behaviour by inferring links between accounts based on related interactions is not unique. Cao et al. (2015) and Giglietto et al. (2019) identified groups of accounts based on the URLs they shared in common, while Lee et al. (2013), Keller et al. (2019) and Graham et al. (2020a) relied on the similarity of the content posted by accounts to do the same. Giglietto et al. explicitly add the temporal element by considering potential links only between accounts that share a URL within a constrained time frame. Their “rational is that, while it may be common that several entities share the same URLs, it is unlikely, unless a consistent coordination exists, that this occurs within the time threshold and repeatedly.”181818Quoted from the README of Giglietto et al.’s open source code (as of 2021-01-19): https://github.com/fabiogiglietto/CooRnet. To the knowledge of the authors, only three other proposed approaches appear to generalise the idea to allow links between accounts to be inferred based on a variety of common behaviours: Pacheco et al. (2020b), Graham et al. (2020a), and Nizzoli et al. (2020).
Pacheco et al.’s method creates strong ties between accounts that share similar behavioural traits. Behavioural traits are extracted from social media data (e.g., hashtags or URLs) and, together with the accounts using them, a bigraph is created. An (weighted) account network is projected from this bipartite network, linking accounts that have edges to the same trait node. The more shared traits, the heavier the edge between accounts. Finally, the account network undergoes cluster analysis, specific to the nature of coordination sought. In their examples, Twitter accounts linked by sharing the same account handle are divided into clusters by virtue of the connected component in which they appear. A second example examining share market “pump and dump” scams links accounts based on the similarity of the text they post, using text frequency/inverse document frequency (TF-IDF), and then clusters are discovered by simply filtering edges with a final weight less than . A third example connects accounts that use multiple hashtags in the same order in their tweets. The approach was employed searching for co-retweeting communities spreading propaganda attacking the Syrian White Helmet movement by linking accounts that retweeted the same tweet within seconds of the original being posted (Pacheco et al., 2020a).
In contrast, Graham et al.’s “coordination network toolkit”191919https://pypi.org/project/coordination-network-toolkit/ (CNT) is also written in Python, but relies on a populated database of information extract from tweets to search for coordinated retweeting (retweeting the same tweet), co-tweeting (tweeting identical text), co-similarity (tweeting similar text), co-linking (sharing the same URL), and co-replying (replying to the same tweet). The database implementation is used to improve the performance of searching for evidence of coordination between pairs of accounts (which, as in our approach, requires pairwise comparison of all accounts in the dataset) by using an inner join. This implementation would need to be modified to suit a streaming data source, but could conceptually be applied to data from a variety of OSNs.
The approach of Nizzoli et al. (2020) is very similar to ours, however it explicitly begins by selecting a set of users of interest, whereas we begin with a corpus of posts and our set of users is defined by those present in it. Nizzoli et al. makes clear that users may be defined by the starting corpus of posts in the same way, or may be otherwise selected as superproducers or superspreaders or followers of a prominent account. They also introduce a filter step before the extraction of HCCs. Pacheco et al. (2020b) filter their user similarity network with an arbitrary filter, which, as pointed out by Nizzoli et al., results in a binary classification of coordinating and non-coordinating users, but importantly disregards the effect of the network structure. Instead, Nizzoli et al. rely on multiscale filtering approaches for complex networks, which retain network structures (not just individual edges) based on statistical significance. Furthermore, these can be scaled to retain more or less of the network, permitted examination of the ‘degree’ of coordination, not just a binary decision on whether or not it is present. They propose an iterative algorithm at this point for detecting clusters of coordinating users, which makes use of an increasingly strict definition of user similarity (i.e., coordination) and each time relies on the communities found in the previous step as the starting point, guaranteeing they are kept in some form. This makes it possible to track communities at different levels of coordination, similar to how core decomposition provides insight into how deeply particular nodes and structures are embedded within a network. Finally, they apply a validation step, studying the resulting networks with network measures, and text analysis of the posts of the HCCs, but all as a function of the resolution at which the HCCs were detected. The FSA_V algorithm is our alternative to their filtering and cluster detection steps. The ability for Nizzoli et al. to examine different degrees of coordination is a distinguishing factor, however they also (just like Pacheco et al., 2020b) must decide beforehand what similarity measure to connect users with – this is equivalent to the behaviours that underpin the coordination strategies we discussed in Section 2, however they make the point that the similarity measure may involve any relevant information about the user profiles, not just their behaviour within the corpus. The temporal aspect of the coordination is not discussed, presumably as it is assumed to be a component of the user similarity measure.
Giglietto et al.’s CoorNet R package does not allow specification of a time window directly, but instead uses a proportion threshold to determine what to regard as an anomalously small but active time window, and thus requires access to an entire dataset. It is designed to study Coordinated Link Sharing Behaviour (CLSB) (Giglietto et al., 2020a) and thus only considers URLs in posts, however, it accepts URLs from a variety of OSNs and sources, including via CrowdTangle202020https://www.crowdtangle.com/ and MediaCloud212121https://www.media.mit.edu/projects/media-cloud/overview/.
Our method is similar to all of these but is described in greater detail, relies upon a discrete window-based approach to apply temporal constraints, and we provide and evaluate a novel cluster extraction algorithm, and an open source implementation is available. By apply time constraints in discrete windows, connections may be missed across windows, but this makes it easier to apply in near real-time streaming settings. If one were to infer connections between accounts as each new tweet is posted, it could create a potentially significant, ongoing processing cost depending on the number of unique accounts observed in the current time window. As new posts arrive, new nodes may need to be added to the account graph, while others may need to be removed, along with their adjacent edges (which, let us recall, represent evidence of coordination, not individual timestamped interactions as one might find in a social network based on direct retweets, mentions or replies). Furthermore, this constantly updated account graph must be complete, i.e., edges should always be added in case the evidence they represent may be consolidated by future posts.
If the choice of time window is very short (e.g., seconds, as per Pacheco et al., 2020a), and LCNs from adjacent windows are aggregated (as per our method), the absence of a truly sliding window like Graham et al.’s may not significantly affect results, as ongoing high levels of coordination will appear over multiple windows. In contrast, if the time window is longer (e.g., or more minutes), then the hard boundary between windows may cause coordinated activities to be missed. The question is, then, what kind of coordination is being sought. Teams of bots tweeting or retweeting the same tweet within small time frames will be vulnerable to detection, however a deliberate covert human team with sockpuppet accounts may escape detection (at least initially) by varying the time frame over which retweets are posted (e.g., spread them unevenly over an hour or more), but if the same accounts cooperate for long, our method will find them once their activities are aggregated. One type of coordination that is very difficult to detect is single event boosts of a post: e.g., when, say, paid accounts retweet a single tweet. In a large discussion, tweets will not stand out especially, but, depending on how connected the paid accounts are to the broader discussion, they may spread the content a considerable distance through the network. Furthermore, gaming OSN trending algorithms may not be difficult222222OSN gaming efforts of the form “Let’s get X trending” are quite common in Australia, e.g., https://twitter.com/Timothyjgraham/status/1351742513044807680, and even a thousand retweets may result in a valuable degree of influence in comparatively smaller communities (e.g., Australia).
As a final comment, all methods discussed in this section are suited to post-collection analysis, with Graham et al.’s relying on the power of database systems to build the LCN but avoids clustering analysis for HCCs, Giglietto et al.’s relies on R’s expressivity and filtering based on anomaly detection, while our implementation uses Python and batch mode processing to enable flexibility in the choice of cluster analysis technique. Pacheco et al.’s implementation is not available, but has been applied to very large datasets, and so may rely on a high performance programming language (e.g., Java) or distributed processing platform, such as Hadoop (2020b). Nizzoli et al. do not mention the availability of their implementation, only that their test dataset will be forthcoming (2020).
Our paper is the only one of these to address the concept of searching for multiple coordination criteria, and how to treat the combination of their evidence, and the attendant complications explored in Step 4 of Section 3.1. In fact, the other papers primarily treat the coordination criteria (i.e., user similarity measure) as entirely dependent on the current investigation and no generalisation of the concepts are discussed.
6 Future Work
The most important directions to take this work relate to the following aspects:
-
1.
Temporal analysis. Temporal analysis of the evolution of HCCs and their influence on the broader discussion over time will provide insight into how the HCCs operate and achieve their goals, and perhaps will reveal distinct classes of HCCs and employed strategies. Further theoretical research relying on the temporal aspect is also required to determine the real meaning of edges in the LCN. Figure 25 provides an illustration of where there is ambiguity. If accounts A, B, and C all retweet tweet 1 within a single time window, or at least overlapping time windows, then we join A to B and B to C in the LCN, and there is a reasonable assumption that A and C may be related somehow. This is less clear when A and B retweet tweet 1 and B and C retweet tweet 2, both in different time windows; it is much less reasonable to assume a relationship between A and C in that case (especially if the time windows are far apart), but both situations result in the same LCN structure: A connects to B and B connects to C.
(a) Situation 1. (b) Situation 2. Figure 25: The semantics of edges in LCNs requires clarification. If A is connected to B and B is connected to C, it may be due to the events in Situation 1 or Situation 2, but any inference of a relationship between A and C will be different depending on which it actually occurred. Each window, , starts at timestamp and has duration . An arrow between, e.g., A and the Twitter bird 1 implies that account A retweeted tweet 1 within a specific window. -
2.
Types of coordination. Not all coordination occurs in short periods of time nor in adjacent time windows. For example, an unscrupulous political campaign may purchase the services of bots to retweet a single campaign post once at a particular time to maximise its political effect. Alternatively, a small group of trolls may post the same tweets attacking a public figure each day at 5pm on weekends but not weekdays. Considering how to address broader definitions of coordination will be an ongoing challenge as OSNs change their features and people find new ways to use and abuse them.
-
3.
Distinguishing authentic and inauthentic behaviour. The issue of astroturfing (e.g., Metaxas and Mustafaraj, 2012) brings this into sharp relief: some campaigns are genuine grass roots movements driven by a broad desire to see policies change (e.g., campaigning to address climate change), however some are artificially organised, aimed at gaming OSN trending algorithms to spread their narrative further and to give the appearance of genuine wide public support (e.g., efforts to convince US Congress to release a politically controversial FBI memo, McKew, 2018). As covert campaigns become more sophisticated and numerous, it will become more important for OSN, law enforcement and relevant agencies to focus their efforts on the relevant activities and therefore to be able to dismiss harmless fan campaigns from the next disinformation campaign. Others have noted that this problem remains unresolved at this time (Vargas et al., 2020).
-
4.
Process improvement. Improving the implementation of the process, and how HCC extraction is performed, and how validation is conducted will be an ongoing activity. As demonstrated by Pacheco et al. (2020b) and Graham et al. (2020a), different types of community extraction will suit different types of coordination strategy, just as will the choice of strategy to search for (e.g., pollution or boost). Introducing a genuine sliding window (cf. our distinct, adjacent windows) will prevent missing further instances of coordination, but modification of the approach will be required to apply it in a near real-time setting (e.g., Carnein et al., 2017; Assenmacher et al., 2020). Finally, to bring some statistical robustness to the question of validity, there are measures that can be used to determine if the accounts in HCCs engage in statistically significant greater or lesser levels of, say, retweeting than the general population. These measures will help determine in what ways HCC behaviour differs, but will leave unresolved the question of intent and authenticity of that behaviour.
7 Conclusion
As online influence operations grow in sophistication, our automation and campaign detection methods must also expose the accounts covertly engaging in “orchestrated activities” (Grimme et al., 2018). We have described several coordination strategies, their purpose and execution methods, and demonstrated a novel pipeline-based approach to finding sets of accounts engaging in such behaviours in two relevant datasets. Using discrete time windows, we temporally constrain potentially coordinated activities, successfully identifying groups operating over various time frames. Guided by research questions posed in Section 1, our results were validated by using a variety of techniques, including developing three one-class classifiers to compare the HCCs found in two relevant datasets, plus a randomised one, with HCCs from a ground truth subset. Two case studies were also presented, in which our technique was applied to reveal insights into the activity of polarised groups in one and the activity of social bots and bot-like accounts in the other. The algorithmic complexity of our approach was discussed, as well as comparison with several similar contemporary approaches. The temporal analysis of HCC evolution and their impact on the broader discussion, theoretical questions of the semantics of edges in LCNs, the ability to distinguish between authentic and inauthentic coordinated behaviour, improvement of HCC extraction and validation techniques and application in near real-time processing environments all provide opportunities for future research in this increasingly important field.
Declarations
Funding
Not applicable.
Conflicts of interest/Competing interests
Not applicable.
Availability of data and material
The data (the identifiers of tweets only, as per Twitter’s terms and conditions) used in this study are available at https://github.com/weberdc/find_hccs.
Code availability
The data manipulation and analysis software written for this study is available at https://github.com/weberdc/find_hccs.
Authors’ contributions
Both authors contributed to the conception of the study, and Derek Weber performed the data collection, analysis and writing of the first draft. Both authors read and approved the final manuscript.
Ethics approval
All data were collected, stored, processed and analysed according to the ethics protocol H-2018-045, approved by the University of Adelaide’s human research and ethics committee.
Consent to participate
(include appropriate statements)
Consent for publication
All authors consent to this work being published.
References
- Adjali et al. (2020) Adjali O, Besançon R, Ferret O, Borgne HL, Grau B (2020) Multimodal entity linking for tweets. In: Lecture Notes in Computer Science, Springer International Publishing, pp 463–478, DOI 10.1007/978-3-030-45439-5˙31
- Alizadeh et al. (2020) Alizadeh M, Shapiro JN, Buntain C, Tucker JA (2020) Content-based features predict social media influence operations. Science Advances 6(30):eabb5824, DOI 10.1126/sciadv.abb5824
- Assenmacher et al. (2020) Assenmacher D, Adam L, Trautmann H, Grimme C (2020) Towards real-time and unsupervised campaign detection in social media. In: The Thirty Third International FLAIRS Conference, AAAI Publications
- Bacco et al. (2017) Bacco CD, Power EA, Larremore DB, Moore C (2017) Community detection, link prediction, and layer interdependence in multilayer networks. Physical Review E 95(4):042317, DOI 10.1103/physreve.95.042317
- Badawy and Ferrara (2018) Badawy A, Ferrara E (2018) The rise of jihadist propaganda on social networks. Journal of Computational Social Science 1(2):453–470, DOI 10.1007/s42001-018-0015-z
- Bedru et al. (2020) Bedru HD, Yu S, Xiao X, Zhang D, Wan L, Guo H, Xia F (2020) Big networks: A survey. Computer Science Review 37:100247, DOI 10.1016/j.cosrev.2020.100247
- Bessi and Ferrara (2016) Bessi A, Ferrara E (2016) Social bots distort the 2016 U.S. Presidential election online discussion. First Monday 21(11)
- Blondel et al. (2008) Blondel VD, Guillaume JL, Lambiotte R, Lefebvre E (2008) Fast unfolding of communities in large networks. Journal of Statistical Mechanics: Theory and Experiment 2008(10):P10008, DOI 10.1088/1742-5468/2008/10/p10008
- Borgatti et al. (2009) Borgatti SP, Mehra A, Brass DJ, Labianca G (2009) Network analysis in the social sciences. Science 323(5916):892–895, DOI 10.1126/science.1165821
- Brandes et al. (2007) Brandes U, Gaertler M, Wagner D (2007) Engineering graph clustering: Models and experimental evaluation. ACM Journal of Experimental Algorithmics 12:1.1:1–1.1:26, DOI 10.1145/1227161.1227162, URL https://doi.org/10.1145/1227161.1227162
- Brooking and Singer (2016) Brooking ET, Singer PW (2016) War goes viral: How social media is being weaponized across the world. The Atlantic URL https://www.theatlantic.com/magazine/archive/2016/11/war-goes-viral/501125/
- Burgess and Matamoros-Fernández (2016) Burgess J, Matamoros-Fernández A (2016) Mapping sociocultural controversies across digital media platforms: one week of #gamergate on twitter, YouTube, and Tumblr. Communication Research and Practice 2(1):79–96, DOI 10.1080/22041451.2016.1155338
- Cao et al. (2015) Cao C, Caverlee J, Lee K, Ge H, Chung J (2015) Organic or organized?: Exploring URL sharing behavior. In: Proceedings of the 24th ACM International on Conference on Information and Knowledge Management, ACM, pp 513–522, DOI 10.1145/2806416.2806572
- Carnein et al. (2017) Carnein M, Assenmacher D, Trautmann H (2017) Stream clustering of chat messages with applications to twitch streams. In: Lecture Notes in Computer Science, Springer International Publishing, pp 79–88, DOI 10.1007/978-3-319-70625-2˙8
- Carvin (2012) Carvin A (2012) Distant Witness: Social media, the Arab Spring and a journalism revolution. CUNY Journalism Press, New York, NY
- Chavoshi et al. (2017) Chavoshi N, Hamooni H, Mueen A (2017) Temporal patterns in bot activities. In: Companion Proceedings of the Web Conference 2017, ACM Press, pp 1601–1606, DOI 10.1145/3041021.3051114
- Chen (2015) Chen A (2015) The Agency. The New York Times Magazine URL https://www.nytimes.com/2015/06/07/magazine/the-agency.html
- Chu et al. (2012) Chu Z, Widjaja I, Wang H (2012) Detecting Social Spam Campaigns on Twitter. In: Applied Cryptography and Network Security, Springer Berlin Heidelberg, pp 455–472, DOI 10.1007/978-3-642-31284-7˙27
- Cresci (2020) Cresci S (2020) A decade of social bot detection. Communications of the ACM 63(10):72–83, DOI 10.1145/3409116
- Cresci et al. (2017) Cresci S, Pietro RD, Petrocchi M, Spognardi A, Tesconi M (2017) The Paradigm-Shift of Social Spambots. In: Companion Proceedings of the Web Conference 2017, ACM Press, pp 963–972, DOI 10.1145/3041021.3055135, arXiv:1701.03017
- Damashek (1995) Damashek M (1995) Gauging similarity with n-grams: Language-independent categorization of text. Science 267(5199):843–848, DOI 10.1126/science.267.5199.843
- Datta and Adar (2019) Datta S, Adar E (2019) Extracting inter-community conflicts in Reddit. In: ICWSM, AAAI Press, pp 146–157
- Davis et al. (2016) Davis CA, Varol O, Ferrara E, Flammini A, Menczer F (2016) BotOrNot: A system to evaluate social bots. In: Companion Proceedings of the Web Conference 2016, ACM Press, pp 273–274, DOI 10.1145/2872518.2889302
- Fang et al. (2019) Fang Y, Huang X, Qin L, Zhang Y, Zhang W, Cheng R, Lin X (2019) A survey of community search over big graphs. The VLDB Journal 29(1):353–392, DOI 10.1007/s00778-019-00556-x
- Ferrara (2017) Ferrara E (2017) Disinformation and social bot operations in the run up to the 2017 French presidential election. First Monday 22(8)
- Ferrara et al. (2016) Ferrara E, Varol O, Davis C, Menczer F, Flammini A (2016) The rise of social bots. Communications of the ACM 59(7):96–104, DOI 10.1145/2818717
- Fisher (2018) Fisher A (2018) Netwar in Cyberia: Decoding the media Mujahadin. CPD Perspectives Paper 5, USC Center on Public Diplomacy, URL https://uscpublicdiplomacy.org/sites/uscpublicdiplomacy.org/files/NetwarinCyberiaWebReady_withdisclosurepage11.08.18.pdf
- Fruchterman and Reingold (1991) Fruchterman TMJ, Reingold EM (1991) Graph drawing by force-directed placement. Software: Practice and Experience 21(11):1129–1164, DOI 10.1002/spe.4380211102
- Giglietto et al. (2019) Giglietto F, Righetti N, Marino G (2019) Understanding coordinated and inauthentic link sharing behavior on Facebook in the run-up to 2018 general election and 2019 European election in Italy. SocArvix DOI 10.31235/osf.io/3jteh
- Giglietto et al. (2020a) Giglietto F, Righetti N, Rossi L, Marino G (2020a) Coordinated Link Sharing Behavior as a signal to surface sources of problematic information on Facebook. In: International Conference on Social Media and Society, ACM, DOI 10.1145/3400806.3400817
- Giglietto et al. (2020b) Giglietto F, Righetti N, Rossi L, Marino G (2020b) It takes a village to manipulate the media: Coordinated link sharing behavior during 2018 and 2019 Italian elections. Information, Communication & Society pp 1–25, DOI 10.1080/1369118x.2020.1739732
- Graham and Ackland (2017) Graham T, Ackland R (2017) Do socialbots dream of popping the filter bubble? The role of socialbots in promoting participatory democracy in social media. In: Gehl RW, Bakardjieva M (eds) Socialbots and Their Friends: Digital Media and the Automation of Sociality, Routledge, London, chap 10, pp 187–206
- Graham and Keller (2020) Graham T, Keller TR (2020) Bushfires, bots and arson claims: Australia flung in the global disinformation spotlight. URL https://theconversation.com/bushfires-bots-and-arson-claims-australia-flung-in-the-global-disinformation-spotlight-129556, (Accessed on 2020-02-07)
- Graham et al. (2020a) Graham T, Ackland R, Mitchell L (2020a) A novel network-based approach to detecting and analysing coordinated inauthentic behaviour on Twitter. Talk presented at the Fifth Annual Australian Social Network Analysis Conference, Adelaide, South Australia, 25–27 November 2020, URL https://pypi.org/project/coordination-network-toolkit/
- Graham et al. (2020b) Graham T, Bruns A, Zhu G, Campbell R (2020b) Like a virus: the coordinated spread of coronavirus disinformation. Tech. rep., Centre for Responsible Technology, The Australia Institute, URL https://apo.org.au/node/305864
- Grimme et al. (2017) Grimme C, Preuss M, Adam L, Trautmann H (2017) Social bots: Human-like by means of human control? Big Data 5(4):279–293, DOI 10.1089/big.2017.0044
- Grimme et al. (2018) Grimme C, Assenmacher D, Adam L (2018) Changing perspectives: Is it sufficient to detect social bots? In: Meiselwitz G (ed) Lecture Notes in Computer Science, Springer International Publishing, Cham, pp 445–461, DOI 10.1007/978-3-319-91521-0˙32
- Gupta et al. (2019) Gupta S, Kumaraguru P, Chakraborty T (2019) MalReG: Detecting and analyzing malicious retweeter groups. In: Proceedings of the ACM India Joint International Conference on Data Science and Management of Data (CoDS-COMAD 2019), ACM, pp 61–69, DOI 10.1145/3297001.3297009
- Hegelich and Janetzko (2016) Hegelich S, Janetzko D (2016) Are social bots on Twitter political actors? Empirical evidence from a Ukrainian social botnet. In: ICWSM, AAAI Press, pp 579–582
- Hine et al. (2017) Hine GE, Onaolapo J, Cristofaro ED, Kourtellis N, Leontiadis I, Samaras R, Stringhini G, Blackburn J (2017) Kek, cucks, and God Emperor Trump: A measurement study of 4chan’s politically incorrect forum and its effects on the web. In: ICWSM, AAAI Press, pp 92–101
- Holme and Saramäki (2012) Holme P, Saramäki J (2012) Temporal networks. Physics Reports 519(3):97–125, DOI 10.1016/j.physrep.2012.03.001
- Howard and Kollanyi (2016) Howard PN, Kollanyi B (2016) Bots, #StrongerIn, and #Brexit: Computational propaganda during the UK-EU referendum. Research Note 2016.1, Oxford, UK: The Computational Propaganda Research Project, URL http://comprop.oii.ox.ac.uk/wp-content/uploads/sites/89/2016/06/COMPROP-2016-1.pdf
- Keller et al. (2017) Keller FB, Schoch D, Stier S, Yang J (2017) How to manipulate social media: Analyzing political astroturfing using ground truth data from South Korea. In: ICWSM, AAAI Press, pp 564–567
- Keller et al. (2019) Keller FB, Schoch D, Stier S, Yang J (2019) Political astroturfing on Twitter: How to coordinate a disinformation campaign. Political Communication 37(2):256–280, DOI 10.1080/10584609.2019.1661888
- King et al. (2017) King G, Pan J, Roberts ME (2017) How the Chinese Government Fabricates Social Media Posts for Strategic Distraction, Not Engaged Argument. American Political Science Review 111(3):484–501, DOI 10.1017/S0003055417000144
- Kumar et al. (2018) Kumar S, Hamilton WL, Leskovec J, Jurafsky D (2018) Community Interaction and Conflict on the Web. In: Proceedings of the 2018 World Wide Web Conference on World Wide Web - WWW’18, ACM Press, pp 933–943, DOI 10.1145/3178876.3186141
- Kuran and Sunstein (1999) Kuran T, Sunstein CR (1999) Availability cascades and risk regulation. Stanford Law Review 51(4):683, DOI 10.2307/1229439
- Lee et al. (2013) Lee K, Caverlee J, Cheng Z, Sui DZ (2013) Campaign extraction from social media. ACM TIST 5(1):9:1–9:28, DOI 10.1145/2542182.2542191
- Lim et al. (2019) Lim KH, Jayasekara S, Karunasekera S, Harwood A, Falzon L, Dunn J, Burgess G (2019) RAPID: Real-time Analytics Platform for Interactive Data Mining. In: Machine Learning and Knowledge Discovery in Databases, LNCS, vol 11053, Springer International Publishing, pp 649–653, DOI 10.1007/978-3-030-10997-4˙44
- Malone and Crowston (1994) Malone TW, Crowston K (1994) The interdisciplinary study of coordination. ACM Computing Surveys 26(1):87–119, DOI 10.1145/174666.174668
- Mariconti et al. (2017) Mariconti E, Onaolapo J, Ahmad SS, Nikiforou N, Egele M, Nikiforakis N, Stringhini G (2017) What’s in a name?: Understanding profile name reuse on Twitter. In: Proceedings of the 26th International Conference on World Wide Web, ACM, pp 1161–1170, DOI 10.1145/3038912.3052589
- Mariconti et al. (2019) Mariconti E, Suarez-Tangil G, Blackburn J, Cristofaro ED, Kourtellis N, Leontiadis I, Serrano JL, Stringhini G (2019) “You know what to do”: Proactive detection of YouTube videos targeted by coordinated hate attacks. Proceedings of the ACM on Human-Computer Interaction 3(CSCW):1–21, DOI 10.1145/3359309
- Mazza et al. (2019) Mazza M, Cresci S, Avvenuti M, Quattrociocchi W, Tesconi M (2019) RTbust. In: Proceedings of the 10th ACM Conference on Web Science - WebSci’19, ACM Press, DOI 10.1145/3292522.3326015
- McGregor (2014) McGregor A (2014) Graph stream algorithms: A survey. ACM SIGMOD Record 43(1):9–20, DOI 10.1145/2627692.2627694
- McKew (2018) McKew MK (2018) How Twitter bots and Trump fans made #ReleaseTheMemo go viral. Politico URL https://www.politico.eu/article/how-twitter-bots-and-trump-fans-made-releasethememo-go-viral/amp/, 4th February 2018
- Metaxas and Mustafaraj (2012) Metaxas PT, Mustafaraj E (2012) Social media and the elections. Science 338(6106):472–473, DOI 10.1126/science.1230456
- Mordelet and Vert (2014) Mordelet F, Vert JP (2014) A bagging SVM to learn from positive and unlabeled examples. Pattern Recognition Letters 37:201–209, DOI 10.1016/j.patrec.2013.06.010
- Morstatter et al. (2018) Morstatter F, Shao Y, Galstyan A, Karunasekera S (2018) From Alt-Right to Alt-Rechts: Twitter analysis of the 2017 German Federal Election. In: Companion Proceedings of the Web Conference 2018, ACM Press, pp 621–628, DOI 10.1145/3184558.3188733
- Nasim et al. (2018) Nasim M, Nguyen A, Lothian N, Cope R, Mitchell L (2018) Real-time detection of content polluters in partially observable Twitter networks. In: Companion Proceedings of the Web Conference 2018, ACM Press, pp 1331–1339, DOI 10.1145/3184558.3191574
- Neudert (2018) Neudert LMN (2018) Germany: A cautionary tale. In: Woolley SC, Howard PN (eds) Computational Propaganda: Political Parties, Politicians, and Political Manipulation on Social Media, Oxford University Press, chap 7, pp 153–184, DOI 10.1093/oso/9780190931407.003.0008
- Nizzoli et al. (2020) Nizzoli L, Tardelli S, Avvenuti M, Cresci S, Tesconi M (2020) Coordinated behavior on social media in 2019 UK general election. CoRR abs/2008.08370, 2008.08370v1
- Pacheco et al. (2020a) Pacheco D, Flammini A, Menczer F (2020a) Unveiling coordinated groups behind White Helmets disinformation. In: Companion Proceedings of the Web Conference 2020, ACM, pp 611–616, DOI 10.1145/3366424.3385775
- Pacheco et al. (2020b) Pacheco D, Hui PM, Torres-Lugo C, Truong BT, Flammini A, Menczer F (2020b) Uncovering coordinated networks on social media. CoRR abs/2001.05658
- Pedregosa et al. (2011) Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O, Blondel M, Prettenhofer P, Weiss R, Dubourg V, VanderPlas J, Passos A, Cournapeau D, Brucher M, Perrot M, Duchesnay E (2011) Scikit-learn: Machine Learning in Python. Journal of Machine Learning Research 12(85):2825–2830, URL http://dl.acm.org/citation.cfm?id=2078195
- Radicioni et al. (2020) Radicioni T, Pavan E, Squartini T, Saracco F (2020) Analysing Twitter semantic networks: The case of 2018 Italian elections. CoRR abs/2009.02960
- Ratkiewicz et al. (2011) Ratkiewicz J, Conover MD, Meiss MR, Gonçalves B, Flammini A, Menczer F (2011) Detecting and tracking political abuse in social media. In: ICWSM, AAAI Press
- Rizoiu et al. (2018) Rizoiu MA, Graham T, Zhang R, Zhang Y, Ackland R, Xie L (2018) #DebateNight: The role and influence of socialbots on Twitter during the 1st 2016 U.S. Presidential debate. In: ICWSM, AAAI Press, pp 300–309
- Şen et al. (2016) Şen F, Wigand R, Agarwal N, Tokdemir S, Kasprzyk R (2016) Focal structures analysis: Identifying influential sets of individuals in a social network. Social Network Analysis and Mining 6(1):17:1–17:22, DOI 10.1007/s13278-016-0319-z
- Shearer and Grieco (2019) Shearer E, Grieco E (2019) Americans are wary of the role social media sites play in delivering the news. Tech. rep., Pew Research Center
- Starbird and Wilson (2020) Starbird K, Wilson T (2020) Cross-Platform Disinformation Campaigns: Lessons Learned and Next Steps. Harvard Kennedy School Misinformation Review DOI 10.37016/mr-2020-002
- Starbird et al. (2019) Starbird K, Arif A, Wilson T (2019) Disinformation as collaborative work: Surfacing the participatory nature of strategic information operations. Proceedings of the ACM on Human-Computer Interaction 3(CSCW):1–26, DOI 10.1145/3359229
- Stilgherrian (2020) Stilgherrian (2020) Twitter bots and trolls promote conspiracy theories about Australian bushfires — ZDNet. URL https://www.zdnet.com/article/twitter-bots-and-trolls-promote-conspiracy-theories-about-australian-bushfires/
- Tversky and Kahneman (1973) Tversky A, Kahneman D (1973) Availability: A heuristic for judging frequency and probability. Cognitive Psychology 5(2):207–232, DOI 10.1016/0010-0285(73)90033-9
- Vargas et al. (2020) Vargas L, Emami P, Traynor P (2020) On the detection of disinformation campaign activity with network analysis. CoRR abs/2005.13466
- Varol et al. (2017) Varol O, Ferrara E, Menczer F, Flammini A (2017) Early detection of promoted campaigns on social media. EPJ Data Science 6(1):13, DOI 10.1140/epjds/s13688-017-0111-y
- Verma and Aggarwal (2020) Verma V, Aggarwal RK (2020) A comparative analysis of similarity measures akin to the jaccard index in collaborative recommendations: empirical and theoretical perspective. Social Network Analysis and Mining 10(1), DOI 10.1007/s13278-020-00660-9
- Vo et al. (2017) Vo N, Lee K, Cao C, Tran T, Choi H (2017) Revealing and detecting malicious retweeter groups. In: Proceedings of the 2017 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining 2017, ACM, pp 363–368, DOI 10.1145/3110025.3110068
- Wasserman and Faust (1994) Wasserman S, Faust K (1994) Social network analysis: Methods and applications, vol 8. Cambridge University Press, DOI 10.1017/cbo9780511815478
- Weber (2019) Weber D (2019) On coordinated online behaviour. Poster presented at the Fourth Australian Social Network Analysis Conference, ASNAC’19, 27–29 November, Adelaide, Australia, URL https://www.slideshare.net/derekweber/on-coordinated-online-behaviour
- Weber and Neumann (2020) Weber D, Neumann F (2020) Who’s in the gang? Revealing coordinating communities in social media. In: Proceedings of the 2020 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining 2020, ACM
- Weber et al. (2020) Weber D, Nasim M, Falzon L, Mitchell L (2020) #ArsonEmergency and Australia’s “Black Summer”: Polarisation and misinformation on social media. Lecture Notes in Computer Science pp 159–173, DOI 10.1007/978-3-030-61841-4˙11, arXiv:2004.00742
- Woolley (2016) Woolley SC (2016) Automating power: Social bot interference in global politics. First Monday 21(4), DOI 10.5210/fm.v21i4.6161
- Woolley and Guilbeault (2018) Woolley SC, Guilbeault DR (2018) United States: Manufacturing consensus online. In: Woolley SC, Howard PN (eds) Computational Propaganda: Political Parties, Politicians, and Political Manipulation on Social Media, Oxford University Press, chap 8, pp 185–211, DOI 10.1093/oso/9780190931407.001.0001
- Wu et al. (2018) Wu T, Wen S, Xiang Y, Zhou W (2018) Twitter spam detection: Survey of new approaches and comparative study. Computers & Security 76:265–284, DOI 10.1016/j.cose.2017.11.013
- Zhao et al. (2015) Zhao Q, Erdogdu MA, He HY, Rajaraman A, Leskovec J (2015) SEISMIC: A self-exciting point process model for predicting tweet popularity. In: Proceedings of the 21st ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, ACM, pp 1513–1522, DOI 10.1145/2783258.2783401