This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Well-Connected Communities in Real-World and Synthetic Networks

Minhyuk Park,1{}^{1\textsuperscript{\textdagger}} Yasamin Tabatabaee,1{}^{1\textsuperscript{\textdagger}} Vikram Ramavarapu,1{}^{1\textsuperscript{\textdagger}}
Baqiao Liu,1 Vidya K. Pailodi,1 Rajiv Ramachandran,1 Dmitriy Korobskiy,2
Fabio Ayres,3 George Chacko,1,4∗ Tandy Warnow1∗

1Department of Computer Science, University of Illinois Urbana-Champaign, IL 61801, USA
2NTT DATA, McLean, VA 22102, USA
3Insper Institute, São Paulo, Brazil
4Grainger College of Engineering, University of Illinois Urbana-Champaign, IL 61801, USA

To whom correspondence should be addressed; E-mail: [email protected]; [email protected]
Contributed equally to this manuscript

Integral to the problem of detecting communities through graph clustering is the expectation that they are “well connected”. In this respect, we examine five different community detection approaches optimizing different criteria: the Leiden algorithm optimizing the Constant Potts Model, the Leiden algorithm optimizing modularity, Iterative K-Core Clustering (IKC), Infomap, and Markov Clustering (MCL). Surprisingly, all these methods produce, to varying extents, communities that fail even a mild requirement for well connectedness. To remediate clusters that are not well connected, we have developed the “Connectivity Modifier” (CM), which, at the cost of coverage, iteratively removes small edge cuts and re-clusters until all communities produced are well connected. Results from real-world and synthetic networks illustrate a tradeoff users make between well connected clusters and coverage, and raise questions about the “clusterability” of networks and models of community structure.

Introduction

Community detection is of broad interest and is typically posed as a graph partitioning problem, where the input is a graph and the objective is a partitioning of its vertices into disjoint subsets, so that each subset represents a community  (?, ?, ?). The terms community and cluster overlap heavily, so we use them interchangeably herein. Our interest in community detection is for the purpose of identifying research communities from the global scientific literature, so we are especially focused on methods that can scale to large networks consisting of documents linked by citation  (?, ?, ?, ?).

A unifying definition of community does not exist but a general expectation is that the vertices within a community are better connected to each other than to vertices outside the community  (?), implying greater edge density within a community. However, a cluster may be dense while still having a small edge cut  (?). Therefore, the minimum edge cut size (min cut) for a community should not be small  (?). Thus, edge density and well-connectedness, i.e., not having a small edge cut, are two separable and expected properties of communities.

The Leiden algorithm  (?), which builds upon the Louvain algorithm  (?), is commonly used for community detection, with default quality function the Constant Potts Model (CPM)  (?). Clusters produced by CPM-optimization have the desirable property that if the edge cut splits the cluster into components AA and BB, then the edge cut will be at least r×|A|×|B|r\times|A|\times|B|  (?), where rr is a user-provided resolution parameter. This guarantee is strong when the edge cut splits a cluster into two components of approximately equal size, but is weaker when it produces an imbalanced split and weakest when the cut separates a single node from the remaining nodes in the cluster. Importantly, the guarantee depends on rr, and small values of rr produce weak bounds. Finally, we note that this guarantee applies to CPM-optimal clusterings but not to clusterings found by heuristics.

In using the Leiden software optimizing CPM, we observed that it produces clusters with small min cuts on seven different networks of varied origin ranging in size from approximately 34,000 to 75 million nodes. We also observed that the number of clusters with small min cuts increases as the resolution parameter decreases. Intrigued by this observation, we performed a broader study to evaluate the extent to which clusters produced by algorithms of interest meet even a mild standard for a well connected cluster.

To evaluate whether a cluster is well connected, we use a slow growing function f(n)f(n) so that a cluster with nn nodes whose min cut size is at most f(n)f(n) will not be considered well connected. By design, we ensure that (i) f(n)f(n) grows more slowly than the lower bound on the min cut size for clusters in CPM-optimal clusterings in the Leiden algorithm and (ii) that f(n)f(n) provides a meaningful lower bound on the small-to-moderate values of nn where the bound in  (?) is weak. We selected f(n)=log10nf(n)=\log_{10}n for this function.

We constructed min cut profiles from four additional clustering methods with different optimality criteria on the seven networks above: Leiden with modularity  (?) as quality function; the k-core based Iterative kk-core Clustering (IKC)  (?) using k=10k=10; and two flow-based methods, Infomap  (?) and Markov clustering (MCL)  (?). None of these methods offer guarantees of well connected clusters. While only IKC and Leiden optimizing either CPM or modularity scaled to the largest network we studied, all tested methods produced poorly connected clusters (i.e., clusters with min cuts of size at most f(n)f(n)) on these networks. These observations reveal a gap between the expectation of well connected clusters and what is actually being produced by these community finding methods. These findings also raise questions about the “clusterability”  (?) of networks and whether only portions of a network exhibit community structure.

For practical remediation, we developed a tunable meta-method, the Connectivity Modifier (CM). CM takes a clustering as input and returns well connected clusters that are also at least a minimum size BB, where the setting for BB as well as the definition of “well connected” can be set by the user (our default setting uses the function f(n)f(n) given above and B=11B=11). CM presently provides support for Leiden optimizing either CPM or modularity and IKC, the methods that scaled to the largest network we studied. On real-world networks, using CM in conjunction with the Leiden method produces well connected clusters but with a reduction in node coverage. We observed similar results of lower magnitude with IKC. Analyses of synthetic networks with ground truth communities show somewhat different trends, revealing intriguing differences between synthetic and real-world networks.

The rest of this manuscript is organized as follows. First, we present an initial study on a large citation network showing conditions under which Leiden clusters are not well connected. Next we present comparable results from additional methods and networks. We then describe the design of Connectivity Modifier (CM) and show the impact of using CM on clusterings by Leiden and IKC on real world and synthetic networks. Finally, we close with a discussion of our findings.

Results

Refer to caption
Refer to caption
Figure 1: Node coverage, connectivity, and size distribution of clusters generated by Leiden optimizing either CPM or modularity on the Open Citations network (75,025,194 nodes). Connectivity (y-axis) is the minimum edge cut size (min cut) of each cluster. Node coverage, the percentage of nodes in clusters of at least size 11, is reported in blue text. Results are shown for clusters of at least size 11 from Leiden optimizing either CPM at five different resolution values or modularity. Higher node coverage is associated with reduced connectivity. Within each clustering, larger clusters are better connected although lower resolution values and modularity trend towards larger clusters that are less well connected.

In a study of the Open Citations network  (?) consisting of 75,025,194 nodes, we computed the min cut of all clusters generated using the Leiden algorithm optimizing either CPM or modularity (Fig. 1). Designating singleton clusters and very small clusters as not being of practical interest, we report node coverage throughout this manuscript as the percentage of nodes in clusters of size at least 11; the exception to this is Figure 4, where we report node coverage based on non-singleton clusters. For CPM, we used five different resolution values (0.5,0.10,0.01,0.001,0.5,0.10,0.01,0.001, and 0.00010.0001) that resulted in node coverage values ranging from 6-99%.

For Leiden-CPM clusterings, we see that as the resolution value is decreased, (i) node coverage increases, (ii) the frequency of small mincuts increases, and (iii) cluster sizes increase (Fig. 1). Clustering under modularity is most similar to clustering under CPM at the lowest resolution value used. Strikingly, 98.7% of 2,184 clusters produced under modularity had a minimum edge cut size of 1 (i.e., could be split by removing a single edge) while accounting for 99.4% node coverage. Additionally, where node coverage is high, clusters tend to be larger and fewer although intermediate resolution values in the range we used result in an increase in the number of clusters (Fig. 1 and Table S1). Thus, these data illustrate a tradeoff that users make with the Leiden algorithm between small clusters, lower node coverage, and few small min cuts (achieved by CPM-optimization with larger resolution values) versus larger clusters, higher node coverage, and many more small min cuts (achieved by modularity-optimization or CPM-optimization with small resolution values).

While a large fraction of small min cuts intuitively signals “poorly-connected” clusters, there are different ways of formalizing this notion. Here we offer a formal definition for the purposes of this study. Briefly, we consider functions f(n)f(n) with the interpretation that if a cluster of size nn has an edge cut of size at most f(n)f(n) then the cluster will be considered poorly connected. We want f(n)f(n) to grow very slowly so that it serves as a mild bound. We also want f(n)1f(n)\geq 1 for all nn that are large enough for the cluster to be considered a potential community. From three examples of slowly growing functions (Materials and Methods), we choose f(n)=log10nf(n)=\log_{10}n, the function that imposes the mildest constraint on large clusters and grows more slowly than the bound in  (?).

In addition to the Open Citations network, we clustered six other networks, ranging in size from 34,546 nodes to 13,989,436 nodes, with Leiden, IKC, Infomap, and MCL and computed the percentage of clusters that we consider well connected (i.e., whose min cuts were greater than f(n)f(n)). Under the conditions used, Leiden and IKC ran to completion on all seven networks, although IKC did not return any clusters from wiki_talk because a 10-core does not exist in this sparse network. Infomap failed on the largest network, and MCL returned output only from the smallest network (cit_hepph) we analyzed (Fig. 2).

For Leiden clustering optimizing CPM (Fig. 1), the frequency of well connected clusters decreases with resolution value, and results from modularity are similar to the lowest resolution value for CPM that was tested. IKC completed on all networks returning well connected clusters that varied between 85.9% and 94% of the total number of clusters. In comparison to Leiden, IKC clustering resulted in lower node coverage, which is consistent with its more conservative formulation. Infomap produced well connected clusters varying from 5% (orkut) to 92.4% (cit_patents). MCL ran only on the cit_hepph network with 81.3% of the clusters being well connected. Interestingly, while neither Leiden nor IKC generated disconnected clusters, Infomap generated disconnected clusters for some networks (maximized at 75% for orkut) and MCL also generated disconnected clusters (7.2%) on the cit_hepph network. These observations reveal the widespread existence of clusters that are not well connected, with the extent dependent on clustering method and network.

Refer to caption
(a) Leiden CPM & Modularity
Refer to caption
(b) IKC, Infomap, MCL
Figure 2: Percentage of Well Connected Clusters. Clustering of seven networks by five different community finding approaches. Networks analyzed are Open Citations (75,025,194), Curated Exosome Network (13,989,436), cit_hepph (34,546), and cit_patents (3,774,768) are citation networks; orkut (3,072,441) is a social network; wiki_talk (2,394,385) and wiki_topcats (1,791,489) are Wikipedia communication and hyperlink networks respectively. Only Leiden and IKC ran to completion on all networks although IKC did not return any clusters from the wiki_talk network. Infomap completed on all but Open Citations. MCL completed only on cit_hepph.

Towards well connected clusters, we designed the Connectivity Modifier (CM)  (?, ?) a remediation tool that can be used to modify a given clustering to ensure that each final cluster is well connected (Materials and Methods). Based on our preliminary findings, we chose to initially evaluate CM paired with the Leiden and IKC methods since these two clustering methods were sufficiently scalable and did not produce disconnected clusters. We also restricted our attention to the two largest networks, Open Citations and CEN.

We implemented CM in a pipeline (Fig. 3) that presently takes a Leiden or IKC clustering as input. A pre-processing (filtering) step discards very small clusters, i.e., those of size at most 10, but this bound can be changed by the user. Tree clusters are also discarded in this pre-processing step (given our definition of f(n)f(n), any tree of size 10 or larger is not well connected). CM then iteratively computes and removes any min cuts of size at most f(n)f(n) and re-clusters until only well connected clusters remain. A post-processing step removes any small clusters of size at most 10 that may have resulted from repeated cutting.

Refer to caption
Figure 3: Connectivity Modifier Pipeline Schematic. The four-stage pipeline depends on user-specified algorithmic parameters: BB, the minimum allowed size of a cluster, and f(n)f(n), a bound on the minimum edge cut size for a cluster with nn nodes, and clustering method. Stage 1: a clustering is computed. Stage 2: clusters are pre-processed by removing trees and those clusters of size less than BB. Stage 3: the CM is applied to each cluster, removing edge cuts of sizes at most f(n)f(n), reclustering, and recursing. Stage 4: clusters are post-processed by removing those of size less than BB. All clusters returned are well connected according to f(n)f(n) and have size at least BB. Our study explored default settings with B=11B=11 and f(n)=log10nf(n)=\log_{10}n.

.

Effect of CM on node coverage

We assessed the impact of CM on node coverage, here specifically examining clusters of size at least two. As above, we examined Leiden clustering of Open Citations and CEN networks using CPM-optimization at five resolution values and also using modularity, examining the change in node coverage before and after CM treatment (Fig. 4). The results are resolution dependent and network sensitive. For CPM-optimization, the impact of filtering out clusters of size at most 10 is large for the two larger resolution values, but then decreases as the resolution value decreases. In contrast, the impact of the Connectivity Modifier (the middle component of the CM pipeline that iteratively finds and removes small edge cuts and reclusters) also depends on the resolution parameter, with a minimal impact for the large resolution values and an increasing impact as the resolution value decreases. Modularity returned results most similar to CPM-optimization with the smallest tested resolution value. Since the pre-processing (filtering) and post-processing both remove all clusters of size at most 10, the node coverage reported for these stages are with respect to clusters of size at least 11. Given this, we observe that post-CM node coverage is low compared to pre-CM for both networks and clustering methods, and was smallest when using CPM-optimization with resolution value r=0.5r=0.5 and largest when using CPM-optimization with one of the two smallest resolution values, r=0.001r=0.001 for CEN and r=0.0001r=0.0001 for Open Citations. Overall, post-CM node coverage of any Leiden clustering never exceeded 24.6% for CEN and 68.7% for Open Citations.

Refer to caption
(a) Open Citations
Refer to caption
(b) CEN
Figure 4: Reduction in node coverage after CM treatment of Leiden clusters. The Open Citations (left panel) and CEN (right panel) networks were clustered using the Leiden algorithm under CPM at five different resolution values or modularity. Node coverage (defined as the percentage of nodes in cluster of size at least 2) was computed for Leiden clusters (lime green), Leiden clusters with trees and/or clusters of size 10 or less filtered out (soft orange), and after CM treatment of filtered clusters (desaturated blue).

Cluster fate

To understand the nature of the modifications effected by CM, we further classified the Leiden clusters based on the impact of CM-processing: extant, reduced, split, and degraded, where “extant” indicates that the cluster was not modified by CM, “reduced” indicates that the cluster is reduced in size, “split” indicates that the cluster was divided into at least two smaller clusters, and “degraded” indicates that the cluster was reduced to singletons or a cluster of size at most 10. We report the fraction in each category (Fig. 5) for six different clustering conditions on the Open Citations and CEN networks. When using CM with CPM-optimization with a very large resolution value, most clusters are extant, the fraction of extant clusters decreases for CPM-optimization as the resolution value decreases, and is low also for modularity (Fig. 2).

An interesting trend with split clusters is seen in Figure 5, indicating the initial cluster contained two or more well connected clusters; this is similar to the observation by Fortunato and Barthélemy in  (?) of modularity’s behaviour on a ring-of-cliques, where modularity returns clusters that contain two or more cliques, instead of returning the individual cliques, under some conditions. Here we see that clusters that are split by CM unsurprisingly occur in both networks with modularity, but also occur in a noticeable way for CPM-optimization with the smallest resolution value for the Open Citations network. Finally, we note that this occurs for CPM-optimization with other resolution values on both networks (Table S5 and Figure S1). The most extreme cluster fate is of course when it is degraded to singletons, which occurs in modularity clusterings and CPM-based clusterings at lower resolution values; however, the degree to which this occurs depends on the network and resolution value in a way that does not suggest any particular pattern (Figure S1).

Refer to caption
Figure 5: Cluster Fate for Leiden. CM-processed clusters are classified as extant (unaffected by CM), reduced (replaced by a single smaller cluster), split (replaced by two or more clusters), or degraded (replaced entirely by singletons). Top: Cluster Fate for Open Citations (OC); Bottom: Cluster fate for the CEN. Data are shown for CM treatment of Leiden clusters, under either CPM for 5 different resolution values or modularity. Cluster counts are expressed as a percentage of the input clusters.

Clustering with IKC

We examine the impact of CM on IKC (with k=10k=10) on the Open Citations and CEN networks (Fig. 6). In comparison to Leiden, IKC-clustering results in relatively low node coverage, 23.6% and 3.8% in the case of the Open Citations and CEN networks respectively. CM treatment of these clusterings also has a small effect on node coverage. We also see that most clusters are extant and some are split, but none are reduced or degraded.

In summary, the response to CM-processing differed across methods. On the CEN and Open Citations networks, and for both IKC and Leiden under modularity or CPM, node coverage in a post-CM clustering did not exceed 68%, either because CM-processing reduces node coverage substantially from the initial clustering in the case of Leiden with modularity or CPM-clustering with smaller resolution values or because the initial clustering was conservative and already had low node coverage. This suggests the possibility that in these real-world networks, only a fraction of the nodes may be in clusters that are sufficiently well connected and sufficiently large. In other words, real-world networks may not be fully covered by what we consider “valid” communities.

Refer to caption
(a) Node coverage
Refer to caption
(b) Cluster Fate
Figure 6: Node coverage (a) and cluster fate (b) for IKC clusters modified by CM on the Open Citations and CEN. For IKC on these two networks, node coverage is small on both the CEN and Open Citations, and slightly reduced by CM on the Open Citations network. We also see that most clusters are extant (unaffected by CM-processing) and some are split (replaced by two or more smaller clusters), but none are reduced (replaced by a single smaller cluster) or degraded (replaced by only singletons).

Analysis of synthetic networks

Refer to caption
Figure 7: The effect on node-coverage (percentage of nodes in clusters of size at least 11) produced by the CM processing pipeline on LFR and real-world networks. Each row corresponds to a real-world network, and each column corresponds to a Leiden clustering (either modularity or CPM-optimization with the specified resolution value). Within each entry, we show node coverage for the Leiden clustering as it goes through the CM pipeline; results on the empirical network are shown in green and results for the LFR network are shown in red. “Pre-CM” indicates the output of the Leiden clustering, “filter” is after removing trees and clusters below size 11, and “post-CM” is after the entire pipeline (which also removes any final clusters below size 11). No results are shown for some combinations (specifically, data are not shown for LFR for any clustering of the Orkut network nor for the wiki_topcats and wiki_talk networks with resolution value r=0.5r=0.5) due to LFR failing to generate networks for those settings.
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 8: Impact of CM-processing on accuracy of synthetic networks. The left panels show accuracy measured in terms of NMI and AMI, with respect to the LFR ground-truth communities. Each condition on the x-axis corresponds to a different LFR network, generated based on Leiden-modularity or Leiden-CPM with that specific resolution parameter. In total, there are 34 different LFR networks. The right panels show percentage of the LFR ground-truth communities that are small or disconnected. In most conditions, CM improves the accuracy of the original Leiden clustering, except for some of the conditions when the ground-truth communities have many (at least 60%60\%) disconnected clusters, or the node coverage by clusters of size at least 11 is low (at most 70%).

Given our hypothesis that the large reduction in node coverage that we observe on real-world networks may be the result of them not being universally covered by well connected true communities, we examined synthetic networks, noting that synthetic networks generated using LFR software  (?) assign every node to a non-singleton ground-truth community. If node coverage after CM-processing of LFR networks were similar to real-world networks, it would argue against this hypothesis. Examining synthetic networks also enables us to evaluate the impact of CM-processing on accuracy.

For this experiment, we computed statistics for the Leiden clusterings of the 7 real-world networks we explored, Open Citations, CEN, and 5 networks from the SNAP collection  (?)), and used them as input to the LFR software  (?) (see Materials and Methods). Using this software, we were not able to build synthetic networks for the CEN and Open Citations with the number of nodes in the empirical networks; therefore, for these two specific cases, we constructed the LFR networks to 33M nodes. For some combinations of network and Leiden clustering, we were unable to produce any synthetic networks, for example, we were unable to produce LFR networks for the Orkut social network. We produced a collection of 34 LFR networks with ground truth communities that had similar empirical statistics as their corresponding real-world networks (Supplementary Materials Section S3). We clustered each of these 34 LFR networks using the same clustering method used to provide empirical statistics to LFR.

Results in Figure 7 show node coverage after clustering for both the empirical network and its corresponding LFR network. The general trends we observed in previous experiments for OC and CEN also hold for the five other empirical networks examined here both with respect to node coverage in Leiden clustering and how CM-processing impacts node coverage. An interesting trend we see here is that node coverage drops for some CPM clusterings of empirical networks though not for LFR networks after the filtering stage. Since node coverage in this figure is with respect to clusters of size 11 or larger, any drop in node coverage resulting from filtering is due only to removing trees. The large drop in node coverage for Leiden with CPM-optimization in these cases means that CPM-optimization in some conditions produces a large number of tree clusters. We note this did not occur for resolution value r=0.5r=0.5, where clusters tend to be small but did occur for other resolution values (Supplementary Tables S3 and S4).

Of specific interest is whether clusterings of the LFR networks respond similarly to CM-processing as clusterings of their corresponding empirical networks, as this addresses our hypothesis regarding why CM-processing impacts node coverage in empirical networks. While node coverage does drop to the same degree for some LFR networks as for the empirical networks on which they are based, there are many cases where node coverage drops much more for the real-world network than for the corresponding LFR network, and no cases where there is a bigger drop in node coverage for the LFR network than for the real-world network. Thus, the idea that real-world networks not having universal coverage by valid communities cannot be ruled out.

We examined the impact of CM-processing on clustering accuracy; results for Normalized Mutual Information (NMI) and Adjusted Mutual Information (AMI) are shown in Figure 8, and results for the Adjusted Rand Index (ARI) are similar to AMI and are provided in Supplementary Materials, Figure S11  (?, ?, ?). We see that CM-processing improves NMI accuracy for modularity and also for CPM-optimization when used with small resolution values. CM-processing tends to be neutral for NMI otherwise, and was only detrimental on two network:clustering pairs. The impact on AMI accuracy is more variable. For example, CM-processing reduced AMI accuracy for all wiki_talk network:clustering pairs except for CPM-optimization with r=0.1r=0.1 where accuracy was very low and the impact was neutral. CM-processing also reduced accuracy for CPM-optimization on some network:clustering pairs for large resolution values.

However, when examining the properties of the ground-truth clusterings of these networks, we see that the cases where CM-processing produced a noteworthy reduction in accuracy for NMI or AMI are those where there are many disconnected ground truth clusters, for example all wiki_talk clusters, or where the node coverage by clusters of size at least 11 is small, for example cit_patents at the three largest resolution values and wiki_topcats with the two largest resolution values. However, there are conditions with low node coverage or a large fraction of disconnected clusters where CM-processing is neutral or even beneficial.

The occurrence of disconnected ground-truth clusters in the LFR networks is striking and problematic, since a basic expectation of a community is that it is connected, if not well-connected  (?). Hence, we assert that it is unreasonable to evaluate accuracy with respect to a ground-truth set of communities if the communities are not connected. In other words, it does not make sense to evaluate clustering accuracy for those LFR networks that contain many disconnected ground truth communities, including the entire set of LFR networks constructed for wiki_talk invalid, which is one of the conditions where CM-processing reduces AMI accuracy. The fact that LFR networks had ground truth clusters that were not connected also indicates the failure of LFR software to reproduce features of the input network:clustering pairs, which by construction always have 100% of the clusters connected.

It is easy to see why a low node coverage by clusters of size at least 11 could reduce accuracy for CM-processing, since CM automatically removes all small clusters. However, this property depends on the network:clustering pair, which depends on network features as well as the clustering methodology. In this study, clustering real-world networks produced a large fraction of small clusters when we used CPM-optimization with large resolution values; the conditions where CM-processing reduced AMI accuracy on the LFR networks with low node coverage by clusters of size at least 11 are drawn from those conditions. We also note that users typically select the clustering method that produces cluster sizes that match their interest. Therefore, CM-processing will not be beneficial where there is interest in recovering small communities unless the bound BB is replaced by a smaller value.

In interpreting these results, we also note the discrepancy between some empirical statistics of LFR networks and those of the real-world network:clustering pairs that were used to simulate the LFR network. Beyond incomplete matching of features, differences such as the frequency of disconnected clusters and the percentage of clusters that are small make it questionable whether accuracy on LFR networks is suggestive of accuracy on real-world networks.

Discussion

In this study, we considered the question of whether clusters produced by community detection methods are well connected. We use a mild definition to demonstrate that five different clustering paradigms generate, to varying extents, output clusters that are not well connected. An important implication of these results is that portions of a network may not exhibit community structure. Further, at parameter settings that maximize node coverage, weakly connected parts of a network may be forced into communities. Related prior work  (?, ?) address whether a graph in general is clusterable, which is a related question.

We developed CM to convert, through partitioning, poorly connected clusters into well connected ones. For flexibility, the function in CM that defines “well connected” can be modified by users to be more or less stringent. Similarly, we provide a parameter BB, tunable by the user, that specifies the smallest size of a cluster for it to be retained; this too can be modified by the user to be more or less stringent. At present, CM provides support only for Leiden and IKC, the most scalable of the methods we tested. CM is being extended to provide support for Infomap and MCL and concurrently being redesigned for developers to integrate their own clustering methods into it. CM allows the user to explore cluster quality in the input, as it reveals which clusters are poorly connected, and, in some cases, finds substructure within clusters. Thus, CM-processing can be used to evaluate and improve clustering outputs, and interrogate and explore the community structure within a given network.

Several factors affect how significantly CM-processing changes a given clustering. These include the network itself, as some networks seem to be more impacted by CM-processing. We also see that the choice of resolution parameter for Leiden-CPM has an influence on how much the clustering changes, with generally larger impact for small resolution values.

The findings that LFR networks produce different patterns than empirical networks is perhaps not surprising, but here we consider potential explanations. First, the LFR methodology assumes that the degree distribution and cluster size distributions follow a power law, however, it is not clear that this is universal to citation and real-world networks  (?, ?, ?, ?). Furthermore, we also observed that the degree distribution and cluster size distributions were imperfectly fit by LFR software in our study (Supplementary Materials, Figs. S2 and S3). Hence, the assumptions about node degree and cluster size distributions that govern the LFR model may not result in adequate simulation of real-world networks.

Another assumption in the LFR methodology is that every node is in a community, which is one that would benefit from deeper investigation. Intuitively, we posit that the assumption that every node in a real world community is in a community may only be reasonable if the communities can be small and/or poorly connected.

We recognize that our perspective on well connected clusters may result in narrow descriptions of communities, perhaps akin to cores of core-periphery structures  (?, ?, ?). Additionally, informative weaker links  (?) may be lost from communities since CM partitions input clusters that are poorly connected into sets of well connected clusters. However, such weak links are not lost from the network being analyzed.

Having developed CM and ascertained its quantitative effects, a direction for our future work is to evaluate it in the context of specific evaluations that are supported by mixed methods. More generally, we emphasize leaving the definition, use, and interpretation of well connected to users.

Materials and Methods

Defining well connected clusters.

Refer to caption
Figure 9: Comparison of the lower bounds on edge cut sizes for defining well connected clusters. Here we compare four functions; t(n)=0.01(n1)t(n)=0.01(n-1) is Traag’s function, which provides a lower bound of the cut size for a cluster on nn nodes in a CPM-optimal clustering when r=0.01r=0.01, f(n)=log10nf(n)=\log_{10}n, g(n)=log2ng(n)=\log_{2}n, and h(n)=n5h(n)=\frac{\sqrt{n}}{5}. For n1000n\geq 1000, Traag’s function provides the largest lower bound on the edge cut size, and so is the strongest guarantee of the functions we compare, while f(n)f(n) provides the smallest lower bound for all n239n\geq 239. However, for n238n\leq 238, f(n)t(n)f(n)\geq t(n), so that f(n)f(n) provides a stronger guarantee on these small- to moderate-sized clusters than Traag’s bound.

In  (?), Traag et al. proved that every cluster CC in a CPM-optimal clustering of a network (with resolution parameter rr) would satisfy the following property: any edge cut XX of CC splitting the nodes into two sets AA and BB would have at least r×|A|×|B|r\times|A|\times|B| edges. This function is maximized when |A|=|B||A|=|B| and minimized when |A|=1|A|=1 and |B|=n1|B|=n-1, where CC has nn nodes. Hence in particular, this establishes that the minimum cut for any cluster with nn nodes has size at least r×(n1)r\times(n-1).

We have already observed that this bound (which we refer to as Traag’s bound) is weak when nn is not large and rr is very small. For example, when r=0.01r=0.01 and n=50n=50, the bound only establishes that the minimum edge cut is at least 11, which therefore simply asserts that the cluster is connected. Hence, we seek lower bounds on the size of the minimum edge cut that are larger than Traag’s bound of r×(n1)r\times(n-1) for small values of nn but grow very slowly, and so do not exceed Traag’s bound for larger nn.

In the text, we discussed the bound that requires that for a cluster of nn nodes to be well connected, the size of its min cut must be strictly greater than f(n)=log10nf(n)=\log_{10}n. We compare f(n)f(n) to Traag’s lower bound when r=0.01r=0.01, which we refer to as t(n)t(n), as well as two other functions, g(n)=log2ng(n)=\log_{2}n and h(n)=n5h(n)=\frac{\sqrt{n}}{5}; note that each of f(n),g(n)f(n),g(n) and h(n)h(n) is strictly positive and increasing, and yet grows more slowly than Traag’s function t(n)t(n).

The comparison between these functions in the range of cluster sizes 1n20001\leq n\leq 2000 is shown in Figure 9. As desired, for large enough nn, Traag’s function t(n)t(n) dominates the other functions, thus imposing a much stronger guarantee on the minimum cut size for the cluster. We also note that f(n)<g(n)f(n)<g(n) for all n>1n>1, but that the relationship between f(n),h(n)f(n),h(n), and t(n)t(n) depends on nn. For large nn, however, f(n)f(n) is less than all the other functions, making it the most slowly growing function.

Based on this comparison, we used f(n)f(n) to define when a cluster of nn nodes is well connected: the min cut size is strictly greater than log10(n)\log_{10}(n); otherwise we say that the cluster is poorly connected.

Note that if we had picked g(n)g(n) instead, we would have considered more clusters poorly connected, since f(n)<g(n)f(n)<g(n) for all nn. Similarly, picking h(n)h(n) would have generally been more stringent a requirement (at least for all but very small values of nn), and so would have ruled more clusters as being poorly connected. Thus, f(n)f(n) represents a very modest constraint on the size of the min cut for a cluster, in order for it to qualify as being well connected.

The Connectivity Modifier (CM) Pipeline

To remediate poorly-connected clusters, we developed a modular pipeline that we now describe (Fig. 2). By design, this pipeline is guaranteed to return a clustering where each cluster is well connected according to the function f(n)f(n) and has size at least BB. The values of f(n)f(n) and B used in this study are set as default but can be easily modified by a user. Note that if an input cluster meets these two criteria, it will be present in the output clustering. Furthermore, every cluster in the output will either be one of the input clusters or will be a subset of one of the input clusters.

The CM pipeline requires the user to specify values for three algorithmic parameters:

  • BB, the minimum allowed size of any “valid” vertex community (default B=11B=11)

  • f(n)f(n), so that a cluster is well connected only if its min cut size exceeds f(n)f(n) (default: f(n)=log10nf(n)=\log_{10}n)

  • A clustering method, presently selected from Leiden optimizing CPM, Leiden optimizing modularity, or the Iterative k-core (IKC) method with k=10k=10. Support for additional methods is in active development.

The input to the CM pipeline is a network 𝒢\mathcal{G} with NN nodes and the algorithmic parameters as specified. The pipeline then operates in four stages:

  • Stage 1: A clustering is generated from the input network 𝒢\mathcal{G}.

  • Stage 2: The clustering is filtered to remove clusters below size BB and all trees, which are not well-connected when using f(n)=log10nf(n)=\log_{10}n

  • Stage 3: The Connectivity Modifier (CM) is applied to each cluster that remains. First, all nodes of degree at most f(n)f(n) are removed (where nn is the number of nodes in the cluster), until there are no low degree nodes remaining. Then, for each remaining cluster, a min cut is calculated using VieCut  (?), and if it is not greater than f(n)f(n) in size then the min cut is removed, thus splitting the cluster into two components. These components are then re-clustered using the selected clustering method, and the process repeats until all clusters are well-connected.

  • Stage 4: Any resultant clusters below size BB are removed.

Additional details for the Connectivity Modifier

The Connectivity Modifier was developed using Python 3.10. The original implementation of VieCut  (?) in C++17 was wrapped into a Python package using PyBind11. In our analysis we use unbalanced minimum cuts computed with the NOI algorithm  (?).

Additional details for IKC

Due to memory constraints, IKC fails to run with the connectivity modifier on the entirety of the OpenCitations network. To solve this, we modified IKC to run in memory as an imported Python module rather than as a separate executable. Moreover, we split the initial IKC clustering of OpenCitations into separate clusters, and the Connectivity Modifier was then run on the largest clusters independently (see Supplementary Section S2 for full details).

Data

A custom-implemented Extract-Transform-Load (ETL) process was designed to process the publicly available Open Citations dataset, downloaded in Aug 2022, and load it into a PostgreSQL table. The resultant network contained 75,025,194 nodes and 1,363,605,603 edges. The CEN is a citation network constructed from the literature on exosome research. From the SNAP repository, we downloaded cit_hepph, an Arxiv High Energy Physics paper citation network; cit_patents, a citation network among US patents; orkut, a social media network; wiki_talk, a network containing users and discussion from the inception of Wikipedia until January 2008; and wiki_topcats, a web graph of Wikimedia hyperlinks. Before clustering, all networks were processed to remove self-loops as well as duplicate and parallel edges.

network nodes edges avg_deg ref
Open Citations 75,025,194 1,363,605,603 36.35  (?)
CEN 13,989,436 92,051,051 13.16  (?)
cit_hepph 34,546 420,877 24.37  (?)
cit_patents 3,774,768 16,518,947 8.75  (?)
orkut 3,072,441 117,185,083 76.28  (?)
wiki_talk 2,394,385 4,659,565 3.89  (?)
wiki_topcats 1,791,489 25,444,207 28.41  (?)
Table 1: Summary statistics for networks used in this study. Average degree is the average of the node degrees across the network.

LFR (synthetic) networks

To create simulated networks with ground truth communities, while attempting to emulate the properties of each empirical network and its corresponding Leiden clustering, we used the LFR software from  (?). The generative model of the LFR graphs assume that the node degree and the community size distributions are power-law distributions  (?). The software for generating LFR benchmark graphs  (?) takes the following eight parameters as input:

  • Network properties: Number of nodes NN, average and maximum node degrees (kk and kmaxk_{max} respectively), and negative exponent for degree sequence (τ1\tau_{1}) that is assumed to be a power-law.

  • Community properties: Maximum and minimum community sizes (cmaxc_{max} and cminc_{min}), and negative exponent for the community size distribution (τ2\tau_{2}), also modeled as a power-law.

  • Mixing parameter μ\mu, that is the ratio between the degree of a node outside its community and its total degree, averaged over all nodes in the network. Lower μ\mu values suggest that a network consists of well-separated communities, as nodes are mostly connected to other nodes inside their communities, rather than outside of it.

Parameter Estimation.

To emulate the empirical networks using LFR graphs, we estimated all eight parameters described above for a given pair of network 𝒢\mathcal{G} and a clustering 𝒞\mathcal{C}. Computing N,k,kmax,cminN,k,k_{max},c_{min} and cmaxc_{max} is straightforward using networkX  (?, ?). To estimate μ\mu, we perform a single iteration over all edges of the network, and for each edge, if the nodes on the two sides of it were in different communities, that edge contributes to the ratio μ\mu of these two nodes. The total μ\mu of the network:clustering pair is the average μ\mu across all the nodes.

To estimate τ1\tau_{1} and τ2\tau_{2}, we fit a power-law distribution to the node degree sequence and the community size distribution, using the approach from  (?) that is implemented in the power-law Python package  (?). Note that the power-law property may hold for the tail of the degree or community size sequence and not the whole distributions. Therefore, following  (?), we estimate xminx_{min}, the minimum value for which the power-law property holds as well as the exponent α\alpha for the tail of the distribution.

Generating LFR networks.

After computing these parameters based on the Leiden clusterings of the empirical networks using both modulatiry and CPM with a range of resolution parameters, we simulated LFR networks using the software from  (?), producing empirical statistics reported in Supplementary Tables S7 and S8.

For networks with more than 10 million nodes, i.e., Open Citations and the CEN, we limited the number of vertices to 3 million, due to scalability limitations of the LFR benchmark graph generator  (?), while preserving the edge density reflected by average degree, and the mixing parameter. The numbers of nodes of the other LFR graphs exactly match the number of nodes in the corresponding empirical network. In some cases, due to the inherent limitations of the LFR graph generator, we had to modify the ranges of the community sizes, i.e., increase cminc_{min} and decrease cmaxc_{max}, to generate the network. These values are available with the datasets.

As shown in the statistics reported in Supplementary Tables S7 and S8, the average node degree, the mixing parameter, and the exponents for the degree and community size distributions are in all cases very well-preserved. Further details about the pipeline for producing LFR graphs and the statistics of the graphs are provided in Supplementary Materials Section 3. We calculated NMI, AMI, and ARI using the Python Scikit-Learn package  (?). Other software used includes Leiden and IKC  (?, ?).

References

  • 1. M. E. Newman, M. Girvan, Finding and evaluating community structure in networks. Physical review E 69, 026113 (2004).
  • 2. P. J. Mucha, T. Richardson, K. Macon, M. A. Porter, J.-P. Onnela, Community structure in time-dependent, multiscale, and multiplex networks. Science 328, 876–878 (2010).
  • 3. S. Fortunato, M. E. J. Newman, 20 years of network community detection. Nature Physics 18, 848–850 (2022).
  • 4. E. Wedell, M. Park, D. Korobskiy, T. Warnow, G. Chacko, Center-periphery structure in research communities. Quantitative Science Studies 3, 289–314 (2022).
  • 5. K. W. Boyack, R. Klavans, Creation of a highly detailed, dynamic, global model and map of science. Journal of the Association for Information Science and Technology 65, 670–685 (2013).
  • 6. L. Waltman, N. J. van Eck, A new methodology for constructing a publication-level classification system of science. Journal of the American Society for Information Science and Technology 63, 2378–2392 (2012).
  • 7. K. W. Boyack, D. Newman, R. J. Duhon, R. Klavans, M. Patek, J. R. Biberstine, B. Schijvenaars, A. Skupin, N. Ma, K. Börner, Clustering more than two million biomedical publications: Comparing the accuracies of nine text-based similarity approaches. PLoS ONE 6, e18029 (2011).
  • 8. M. Coscia, F. Giannotti, D. Pedreschi, A classification for community discovery methods in complex networks. Statistical Analysis and Data Mining 4, 512–546 (2011).
  • 9. F. Bonchi, D. García-Soriano, A. Miyauchi, C. E. Tsourakakis, Finding densest k-connected subgraphs. Discrete Applied Mathematics 305, 34–47 (2021).
  • 10. V. A. Traag, L. Waltman, N. J. Van Eck, From Louvain to Leiden: guaranteeing well-connected communities. Scientific reports 9, 1–12 (2019).
  • 11. V. D. Blondel, J.-L. Guillaume, R. Lambiotte, E. Lefebvre, Fast unfolding of communities in large networks. Journal of Statistical Mechanics: Theory and Experiment 2008, P10008 (2008).
  • 12. V. A. Traag, P. V. Dooren, Y. Nesterov, Narrow scope for resolution-limit-free community detection. Physical Review E 84 (2011).
  • 13. M. Rosvall, C. T. Bergstrom, Maps of random walks on complex networks reveal community structure. Proceedings of the National Academy of Sciences 105, 1118–1123 (2008).
  • 14. S. V. Dongen, Graph clustering via a discrete uncoupling process. SIAM Journal on Matrix Analysis and Applications 30, 121–141 (2008).
  • 15. P. Miasnikof, A. Y. Shestopaloff, A. Raigorodskii, Statistical power, accuracy, reproducibility and robustness of a graph clusterability test. International Journal of Data Science and Analytics 15, 379–390 (2023).
  • 16. S. Peroni, D. Shotton, OpenCitations, an infrastructure organization for open scholarship. Quantitative Science Studies 1, 428–444 (2020).
  • 17. V. Ramavarapu, F. Ayres, M. Park, V. K. Pailodi, G. Chacko, T. Warnow, Connectivity modifier, https://github.com/illinois-or-research-analytics/cm_pipeline (2023).
  • 18. B. Liu, M. Park, Connectivity modifier, https://github.com/RuneBlaze/connectivity-modifier (2022).
  • 19. S. Fortunato, M. Barthelemy, Resolution limit in community detection. Proceedings of the National Academy of Sciences 104, 36–41 (2007).
  • 20. A. Lancichinetti, S. Fortunato, F. Radicchi, Benchmark graphs for testing community detection algorithms. Physical review E 78, 046110 (2008).
  • 21. J. Leskovec, A. Krevl, SNAP Datasets: Stanford large network dataset collection (2014). Http://snap.stanford.edu/data.
  • 22. L. Danon, A. Diaz-Guilera, J. Duch, A. Arenas, Comparing community structure identification. Journal of statistical mechanics: Theory and experiment 2005, P09008 (2005).
  • 23. N. X. Vinh, J. Epps, J. Bailey, Proceedings of the 26th annual international conference on machine learning (2009), pp. 1073–1080.
  • 24. L. Hubert, P. Arabie, Comparing partitions. Journal of classification 2, 193–218 (1985).
  • 25. C. Gao, J. Lafferty, Testing for global network structure using small subgraph statistics (2017).
  • 26. F. Radicchi, S. Fortunato, C. Castellano, Universality of citation distributions: Toward an objective measure of scientific impact. Proceedings of the National Academy of Sciences 105, 17268–17272 (2008).
  • 27. M. J. Stringer, M. Sales-Pardo, L. A. N. Amaral, Statistical validation of a global model for the distribution of the ultimate number of citations accrued by papers published in a scientific journal. Journal of the American Society for Information Science and Technology 61, 1377–1385 (2010).
  • 28. I. Artico, I. Smolyarenko, V. Vinciotti, E. C. Wit, How rare are power-law networks really? Proceedings of the Royal Society A 476, 20190742 (2020).
  • 29. M. Brzezinski, Power laws in citation distributions: evidence from scopus. Scientometrics 103, 213–228 (2015).
  • 30. R. Breiger, Explorations in Structural Analysis (RLE Social Theory) (Routledge, 2014).
  • 31. P. Rombach, M. A. Porter, J. H. Fowler, P. J. Mucha, Core-periphery structure in networks (revisited). SIAM Review 59, 619–646 (2017).
  • 32. M. S. Granovetter, The strength of weak ties. American Journal of Sociology 78, 1360–1380 (1973).
  • 33. M. Henzinger, A. Noe, C. Schulz, D. Strash, Practical minimum cut algorithms. ACM Journal of Experimental Algorithmics 23 (2018).
  • 34. H. Nagamochi, T. Ono, T. Ibaraki, Implementing an efficient minimum capacity cut algorithm. Math. Program. 67, 325–341 (1994).
  • 35. A. Jakatdar, B. Liu, T. Warnow, G. Chacko, AOC: Assembling overlapping communities. Quantitative Science Studies 3, 1079–1096 (2022).
  • 36. J. Leskovec, J. Kleinberg, C. Faloutsos, Proceedings of the eleventh ACM SIGKDD international conference on Knowledge discovery in data mining (ACM, 2005), pp. 177–187.
  • 37. J. Yang, J. Leskovec, Defining and evaluating network communities based on ground-truth. Knowledge and Information Systems 42, 181–213 (2013).
  • 38. J. Leskovec, D. Huttenlocher, J. Kleinberg, Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (ACM, 2010), pp. 1361–1370.
  • 39. H. Yin, A. R. Benson, J. Leskovec, D. F. Gleich, Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (ACM, 2017), pp. 555–564.
  • 40. R. Albert, A.-L. Barabási, Statistical mechanics of complex networks. Reviews of modern physics 74, 47 (2002).
  • 41. S. Fortunato, Resources (2023). https://www.santofortunato.net/resources.
  • 42. A. Hagberg, P. Swart, D. S Chult, Exploring network structure, dynamics, and function using networkx, Tech. rep., Los Alamos National Lab.(LANL), Los Alamos, NM (United States) (2008).
  • 43. Y. Tabatabaee, Emulating real networks using LFR graphs, https://github.com/ytabatabaee/emulate-real-nets (2023).
  • 44. A. Clauset, C. R. Shalizi, M. E. Newman, Power-law distributions in empirical data. SIAM review 51, 661–703 (2009).
  • 45. J. Alstott, E. Bullmore, D. Plenz, powerlaw: a python package for analysis of heavy-tailed distributions. PloS one 9, e85777 (2014).
  • 46. G. M. Slota, J. W. Berry, S. D. Hammond, S. L. Olivier, C. A. Phillips, S. Rajamanickam, Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis (2019), pp. 1–14.
  • 47. F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, et al., Scikit-learn: Machine learning in python. the Journal of machine Learning research 12, 2825–2830 (2011).
  • 48. V. Traag, Leiden algorithm: leidenalg, https://github.com/vtraag/leidenalg (2019).
  • 49. E. Wedell, M. Park, Iterative k-core software, https://github.com/chackoge/ERNIE_Plus/blob/master/Illinois/clustering/eleanor/code/IKC.py (2021).

Acknowledgements: The authors thank Nathan Bryans, Christine Ballard, and Bryan Barker from Oracle Research for their assistance in setting up and using the Oracle Cloud Infrastructure,
Funding: This work was supported in part by Insper-Illinois Collaboration and by a Research Award to TW from Oracle Research..
Author Contributions GC and TW conceived the research; GC, TW, and YT designed the analyses. VR, FA, MP, BL, YT, VP, DK, and RR developed software to support analyses. MP, YT, GC, VR, VP conducted the analyses. GC, TW, YT, and MP interpreted the data and wrote the manuscript.
Competing Interests The authors declare that they have no competing financial interests.
Data and materials availability: Additional data and materials are available online.