Unsupervised Graph Neural Network Reveals the Structure–Dynamics Correlation in Disordered Systems
Abstract
Learning the structure–dynamics correlation in disordered systems is a long-standing problem. Here, we use unsupervised machine learning employing graph neural networks (Gnn) to investigate the local structures in disordered systems. We test our approach on 2D binary A65B35 LJ glasses and extract structures corresponding to liquid, supercooled and glassy states at different cooling rates. The neighborhood representation of atoms learned by a Gnn in an unsupervised fashion, when clustered, reveal local structures with varying potential energies. These clusters exhibit dynamical heterogeneity in the structure in congruence with their local energy landscape. Altogether, the present study shows that unsupervised graph embedding can reveal the structure–dynamics correlation in disordered structures.
I Introduction
Disordered structures exist in distinct phases such as liquid, supercooled and glassy states [1, 2]. Liquids, when cooled fast enough to avoid crystallization, form glasses [2]. While liquids and supercooled liquids are equilibrium states, glasses are considered as metastable state, which continuously relaxes [2]. The cooling pathway, such as the rate and the pressure, directly affects the glass structure and dynamics [3, 4, 5, 6, 7, 8]. Thus, although glasses can be considered as “frozen liquids”, the dynamics of glasses are notably different from that of liquids. However, it is still debated whether the dynamics of these disordered systems are encoded precisely in their local structure.
Conventional approaches on understanding the structure-dynamics relationships rely on identifying structural signatures such as medium range crystalline orders (MRCO) that relate to the dynamical heterogeneity in disordered systems [9, 10, 11]. However, attempts to predict the dynamical heterogeniety directly from the structure have been barely successful, although indirect correlations have been established [12, 13, 14, 15, 16, 17]. Recently, data-driven approaches employing machine learning (ML) was used to predict atomic dynamics directly from the structure. Specifically, “softness” [18, 19], a machine learned parameter, was introduced to predict the tendency of atoms to rearrange purely based on their local structure in a disordered system. Further, Bapst et al. [20] used a graph neural network (Gnn) to learn the structural descriptors capable of predicting propensity [21, 15] of atoms to move under shear stress. Although both these works [18, 20] confirmed that the dynamical heterogeneity could be predicted directly from the structure, they employed a supervised ML approach [22, 23], where the models were trained explicitly to learn the descriptors that predict the dynamics of atoms. In other words, the data on both structure and dynamics was used to learn their interrelationship through supervised ML, which was then used to model unseen systems. This infuses an inherent bias to learn the features that best represent the individual dynamics of atoms. Hence, the descriptors may not be suitable for other tasks such as predicting the collective dynamics such as self-organization or even learning the local motifs in glass structures [24, 25]. Further, training the model require a priori knowledge of both the structure and the dynamics, which may not be available for many systems.
Here, using Gnns, we learn the structural features of disordered systems directly in an unsupervised fashion [26]. Specifically, by considering disordered structures as graphs [27], we allow the model to learn local atomic environment in an unsupervised fashion. We show that the Gnn is able to identify the features that, in turn, discovers local clusters in the disordered structures reminiscent of the MRCO. We test our approach on a well established 2D binary LJ glass of composition A65B35 to form glasses from the melt at different cooling rates ranging over four orders of magnitude. The clusters discovered by the Gnn exhibit a good correlation in terms of their dynamics and energy, revealing the existence of a structure–dynamics correlation.
II Methodology
To learn the local topology of disordered systems, we use Gnns. Figs. 1(a)-(b) describe the transformation from disordered structure to a graph. First, we transform each of the disordered structures as an undirected graph [27] . The node set and edge set represent the atoms and the set of bonds between these atoms, respectively. Here, the bonds of the atoms are with their first neighbors with the cutoff distance as the first minima of the partial pair distribution functions (PDFs). Thus, represents the set of nodes that are neighbors to the node in graph . Further, each node consists of the following node attributes (): (1) one-hot encoding representing the atom (node) type (i.e., type A or B). (2) Coordinates of the node in two dimensions .

Our goal is to understand the local structure of such disordered systems. Towards this end, we utilize GraphSage [26] to learn the node representation in a way that captures topological structure of each node’s neighborhood (see Fig. 1(c),(d)). Hence, for each node , we learn how to aggregate information from its neighboring nodes as
(1) |
where represents the raw feature information of nodes comprising of the atom type and coordinates, represents the node information after rounds of message passing with the neighbors, and represents an activation function which in our case is the sigmoid function.
The purpose of the Gnn is to construct representations for each node integrating both the nodes’ initial raw feature information and the information about the local graph structure that surrounds them. Specifically, each node aggregates its own information and information from its neighborhood to update its representation, i.e., where denotes the depth of the Gnn and denotes the final representation of the node . The term is a learnable parameter of the Gnn at depth . In the case of a disordered atomic structure, learns to aggregate feature information from neighboring atoms and pass this information to generate an improved representation for the central atom incorporating the local neighborhood features. With a higher depth , nodes incrementally gain more and more information from further reaches of the graph. For instance, aggregates information from the hop neighborhood of the node . However, higher values of in Gnns, such as could lead to poor performance [28].
Next, we train the Gnn by minimizing an objective function. Since our aim is to learn the local features of the structure without any additional information, we resort to unsupervised learning. In order to train the model, we generate a fixed-length () random walk [29, 26] from each node in the graph, which visits different nodes in ’s neighborhood. This random walk allows the exploration of the local structure surrounding node . Fig. 1(c) depicts 2 random walks of length 3 from a node . Note that in contrast to previous approaches, here we are not forcing the Gnn to learn any task-specific embedding such as predicting the dynamics. The embeddings are purely a representation of the local neighborhood of a node and hence is purely structural in origin. In other words, the representation learned by Gnn is purely based on the atomic structure and has no bias toward their dynamics.
Through these random walks, we identify nodes that co-occur in ’s neighborhood. Without loss of generality, we consider these node pairs as positive samples, that is representing the neighborhood of . Node pairs that do not co-occur in random walks are considered as negative samples, which is representative of a random node absent from ’s neighborhood. Next, our unsupervised objective function given by Eq. 2 is optimized to learn the parameters of the Graph neural network.
(2) |
Here, , where is the number of layers of the Gnn and determines the number of negative samples from the negative distribution (consisting negative samples). The minimization of the objective function encourages nodes that are in same neighborhood (positive samples) to have similar representation and vice-versa. As evident from Eq. 2, the dot-product (hence, similarity) of representations of positive node pairs that co-occur in random walks is encouraged to be higher, while the opposite is true for the representation of negative node pairs .
III Results and Discussion
We test our approach on a 2D binary LJ system. We perform MD simulation of melt-quenching LJ system [30] of 500 atoms with composition A65B35 to prepare glasses at different cooling rates. All the simulations are carried out using the LAMMPS package [31]. The interaction between the particles is governed by
(3) |
where refers to the distance between two particles, is the distance at which inter-particle potential energy is zero and refers to the depth of the potential well. Here, we use the LJ parameters , , , , and . The mass for all particles is set to . All the quantities are expressed in reduced units with respect to , , and Boltzmann constant . We set the interaction cutoff [32] and the time step for simulations.
We perform all the simulations at constant volume. For preparing the glass structure, first the ensemble is equilibrated at high temperature where it equilibrates in the liquid state. Next, it is cooled down to low temperature of , at different cooling rates ranging from to . The cooling is carried out at NVT, using Nose-Hoover thermostat in temperature steps of . At each of these steps, the atomic structure of the system along with their density and enthalpy is noted. In order to avoid the noise due to thermal fluctuations, the glass structures are minimized at 0 atm pressure to obtain the ground state enthalpy and density [32]. Accordingly, we obtain minimized glass structures, from each cooling rate, representing the ground state structures at different temperatures and states.
To train the Gnn, we set , representing the information aggregation from up to the neighbors. We observe that higher values of could accumulate noise, which is analogous to the PDF in disordered systems which at larger distances saturates to 1. The choice of also resulted in learning better representations as evidenced by the performance on the validation set as discussed in Fig. 10 in section S4 in the Appendix. For performing random walk, we set and and number of random walks starting from each node to be . Effect of random walk length parameter is shown in section S7 of Appendix. We set the size of node representation to be . Finally, we use the training graphs to optimize the parameters of Gnn. Using the learnt parameters of Gnn, we generate node representations for graphs that were not seen during the training phase. The parameters used for training Gnn are specified in detail in section S4 in Appendix.


First, we analyse the change in the enthalpy and density of the disordered structures, when they are cooled at different rates. Figs. 2(a) and (b) represent the variation in enthalpy and density of the structures upon quenching. In agreement with theoretical understanding [32, 5, 4, 8], we observe that lower cooling rates lead to more stable glasses with lower enthalpy and higher densities. We also observe that the density and enthalpy of the structures in the liquid () and supercooled () regimes are indistinguishable, despite the difference in their cooling rates—thanks to their equilibrium nature [2]. In contrast, the solid (glassy, ) state exhibits clear variation for the structures obtained with different cooling rates. Note that the enthalpy of a system have major contributions from the short-range interactions in a structure, while the density could be attributed to the medium- and long-range order of the system. Accordingly, we observe that the disordered structures simulated have notable differences both in terms of their short-range interactions and the long-range order.
Next, in Fig. 3 we compare the total and partial PDFs of the structures obtained at different temperatures and cooling rates. Since liquid and supercooled structures does not show any significant difference of cooling rate (see Fig. 2), only one PDF is plotted representing them. We observe that despite the denser and more stable structure (in terms of enthalpy) of glass obtained at lower cooling rate, the total PDFs obtained are almost identical. Except a minor broadening of the first peak of B-B partial PDFs, the partial PDFs of different glassy structures are also comparable. This results shows that merely distance distribution cannot capture the effect of cooling rate on the local structure. Note that extensive studies have been carried out to understand the local structure of LJ glasses with different cooling rates and readers are referred to these works for further differences and similarities on the local structure with respect to cooling rate [5].
Towards our goal of gaining better understanding of the different local regions in the disordered structure (graph), we use the unsupervised graph embeddings learned by the Gnn. Specifically, we cluster the nodes of a graph based on their node representations using a density-based clustering algorithm OPTICS [33] and cosine similarity as our distance function between two nodes.
Specifically, nodes (atoms) in the graph are ordered such that spatially closest points become neighbors in the ordering. Then based upon a distance parameter, representing the density, a decision is made to accept/reject points to be part of a cluster. Details of the OPTICS clustering algorithms and hyper-parameters used are given in section S2,S3 and S5 in Appendix.

Now, we analyze the LJ system when atoms are clustered based upon the approach described above. Fig. 4(a) shows the potential energy per atom for an arbitrary glass structure from the simulations. We observe that the energy distribution seems to be random with no local order. In contrast, Fig. 4(b) shows the average potential energy of the clusters obtained from OPTICS clustering of node representations generated by the trained Gnn. Note that nodes colored in grey are not part of any cluster. i.e., they are outliers. Additionally, we also show the reachability distance plot generated by OPTICS in section S6 in Appendix. Fig. 4(c) shows the cluster size and average potential energy of the clusters. We clearly observe several high and low energy clusters in the structure. It can be observed from the limits of color bar in Fig. 4(b) that the variation in average potential energies of the obtained clusters, although small in comparison to that of individual atoms, is of higher range than that observed in the cooling curves of Fig. 2. Interestingly, the local clusters observed are reminiscent of the MRCO observed in supercooled and glassy systems based on orientational order parameters [9, 10, 11]. This suggests that the unsupervised embeddings learned by the Gnn is able to capture the local structural order (and disorder) in these systems.
In order to analyze whether the learnt cluster correlates to the dynamical heterogenity observed in these systems, we study the curvature of the local energy landscape of the clusters following an established methodology [34, 2]. First, we perform multiple energy minimizations to ensure that the samples are at their ground state. Then, we provide a sudden energy bump to the system corresponding to a temperature of . The temperature is chosen to be high enough to allow potential motion between low-energy barriers but low enough to avoid glass transition or melting of the system. Finally, we allow the system to evolve in the micro-canonical ensemble (NVE) for 100 thousand steps. The MSD of a cluster is obtained as
(4) |
where represents the number of atoms in a cluster, are the positions of each atom after 100 thousand steps of dynamics following the energy bump, and are the positions in the inherent structure.
Stable low-energy structures represented by narrow and deep local energy landscape exhibit low MSD, while high energy structures, exhibiting a shallow and broad landscape, exhibit high MSD. To understand the correlation between the clusters identified by Gnn and dynamical heterogeniety, we plot the MSD of the clusters with respect to the average potential energy of the clusters (see Fig. 4(d)). Note that each data point in the plot corresponds to a cluster, and the color represent the state (liquid, supercooled, solid) from which it was obtained. We observe a direct positive correlation with the cluster potential energy and the MSD with a Pearson coefficient of 0.747, representing high statistical significance. The variations in the cluster MSD even within a given disordered structure (see Fig. 5 in section S1 of Appendix), in congruence with the local potential energy, suggests the existence of dynamical heterogeneity in these systems. These heterogenieties are captured effectively by the graph embeddings.
IV Conclusion
Altogether, present study reveals a novel method, employing unsupervised Gnn, for learning the local structure in disordered systems. We demonstrate that these local structures exhibit a direct correlation with the dynamical heterogenieties present in the disordered systems. It is worth noting that in contrast to previous approaches, where ML was trained against a data to predict dynamics from structure, the present approach reveal local structural motifs in an unbiased manner purely from the structure. Further analysis of these clusters could potentially reveal hidden order within a disordered structure, that, in turn, govern their dynamics. In addition, the learned embedding could further be potentially used as a representation or a reduced-order feature of the disordered system to predict their properties. Our approach thus paves way for gaining an understanding on the structure–dynamics correlation in disordered systems that are not achieved by conventional methods.
The code is available at : https://github.com/M3RG-IITD/ML-DOS
References
- Angell [1995] C. A. Angell, Formation of glasses from liquids and biopolymers, Science 267, 1924 (1995).
- Debenedetti and Stillinger [2001] P. G. Debenedetti and F. H. Stillinger, Supercooled liquids and the glass transition, Nature 410, 259 (2001).
- Vollmayr et al. [1996a] K. Vollmayr, W. Kob, and K. Binder, Cooling-rate effects in amorphous silica: A computer-simulation study, Phys. Rev. B 54, 15808 (1996a).
- Li et al. [2017] X. Li, W. Song, K. Yang, N. A. Krishnan, B. Wang, M. M. Smedskjaer, J. C. Mauro, G. Sant, M. Balonis, and M. Bauchy, Cooling rate effects in sodium silicate glasses: Bridging the gap between molecular dynamics simulations and experiments, The Journal of chemical physics 147, 074501 (2017).
- Vollmayr et al. [1996b] K. Vollmayr, W. Kob, and K. Binder, How do the properties of a glass depend on the cooling rate? a computer simulation study of a lennard-jones system, The Journal of chemical physics 105, 4714 (1996b).
- Moynihan et al. [1976] C. T. Moynihan, A. J. Easteal, M. A. De Bolt, and J. Tucker, Dependence of the fictive temperature of glass on cooling rate, Journal of the American Ceramic Society 59, 12 (1976).
- Smedskjaer et al. [2014] M. M. Smedskjaer, R. E. Youngman, S. Striepe, M. Potuzak, U. Bauer, J. Deubener, H. Behrens, J. C. Mauro, and Y. Yue, Irreversibility of pressure induced boron speciation change in glass, Scientific reports 4, 1 (2014).
- Bhaskar et al. [2020] P. Bhaskar, R. Kumar, Y. Maurya, R. Ravinder, A. R. Allu, S. Das, N. N. Gosvami, R. E. Youngman, M. S. Bødker, N. Mascaraque, et al., Cooling rate effects on the structure of 45s5 bioglass: Insights from experiments and simulations, Journal of Non-Crystalline Solids 534, 119952 (2020).
- Watanabe and Tanaka [2008] K. Watanabe and H. Tanaka, Direct observation of medium-range crystalline order in granular liquids near the glass transition, Physical review letters 100, 158002 (2008).
- Kawasaki et al. [2007] T. Kawasaki, T. Araki, and H. Tanaka, Correlation between dynamic heterogeneity and medium-range order in two-dimensional glass-forming liquids, Physical review letters 99, 215701 (2007).
- Tah et al. [2018] I. Tah, S. Sengupta, S. Sastry, C. Dasgupta, and S. Karmakar, Glass transition in supercooled liquids with medium-range crystalline order, Physical review letters 121, 085703 (2018).
- Coslovich and Pastore [2006] D. Coslovich and G. Pastore, Are there localized saddles behind the heterogeneous dynamics of supercooled liquids?, EPL (Europhysics Letters) 75, 784 (2006).
- Widmer-Cooper et al. [2004] A. Widmer-Cooper, P. Harrowell, and H. Fynewever, How reproducible are dynamic heterogeneities in a supercooled liquid?, Physical review letters 93, 135701 (2004).
- Widmer-Cooper and Harrowell [2006] A. Widmer-Cooper and P. Harrowell, Predicting the long-time dynamic heterogeneity in a supercooled liquid on the basis of short-time heterogeneities, Physical review letters 96, 185701 (2006).
- Widmer-Cooper and Harrowell [2007] A. Widmer-Cooper and P. Harrowell, On the study of collective dynamics in supercooled liquids through the statistics of the isoconfigurational ensemble, The Journal of chemical physics 126, 154503 (2007).
- Cubuk et al. [2015] E. D. Cubuk, S. S. Schoenholz, J. M. Rieser, B. D. Malone, J. Rottler, D. J. Durian, E. Kaxiras, and A. J. Liu, Identifying structural flow defects in disordered solids using machine-learning methods, Physical review letters 114, 108001 (2015).
- Malins et al. [2013] A. Malins, J. Eggers, H. Tanaka, and C. P. Royall, Lifetimes and lengthscales of structural motifs in a model glassformer, Faraday discussions 167, 405 (2013).
- Schoenholz et al. [2016] S. S. Schoenholz, E. D. Cubuk, D. M. Sussman, E. Kaxiras, and A. J. Liu, A structural approach to relaxation in glassy liquids, Nature Physics 12, 469 (2016).
- Schoenholz et al. [2017] S. S. Schoenholz, E. D. Cubuk, E. Kaxiras, and A. J. Liu, Relationship between local structure and relaxation in out-of-equilibrium glassy systems, Proceedings of the National Academy of Sciences 114, 263 (2017), https://www.pnas.org/doi/pdf/10.1073/pnas.1610204114 .
- Bapst et al. [2020] V. Bapst, T. Keck, A. Grabska-Barwińska, C. Donner, E. D. Cubuk, S. S. Schoenholz, A. Obika, A. W. Nelson, T. Back, D. Hassabis, et al., Unveiling the predictive power of static structure in glassy systems, Nature Physics 16, 448 (2020).
- Berthier and Jack [2007] L. Berthier and R. L. Jack, Structure and dynamics of glass formers: Predictability at large length scales, Physical Review E 76, 041509 (2007).
- LeCun et al. [2015] Y. LeCun, Y. Bengio, and G. Hinton, Deep learning, nature 521, 436 (2015).
- Carleo et al. [2019] G. Carleo, I. Cirac, K. Cranmer, L. Daudet, M. Schuld, N. Tishby, L. Vogt-Maranto, and L. Zdeborová, Machine learning and the physical sciences, Reviews of Modern Physics 91, 045002 (2019).
- Jones and Stevanović [2020] E. B. Jones and V. Stevanović, The glassy solid as a statistical ensemble of crystalline microstates, npj Computational Materials 6, 1 (2020).
- Boolchand et al. [2005] P. Boolchand, G. Lucovsky, J. Phillips, and M. Thorpe, Self-organization and the physics of glassy networks, Philosophical Magazine 85, 3823 (2005).
- Hamilton et al. [2017] W. Hamilton, Z. Ying, and J. Leskovec, Inductive representation learning on large graphs, Advances in neural information processing systems 30 (2017).
- Trudeau [2013] R. J. Trudeau, Introduction to graph theory (Courier Corporation, 2013).
- Liu et al. [2020] M. Liu, H. Gao, and S. Ji, Towards deeper graph neural networks, in Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining (2020) pp. 338–348.
- Aldous and Fill [1995] D. Aldous and J. Fill, Reversible markov chains and random walks on graphs (1995).
- Brüning et al. [2008] R. Brüning, D. A. St-Onge, S. Patterson, and W. Kob, Glass transitions in one-, two-, three-, and four-dimensional binary lennard-jones systems, Journal of Physics: Condensed Matter 21, 035117 (2008).
- Thompson et al. [2022] A. P. Thompson, H. M. Aktulga, R. Berger, D. S. Bolintineanu, W. M. Brown, P. S. Crozier, P. J. in ’t Veld, A. Kohlmeyer, S. G. Moore, T. D. Nguyen, R. Shan, M. J. Stevens, J. Tranchida, C. Trott, and S. J. Plimpton, LAMMPS - a flexible simulation tool for particle-based materials modeling at the atomic, meso, and continuum scales, Comp. Phys. Comm. 271, 108171 (2022).
- Singh et al. [2013] S. Singh, M. D. Ediger, and J. J. De Pablo, Ultrastable glasses from in silico vapour deposition, Nature materials 12, 139 (2013).
- Ankerst et al. [2008] M. Ankerst, M. Breunig, H. Kriegel, R. Ng, and J. Sander, Ordering points to identify the clustering structure, in Proc. ACM SIGMOD, Vol. 99 (2008).
- Krishnan et al. [2017] N. M. A. Krishnan, B. Wang, Y. Yu, Y. Le Pape, G. Sant, and M. Bauchy, Enthalpy landscape dictates the irradiation-induced disordering of quartz, Phys. Rev. X 7, 031019 (2017).
- Ester et al. [1996] M. Ester, H.-P. Kriegel, J. Sander, X. Xu, et al., A density-based algorithm for discovering clusters in large spatial databases with noise., in kdd, Vol. 96 (1996) pp. 226–231.
- Ankerst et al. [1999] M. Ankerst, M. M. Breunig, H.-P. Kriegel, and J. Sander, Optics: Ordering points to identify the clustering structure, ACM Sigmod record 28, 49 (1999).
- [37] OPTICS Clustering Scikit-Learn, https://scikit-learn.org/stable/modules/generated/sklearn.cluster.OPTICS.html/, [Online; accessed 07-June-2022].
Supplementary Material
S1. Cluster mean squared displacement in 2D LJ System

In Fig. 5(a) and (b) we show the obtained mean squared displacement of clusters in our 2D LJ system. Clustering the graph embeddings thus captures the existence of dynamic heterogeneity in these disordered systems.
S2. Density based clustering:

Density based clustering identifies groups/clusters in data based upon the intuition that each cluster is a contiguous region of high point density, separated from other clusters by regions of low point density. Fig. 6 shows an example of density based clustering obtained using the DBSCAN algorithm [35]. We can clearly see that DBSCAN generated clusters where density of points is high and is separated by regions of low density of points. Density based clustering can generate clusters of any shape, which has advantage of other methods such as the popular centroid-based technique K-means, which generates clusters that are more or less spherical.
However, DBSCAN fails to identify clusters when data has clusters of varying density. As an example, let us consider Fig. 7(a), which consists clusters of varying densities. DBSCAN fails to identify one of the clusters. The prime reason responsible for this behavior is that DBSCAN uses a fixed cutoff distance between two points to consider them as neighbors. However, a larger distance in one cluster could be a shorter distance in another cluster as can be seen in Fig. 7(a). Further, a solution that uses larger cutoff could lead to merging of multiple clusters and a lower cutoff could lead to missing out points to be part of a cluster as in Fig. 7(a). Hence, finding the optimal cutoff is non-trivial.
To overcome this limitation, the OPTICS (Ordering Points To Identify the Clustering Structure) [36] algorithm was proposed. Specifically, instead of relying on a fixed cutoff distance, points are ordered linearly such that spatially closest points become neighbors in the ordering. This ordering is then utilized to perform clustering. As we can see in Fig. 7(b), using OPTICS, we are able to identify regions having varying densities.

S3. Optics Clustering Algorithm
As discussed above, OPTICS [36] is a density-based clustering algorithm, like DBSCAN, with the additional advantage of being able to identify clusters of varying densities. We next introduce few definitions for better understanding of the algorithm. Please refer to Fig. 8(a) for a visual depiction.
-
•
parameter: Two points are considered neighbors if their distance is less than . In our case it translates to the cosine distance between the learned representations of two atoms (nodes). Specifically for our case, without loss of generality, an atom is neighbor of an atom if . Here refers to the representation of the atom at layer of the Gnn.
(a) Core point (b) Reachability distances Figure 8: OPTICS: Core points and reachability distance of points -
•
-Neighborhood : The points that are within distance of a node .
-
•
: Minimum number of points that must be within distance of an atom in order to form a cluster.
-
•
Core Point: A point is considered a core point if it has more than points within distance.
-
•
Core Distance: Minimum required to make a point a core point.
(5) -
•
Reachability Distance: Reachability distance between a point and is the maximum of the Core Distance of and the distance between and . For point that is reachable from a point :
(6) Fig. 8(b) shows examples of reachability distances for some points.

In OPTICS clustering, first, core distances of all points are computed. Then, we calculate their reachability distances iteratively. In this iterative process, the data points are chosen in ascending order of their lowest reachability distance to the points already visited until that time. Finally, based upon the reachability distances of all points, we construct a reachability plot as in Fig. 9(c).
In Fig. 9(c), we observe valleys of points (sequence of low reachability distances sandwiched between high reachability distances), which represent high density regions. Valleys get separated by regions of low-density. These valleys can be considered to be part of the same cluster(Fig. 9(b). For example, the colored points(other than gray) are those identified as clusters, while gray ones represent noise. For more details on OPTICS, we refer to the original paper [36].
S4. Parameter details of GraphSage
Hyper-parameters used for training the GraphSage Gnn model are given below. Note that these parameters were optimized based on the results on the training and validation data.
-
•
Number of Gnn Layers() = 2
-
•
Hidden layer size = 64
-
•
Random Walk Length() = 4
-
•
No. of Random Walks per node = 10
-
•
Number of Negative Samples () = 10
-
•
Optimizer = Adam
-
•
Learning rate = 0.01
-
•
Graph batch size = 5
-
•
Node batch size = 250
-
•
Neighborhood aggregation function = Mean
-
•
Gradient clipping max norm = 5
Choice of number of layers (): As discussed in the main paper, each node aggregates information from its -hop neighborhood. To identify a suitable value of , we evaluate the model by performing a random walk test on the set of validation graphs (not seen during training). As per our training loss in Eq. 2, the cosine similarity of embeddings of a pair of nodes co-occurring in random walks is expected to be higher than cosine-similarity of embeddings of node pairs that do not co-occur in random walk. In Fig. 10, we show the Area under cover- Receiver Operating characteristics (AUC-ROC) score obtained from random walk test for Gnn model trained at different number of Gnn layers (). We observe that the performance of the model at is significantly higher than at . Further, the performance saturates at . Hence, we use , which provides a good balance between efficacy and efficiency (computation cost grows with the number of layers).

S5. Parameter details of OPTICS
We used OPTICS clustering from Scikit-learn library[37]. We mention the parameters used below:
-
•
: Minimum number of samples in a neighborhood for a point to be considered as a core point. We set this to .
-
•
: Maximum distance between two samples for one to be considered as in the neighborhood of the other. We set this to .
-
•
Distance metric: Cosine similarity
S6. Reachability plot for nodes of 2D LJ System

In Fig. 11 we show the reachability distance plot obtained by OPTICS algorithm for nodes of our 2D LJ system (same as shown in Fig. 4.b in main paper). Each color represents a cluster generated by OPTICS as in Fig. 4.b. The gray colored points(noise) instead of forming a cluster themselves, act as a separator between regions of high density.
S7. Effect of random walk length
We also study the effect of length of random walk in Table 1. We observe that the correlation between MSD of clusters and average cluster potential energy increases monotonically with the length of random walk till , after which it drops. A possible reason is that when the length of the random walk increases beyond , far away nodes could get incorporated in the set of positive samples pairs, and the model might accumulate noise from these node pairs at larger distances.
Random Walk Length | Correlation |
---|---|
2 | 0.658 |
3 | 0.714 |
4 | 0.747 |
5 | 0.627 |