This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Signed Graph Representation Learning: A Survey

Zeyu Zhang1    Peiyao Zhao2    Xin Li2    Jiamou Liu3   
Xinrui Zhang4
   Junjie Huang4&Xiaofeng Zhu5 1Huazhong Agricultural University
2Beijing Institute of Technology
3The University of Auckland
4Southwest University
5University of Electronic Science and Technology of China [email protected], {peiyaozhao, xinli}@bit.edu.cn, [email protected], [email protected], [email protected], [email protected]
Abstract

With the prevalence of social media, the connectedness between people has been greatly enhanced. Real-world relations between users on social media are often not limited to expressing positive ties such as friendship, trust, and agreement, but they also reflect negative ties such as enmity, mistrust, and disagreement, which can be well modelled by signed graphs. Signed Graph Representation Learning (SGRL) is an effective approach to analyze the complex patterns in real-world signed graphs with the co-existence of positive and negative links. In recent years, SGRL has witnesses fruitful results. SGRL tries to allocate low-dimensional representations to nodes and edges which could preserve the graph structure, attribute and some collective properties, e.g., balance theory and status theory. To the best of knowledge, there is no survey paper about SGRL up to now. In this paper, we present a broad review of SGRL methods and discuss some future research directions.

1 Introduction

The widespread impact of online platforms, like social media, business dealings, and cryptocurrency transactions, has led to a huge increase in graph datasets. These datasets, with complex and connected structures, pose a big challenge for analysis. ver the past decade, graph machine learning methods, especially Graph Neural Networks (GNNs) Kipf and Welling (2017); Hamilton et al. (2017); Velickovic et al. (2018), have gained increasing attention in both academic and industrial circles. These approaches have significantly advanced in various applications, such as link prediction, node classification, and graph classification.

Despite significant progress in GNNs, most GNN methods are designed for unsigned graphs (consisting of only positive edges). However, real-world edge connections among nodes frequently extend beyond conveying positive relationships like friend, accept, trust, and support. They may also encompass negative associations, such as foe, rejection, distrust, and opposition. In Figure 1, we find that signed graphs have numerous applications in our daily lives, spanning various domains and contexts. This setup can be thought of as a graph with both positive and negative connections. However, having negative links messes up the usual way information is passed, so we need new signed graph representation learning (SGRL) models, like Signed Graph Neural Networks (SGNN), to handle this mix of positive and negative connections.

Refer to caption
Figure 1: An illustration of signed graph in real world.

For unsigned graph representation learning, there exist several surveys in recent years Cui et al. (2018); Cai et al. (2018); Wu et al. (2020b); Zhang et al. (2020); Zhou et al. (2020). Cai et al. (2018) provides a review of graph embedding methods and classifies the current methods into matrix factorization, deep learning, matrix reconstruction, graph kernel, generative models, and so on. Cui et al. (2018) focuses on categorizing and reviewing current network embedding methods including linear models and deep models. Wu et al. (2020b) only focuses on deep models of graph representation learning and proposes a new taxonomy to classify GNN models into four categories, i.e., recurrent GNNs, convolutional GNNs, graph autoencoder, and spatial-temporal GNNs. Similar to Wu et al. (2020b), Zhang et al. (2020) also presents a comprehensive review of different types of deep learning methods on graphs. Zhou et al. (2020) provides an overview of GNN variants from different angles, i.e., graph type, training methods, and propagation step. In addition to these surveys, there exist several related surveys about the extended application of graph neural networks Lamb et al. (2020); Wu et al. (2020a). For signed graph mining, one of the important surveys is Tang et al. (2016b). This survey gives an overall review of signed graph properties and analysis tasks but does not consider graph representation learning methods as this direction only begins to receive attention in recent years. Another signed graph-related survey is Zheng et al. (2015). This survey provides a comprehensive overview of significant research regarding social balance, one of the important signed graph collective properties, including fundamental measures, detecting algorithms, and so on.

In summary, to the best of our knowledge, no surveys are specifically written for the SGRL methods. Thus, in this paper, we provide a comprehensive and unique survey about SGRL to bridge the gap. We believe this survey can not only help researchers gain a deep understanding of the current progress of SGRL but also provide rich resources and potential future directions in this field.

2 Challenges

Despite significant progress in graph representation learning, research on SGRL remains crucial for the following challenges:

How to handle negative edges. Most Graph Neural Networks (GNNs) for unsigned networks use homophily, assuming connected users share similarities. However, this concept may not directly apply to signed networks, where negative edges represent opposing semantics.

Limitation of balance theory. Many SGRL methods rely on balance theory, rooted in psychology, which simplistically divides nodes into two conflicting groups. However, this falls short of capturing the complexity of real signed graphs. It is essential to explore extensions of current SGNN methods to handle cases with more than two groups (k-groups).

Representation limitation of SGNN models. Unbalanced triangles, common in real-world signed graph datasets, challenge current SGNN methods in accurate representation learning Zhang et al. (2023b). Increased unbalanced triangles due to factors like random noise and adversary attacks impact predictive performance, highlighting the need to explore mitigation strategies.

Diverse learning tasks. Downstream tasks for signed graph representation, like link sign prediction and node classification, differ from unsigned graphs. Node classification is particularly challenging in signed graph datasets without node labels and attributes, relying solely on the network.

3 Preliminaries and Problem Formulation

This section will introduce some preliminaries to lay the foundations for further discussion, including the preliminaries for signed networks, popular signed GNN models, and basic signed social theory.

3.1 Signed Social Graphs

In this subsection, we introduce notations and give the necessary backgrounds on the signed graph.

Notations. We denote a signed graph as 𝒢={𝒱,,s}\mathcal{G}=\{\mathcal{V},\mathcal{E},s\}, where 𝒱\mathcal{V} is the node sets in a signed graph 𝒢\mathcal{G}, and 𝒱×𝒱\mathcal{E}\subset\mathcal{V}\times\mathcal{V} represents the sets of edges in graph 𝒢\mathcal{G} with sign ss. +\mathcal{E}^{+} and \mathcal{E}^{-} denote the sets of positive and negative links, respectively. It means that there is at most one signed edge between any two nodes. \mathcal{E} contains +\mathcal{E}^{+} and \mathcal{E}^{-}. Similarly, we can divide the neighbors of node viv_{i} into positive neighbors 𝒩i+\mathcal{N}_{i}^{+} and negative neighbors 𝒩i\mathcal{N}_{i}^{-}. Also, 𝒢\mathcal{G} can be denoted as the adjacency matrix AA, where Aij=1A_{ij}=1 means there exists a positive link between viv_{i} and vjv_{j}, Aij=1A_{ij}=-1 denotes a negative link between uiu_{i} and uju_{j}, and Aij=0A_{ij}=0 means there is no link between uiu_{i} and uju_{j}. Usually, s{1,1}s\in\{1,-1\} is the sign of the edge111The signed network can be further associated with weights. Kumar et al. (2016) and +=\mathcal{E}^{+}\bigcap\mathcal{E}^{-}=\emptyset;

Problem formulation. Given a signed graph 𝒢={𝒱,,s}\mathcal{G}=\{\mathcal{V},\mathcal{E},s\}, SGRL aims to encode the nodes 𝒱\mathcal{V} into low-dimensional vectors Zd×|𝒱|Z\in\mathbb{R}^{d\times|\mathcal{V}|} as follows:

f(𝒢)Z,f(\mathcal{G})\rightarrow Z, (1)

where dd is the embedding size (d|𝒱|d\ll|\mathcal{V}|), and ff is the encoder function which can be Network Embedding or GNNs. The low-dimensional vectors ZZ should be useful in downstream machine-learning tasks.

Dataset # Nodes # Edges %Pos Sign Semantics Scenarios
Bitcoin-Alpha 3,783 24,186 93.7 Trust/Distrust Social Relationship
Bitcoin-OTC 5,881 35,592 90.0 Trust/Distrust
Wikirfa 11,259 178,096 77.9 Support/Oppose
Slashdot 82,144 549,202 77.4 Like/Dislike
Epinions 131,828 841,372 85.3 Trust/Distrust
RedditHyperlinks 55,863 858,490 92.7 Positive/Negative
Bonanza 9,892 36,543 98.0 Positive/Negative Object Opinion
U.S. House 1,796 114,378 54.0 Support/Oppose
U.S. Senate 1,201 27,083 44.7 Support/Oppose
PeerReview 486 1,170 39.7 Accept/Reject
ML-1M 9,992 1,000,209 57.5 High-rate/Low-rate Interaction Feedback
Amazon-Book 73,857 1,960,674 80.6 High-rate/Low-rate
Yelp 66,418 1,900,308 68.2 High-rate/Low-rate
Zhihu 17,437 554,150 29.3 Interact/Skip
WeChat 8,582 13,557 85.7 Read/Report
Table 1: Common Datasets for Signed Networks.

Datasets. Beyond social relationships, antagonistic relationships also exist in complex systems like diplomacy, biology, and chemistry. Examples include cooperative and hostile relations in international affairs, excitatory and inhibitory relationships among neurons in biology Tang et al. (2016b), and the side effects between chemical drugs Zitnik et al. (2018). While these complex systems also feature signed links, this paper primarily focuses on signed links in social relationships, where the entities involved are individuals. We have collected and organized datasets commonly used for modeling signed networks, as shown in Table 1. From Table 1, it can be seen that the modeling of signed networks can be categorized into three different types based on the context of the links. Thus, the scenarios of signed networks are divided into: social relationships between people, object opinions towards specific topics, and interaction feedback in human-computer interaction.

Machine learning tasks Common graph machine learning tasks, as detailed in Wu et al. (2020b), typically fall into three distinct categories: Node-level, Link-level, and Graph-level. Node-level graph machine-learning tasks usually involve applying machine learning to predict nodes’ attributes. Link-level tasks primarily focus on predicting the relationships between edges in the network222 In signed graphs, link prediction typically includes predicting the sign of a given link (binary classification) and predicting both the existence and the sign of a link (3-class classification).. In graph-level machine learning tasks, there are generally two main subjects of study: subgraph-level and whole-graph-level. At the subgraph level, the focus is on clustering nodes within social networks, with common tasks including Node Clustering He et al. (2022b) and Community Detection Sun et al. (2020). Whole-graph-level tasks mainly encompass Graph Classification Liu et al. (2023) and Graph RegressionLi et al. (2019). It is worth mentioning that the mainstream machine learning task of SGRL is the Link-level.

3.2 Signed Graph Neural Network

The architecture of current popular SGNN models, such as SGCN Derr et al. (2018) and SNEA Li et al. (2020) follows the spatial unsigned GNNs utilizing neural message passing among nodes to define convolutions on graphs, but adopt the following mechanism. The representation of a node viv_{i} at a given layer \ell is defined as

hi()=[hipos(),hineg()],h_{i}^{(\ell)}=[h_{i}^{pos(\ell)},h_{i}^{neg(\ell)}],

where hipos()h_{i}^{pos(\ell)} and hineg()h_{i}^{neg(\ell)} respectively denote the positive and negative representation vectors of node vi𝒱v_{i}\in\mathcal{V} at the \ellth layer, and [,][,] denotes the concatenation operation. The unsigned GNN message passing is defined as

hi(l)=COMBINE(1)(hi(1),AGGREGATE(1)({hj:vj𝒩i})),\scriptsize h_{i}^{(l)}=\text{COMBINE}^{(\ell-1)}\left(h_{i}^{(\ell-1)},\text{AGGREGATE}^{(\ell-1)}(\{h_{j}:v_{j}\in\mathcal{N}_{i}\})\right), (2)

where COMBINE and AGGREGATE are differentiable functions (e.g., Mean, Max). Different from GNNs, SGNNs accommodate positive and negative edges using a two-part representation, and a more involved aggregation scheme. For example, when >1\ell>1, the positive part of the representation for node viv_{i} could aggregate information from the positive-representation of its positive neighbors and the negative-representation of its negative neighbors.

hipos()=\displaystyle h_{i}^{pos(\ell)}= COMBINE()(hipos(1),\displaystyle\text{COMBINE}^{(\ell)}\Biggl{(}h_{i}^{pos(\ell-1)}, (3)
AGGREGATE()({hjpos(1):vj𝒩i+},{hjneg(1):vj𝒩i}))\displaystyle\text{AGGREGATE}^{(\ell)}(\left\{h_{j}^{pos(\ell-1)}:v_{j}\in\mathcal{N}_{i}^{+}\right\},\left\{h_{j}^{neg(\ell-1)}:v_{j}\in\mathcal{N}_{i}^{-}\right\})\Biggl{)}
hineg()=\displaystyle h_{i}^{neg(\ell)}= COMBINE()(hineg(1),\displaystyle\text{COMBINE}^{(\ell)}\Biggl{(}h_{i}^{neg(\ell-1)},
AGGREGATE()({hjneg(1):vj𝒩i+},{hjpos(1):vj𝒩i})),\displaystyle\text{AGGREGATE}^{(\ell)}(\left\{h_{j}^{neg(\ell-1)}:v_{j}\in\mathcal{N}_{i}^{+}\right\},\left\{h_{j}^{pos(\ell-1)}:v_{j}\in\mathcal{N}_{i}^{-}\right\})\Biggl{)},

After \ell SGNN layers, we have the final representation Zi=CONCATENATE(hi,hi)Z_{i}=\text{CONCATENATE}(h_{i}^{\ell},h_{i}^{\ell}). In addition to CONCATENATE, we can also use MLP or ATTENTION mechanisms to fuse positive and negative representations.

3.3 Social Theory

In the realm of traditional social network theory, a considerable amount of research has focused on the concepts of Homophily and Heterophily. However, when negative links are introduced into social networks, many assumptions no longer hold. For example, heterophily assumptions also suggests that connected nodes are prone to have different properties or labels. However, it significantly differs from Balance theory in its inability to capture certain concepts like “an enemy of an enemy is a friend”. In this section, we introduce two important sociological theories that are more prevalent in signed graphs: Balance theory and Status theory.

Balance theory Heider (1946), a concept in social psychology, was developed by Fritz Heider in the 1940s. The theory focuses on the idea that individuals seek cognitive consistency and balance in their attitudes and relationships within social networks. According to balance theory, people strive to maintain a psychological balance in their interactions. In the context of triadic relationships, where three entities are connected by positive and negative ties, balance theory suggests that individuals tend to prefer balanced configurations. For example, if viv_{i} has a positive relationship with vjv_{j} and a negative relationship with person vkv_{k} there is an imbalance in the triad. To restore balance, viv_{i} might either develop a positive relationship with vkv_{k} or a negative relationship with person vjv_{j}. Figure 2 shows all four isomorphism types of triads.

Refer to caption
Figure 2: Four isomorphism types of triads. Blue and red lines represent positive and negative edges, resp.

Status theory Leskovec et al. (2010) is relevant for directed networks compared to balance theory is naturally defined for undirected networks. Social status can be represented in a variety of ways, such as the rankings of nodes in social networks, and it represents the prestige of nodes. In its most basic form, Status theory suggests that viv_{i} has a higher status than vjv_{j} if there is a positive link from vjv_{j} to viv_{i} or a negative link from viv_{i} to vjv_{j}. In Figure 2, we can see the first two satisfy Status theory, while the latter two do not. This status relationship is transitive.

When comparing the two theories, Balance theory can be understood as simulating individual preferences, like liking or disliking, and primarily applies to undirected signed graph. On the other hand, Status theory is based on the assessment of individuals’ social status and is direction-dependent. Both theories can be modeled within signed directed triads (see Figure 2). Chen et al. (2018) propose to use triplets (ii, jj, kk) to comparatively measure and analyze the two theories. They found that in the Slashdot and Epinions datasets, approximately 75% of the triangles satisfy both sociological theories simultaneously, and only about 1% of triangles satisfy neither.

Table 2: A summary of SGRL methods. “Information Source” denotes the information used during the encoding process. “Encoder type” refers to the foundational types on which the encoder design is based. “Loss Type” denotes the loss function of the Optimization goal. “Social Theory” refers to the collective properties of signed graph datasets, i.e., balance theory, status theory. “Task” indicate the learning task of each method.
Method Information Type Encoder Type Loss Type Social Theory Task
Network Embedding
SNE Yuan et al. (2017) Network only Log-Bilinear Model MLE Balance Theory Node Classfication, Link Prediction
SIGNet Islam et al. (2018) Network only word2vec MLE Balance Theory Link Sign Prediction
SIDE Kim et al. (2018) Network only Skip-Gram MLE Balance Theory Link Sign Prediction
ROSE Javari et al. (2020) Network only Node2vec MLE N/A Link Sign Prediction
SPONGE Cucuringu et al. (2019) Network only Spectral Method Four-player minimax game Balance theory Node Classification
Deep Model
SiNE Wang et al. (2017) Network, node feature Neural Nework Reconstruction Balance theory Link Sign Prediction
SSNE Lu et al. (2019) Network only Neural Network Energy-based ranking Status theory Link Sign Prediction
BESIDE Chen et al. (2018) Network Only Neural Network Cross Entropy Balance/Status Link Sign Prediction
SGCN Derr et al. (2018) Network, Node feature GCN Cross Entropy, Structure Balance Theory Link Sign Prediction
SHINE Wang et al. (2018) Multi-Network Autoencoder Reconstruction N/A Link Sign Prediction
SSSNET He et al. (2022b) Network, node feature MLP Reconstruction Balance Theory Node Clustering
SNEA Li et al. (2020) Network, node feature GAT Cross Entropy, Structure Balance Theory Link Sign Prediction
GS-GNN Liu et al. (2021) Network, node feature Dual GCN Cross Entropy N/A Link Sign Prediction
SiGAT Huang et al. (2019) Network only, node feature GAT Reconstruction Balance/Status Link Sign Prediction
Trustworthy Model
ASiNE Lee et al. (2020) Network Only GAN Four-player minimax game Balance Theory Link Sign Prediction
SigGAN Chakraborty et al. (2023) Network Only GAN Two-player minimax game Balance Theory Link Sign Prediction
SGCL Shu et al. (2021) Network, Node feature Multi GAT Cross Entropy, Contrastive Balance Theory Link Sign Prediction
RSGNN Zhang et al. (2023b) Network, Node feature GCN Sparsity, High Balance, etc a Balance/Status Link Sign Prediction
SGA Zhang et al. (2023c) Network Only GCN Cross Entropy Balance Theory Link Sign Prediction
a.a. feature loss, reconstruction and cross entropy loss

4 Categorization

To transform signed graphs from topology space to embedding spaces, there are different methods adopted in current models. In this section, We categorize signed signed graph representation methods into three parts, i.e., network embedding methods, deep models and trustworthy models.

4.1 Shallow Network Embedding

Shallow network embedding typically refers to those methods with simple, linear non-deep learning structures. This type of method is mainly divided into two aspects: methods based on random walks (SNE Wang et al. (2017), SIGNet Islam et al. (2018), SIDE Kim et al. (2018), ROSE Javari et al. (2020)), SSNE Lu et al. (2019) and methods based on matrix factorization SPONGE Cucuringu et al. (2019). Below, we will introduce these methods separately.

SNE Wang et al. (2017) is the first research on signed network embedding, which adopts the log-bilinear model (variant of Skip-Gram model) to capture both node’s path and sign information. One issue present is that the relationship between the target node and multi-hop neighbor nodes is determined by the edge sign between multi-hop neighbors and the previous-hop neighbors. Specifically, for a target node vv and a path h=[u1,u2,,ul,v]h=[u_{1},u_{2},\cdots,u_{l},v], the relationship between node vv and u1u_{1} is decided by u1,u2\mathcal{E}_{u_{1},u_{2}}. If u1,u2>0\mathcal{E}_{u_{1},u_{2}}>0, their relationship is positive. Conversely, it is negative.

SIGNet Islam et al. (2018). This method encodes network structural information based on the word2vec model, considering both undirected and directed graph scenarios. A major innovation of this approach lies in proposing a new negative sampling method. In signed graphs, where relationships between nodes can be positive or negative, the paper suggests that nodes with positive relationships should not be included in the negative sampler set and introduces a method based on balance theory to compute the positive and negative relationships of multi-hop neighbors. In summary, neighbors connected to the target node by an even number of negative edges are assigned to the positive relationship set, while those connected by an odd number are assigned to the negative relationship set.

SSNE Lu et al. (2019) incorporates status theory into signed network embedding. In status theory, a higher status node viv_{i} has a positive link to vjv_{j}, while a lower status node viv_{i} has a negative link to vjv_{j}. SSNE represents positive links as triplets (vi,+,vj)(v_{i},\mathcal{\ell}{+},v_{j}), where the embedding space forces the combined embeddings of hi+h+h_{i}+h_{\mathcal{\ell}{+}} to be close to the embedding of hjh_{j}. Conversely, for negative links, the model encourages the embedding hih+h_{i}-h_{\mathcal{\ell}_{+}} to be close to the embedding of hjh_{j}.

SIDE Kim et al. (2018) proposed signed network embedding method builds upon truncated random walk, introducing a comprehensive likelihood formulation for signed directed connections that consistently represents both positive and negative edges. By incorporating bias factors in the likelihood function to model individual connectivity, the approach enhances the accuracy of the embedding process. Furthermore, the random walk sampling process and likelihood formulation are extended to accommodate multi-step relationships, encompassing both sign and direction. This method intricately links vector space geometry with social phenomena in networks, including homophily, preferential attachment, and balance theoretic behaviors. This association dissects the dual factors influencing link formation, establishing a robust foundation for the broad application of the method in the analysis of signed directed networks.

ROSE Javari et al. (2020). This papers attempts to address two issues. First, existing social theories fail to explain the structure in all signed graphs, meaning that not all structures conform to balance theory and status theory. Second, current signed network embedding only considers predicting the positive or negative nature of existing edges, overlooking the prediction of edge existence. To tackle these challenges, ROSE does not rely on any specific social theory for node encoding. Instead, it adopts a network transformation-based embedding approach. This involves assigning multiple roles to the same node, transforming the signed graph into an unsigned graph. Subsequently, unsigned network embedding methods can be applied to encode the graph structure.

SPONGE (Signed Positive Over Negative Generalized Eigenproblem) Cucuringu et al. (2019) is a kk-way spectral clustering algorithm on signed networks. Inspired by constrained clustering, a new kk-way objective of signed graph cut is designed, which seeks the global clustering assignment for each node by minimizing the trade-off between two measures of badness, i.e., the normalized weight of positive edges cross different clusters and the reciprocal of that of negative edges cross different clusters. The new objective is cast as a regularized spectral algorithm based on solving a generalized eigenproblem. The key of solving it is to derive the goal matrix, and then perform eigenvalue decomposition on the goal matrix to obtain its eigenvectors corresponding to the kk-smallest eigenvalues. Simultaneously, it also requires that the derived embedding vectors be orthonormal for dropping the discreteness constraints. Moreover, this work also provides a detailed theoretical analysis w.r.t. the robustness of SPONGE in the Signed Stochastic Block Model framework.

4.2 Deep Model

Compared to shallow network embedding methods, deep models in the context of network embedding refer to those approaches that involve deep learning architectures with multiple layers in the neural network. This is currently one of the mainstream methods in SGRL.

SiNE Wang et al. (2017) is the first method which employs deep learning model to learn low-dimensional vector representations for nodes of a given signed graph. Follow the assumption of extended structural balance theory, in embedding space, users should sit closer to their “friends” (or users with positive links) than their “foes” (or users with negative links). Based on this assumption, in order to preserve the structure information of signed graph, special triplets (e.g., (vi,vj,vk)(v_{i},v_{j},v_{k})) (with one positive link eij=1e_{ij}=1 and one negative link eij=1e_{ij}=-1) are extracted, the learned similarity function should make sure f(vi,vj)>f(vi,vk)f(v_{i},v_{j})>f(v_{i},v_{k}).

SGCN Derr et al. (2018) tackles the challenge of applying Graph Convolutional Network (GCN) Kipf and Welling (2017) to signed networks. GCN excels in unsigned networks, but struggles with the semantic and structural complexities of negative links. SGCN introduces balance theory to categorize neighbors as positive or negative based on the number of negative edges. It maintains dual representations for each node—balanced set for positive relationships and unbalanced set for negative ones. This approach, using an extended balance theory for judging multi-hop neighbors, enables SGCN to effectively model the nuanced relationships in signed graphs.

DNE-SBP Shen and Chung (2018) pioneers the application of semi-supervised Stacked Autoencoder (SAE) for embedding signature networks. This deep embedding model focuses on learning low-dimensional node vector representations while preserving structural balance in signed networks. By employing semi-supervised stacked autoencoders, DNE-SBP reconstructs the adjacency connections in the given signed network. The model prioritizes sparse negative connections by applying a higher penalty to negative links during reconstruction. To maintain structural balance, pairwise constraints are incorporated, ensuring nodes connected by positive links in the embedding space are closer than those connected by negative links. This innovative approach addresses the limitation of SDNE in capturing crucial structural balance properties in signature networks.

SNEA Li et al. (2020). Attention mechanisms allow for dealing with changeable sized input and focus on the most relevant parts of input to make decisions. Graph attention network (GAT) Velickovic et al. (2018) is the first attention-based graph learning architecture, which assign different weights to different neighbor nodes. SNEA is a GAT-based signed network embedding methods which distinguish neighbor nodes into positive and negative neighbors which follow the similar design from Derr et al. (2018). The difference lies in that the contribution of neighbor nodes to the target node is not fixed; instead, by introducing attention mechanism, the model can learn an appropriate weight for each neighbor node.

SiGAT Huang et al. (2019). SiGAT is another attention-based graph neural network applied to signed networks. Different from previously mentioned SNEA, it primarily defines different relationships by establishing various motifs. These motifs are related to triangles in the balance theory and status theory. Then, it defines an attention aggregation mechanism under different relationships, obtaining the learning of node representations in signed networks.

SDGNN Huang et al. (2021b). Building upon the concept of SiGAT, SDGNN further draws inspiration from GAE, treating the task of learning representations for signed graphs as an encoder-decoder framework. In the encoder part, SDGNN simplifies the definition of different relationships, defining only four types of signed directed relationship. Furthermore, it integrates the depiction of balance theory and status theory in signed networks into the decoder parts. SDGNN propose to recontruct the sign, direction, and trianles in loss functions.

SGCL Shu et al. (2021) is the first to employ graph contrastive representation on signed graph. Contrastive learning is an unsupervised learning paradigm which can help model capture invariant and robust representations under perturbations. Compared to the unsigned graph contrastive learning framework, SGCL has three main differences. Firstly, the data augmentation method is different, taking into consideration the existence of positive and negative edges in edge relationships. Therefore, flipping positive and negative edges is used as a means of adding perturbation. Secondly, SGCL adopts the design of encoding positive and negative edges separately. Thirdly, balance theory is employed in the design of the contrastive loss function.

SSSNET (Semi-Supervised Signed NETwork clustering) He et al. (2022b) designed a new GNN-based aggregation mechanism, Signed Mixed-Path Aggregation (SIMPA), to aggregate up-to-hh-hop contributing neighbors, by assigning weights for different paths. Through statistical analysis of real-world signed graph datasets333Six public datasets are used, including Sampson, Rainfall, Fine-YNet, S&P 1500, PPI, Wiki-Rfa., nearly 20%50%20\%-50\% of triangles violate the social balance theory. Thus a variant of social balance theory is proposed to generate those paths, where a neutral stance is assumed on whether or not the enemy of an enemy is a friend. Specifically, the positive embedding is the weighted combination of the node representations from a “friend path” where all edges need to be positive. For a target node to be an hh-hop enemy neighbor of the source node, exactly one edge on the “enemy path” has to be negative. Besides, a (differentiable) Probabilistic Balance Normalized Cut (PBNC)444Balance Normalized Cut (BNC) is a non-differentiable function with hard clustering assignment distribution Chiang et al. (2012). as a self-supervised loss is introduced to be optimized for training clustering on signed graphs, which minimizes the probability of negative edges assigned in intra-clusters and maximizes the probability of positive edges assigned in inter-clusters. PBNC with supervised classification loss (i.e., Cross Entropy and Triplet Loss) between the ground truths and the predicted labels is leveraged to optimize node embeddings and clustering simultaneously without any intermediate step.

4.3 Trustworthy Model

Trustworthy models typically refer to those representation learning methods which can produce robust, explainable ethical outcomes. Research in this area is very limited, making it a crucial direction for future studies.

ASiNE Lee et al. (2020) is the first attempt that utilizes adversarial learning for signed network. Motivated by generative adversarial networks for unsigned network, ASiNE designed two pairs of generators and discriminators to generate and distinguish false positive edges and false negative edges, respectively. Two generators share an embedding space, similarly, two discriminators share another embedding space. In addition, the negative edge generator can also generate false positive edges based on the balance theory. Generators aim to generate the most indistinguishable edges, while discriminators aim to discriminate between real edges and fake edges, acting as two pairs of opponents in the following four-player minimax game. As the game progresses, the performance of both sides will gradually improve.

RSGNN Zhang et al. (2023b) is the first paper dedicated to the robustness study of the SGNN model, investigating the impact of random noise on SGNN. The paper theoretically analyzes the limitations of the current SGNN, demonstrating that the existing SGNN model fails to learn suitable representations from unbalanced triangles. Furthermore, random noise increases the number of unbalanced triangles, thereby illustrating that the decline in SGNN performance is attributed to the introduction of random noise. Then, RSGNN explore the properties of real-world signed graph to defend the negative effect of noise and propose a novel framework RSGNN which adopts a dual architecture that simultaneously denoises the graph and learns the node representations.

5 Recent Advances

Recently, the field of signed networks has seen several cutting-edge advancements, primarily including the modeling of more complex signed networks, broader applications of signed networks, and the development of tools specifically designed for signed networks.

5.1 More Complex Signed Networks

Compared to traditional signed social networks, researchers are increasingly focusing on more complex signed networks such as signed weighted graphs Kumar et al. (2016), signed temporal graphs Sharma et al. (2023), signed hypergraphs Chen et al. (2020) and so on. Within these domains, the fields of signed bipartite graphs modeling object opinions and signed temporal graphs modeling temporal dynamics have seen the emergence of some typical works, such as SBGNN, SBGCL, and SEMBA.

SBGNN Huang et al. (2021a) is a SGNN model designed for signed bipartite graphs. Like most GNNs message-passing scheme, SBGNNs follow a message-passing scheme. However, in SBGNN, new message functions, aggregation functions, and update functions are defined by applying balance theory. More specifically, message neighborhood propagation is divided into four categories. SBGNN captures the higher-order information through a layer-by-layer design.

SBGCL Zhang et al. (2023a) is another model that applies contrastive learning models to signed bipartite graph representation learning. Unlike SGCL, SBGCL focuses on signed bipartite graphs, attempting to address the issue of SGCL’s inability to capture potential relationships among nodes of the same type. SBGCL enhances a signed bipartite graph through an innovative two-level graph augmentation method and a multi-perspective contrastive loss is employed to unify the node presentations learned from the two perspectives.

SEMBA Sharma et al. (2023) is a signed GNN model that combine sign and dynamci in social networks. In order to addressing temporal-awareness, staleness, and sign-awareness problems, SEMBA use memory modules and balanced aggregation to learn short-term memory encoding and a long-term embedding. More specifically, memories contains both positive and negative parts, and aggregation encode positive and negative embeddings following balance theory.

5.2 Applications

With the rise of social media, social conflicts have intensified, manifesting in an increase in negative links between individuals. In this section, we provide an overview of some of the most attention-grabbing applications currently.

Social Polarization Using signed graphs to study adverse effects is a promising research topic. These adverse effects, such as polarization and echo chambers, can be harmful to the process of democratic deliberation in our societyXiao et al. (2020); Bonchi et al. (2019). For example, POLE Huang et al. (2022) is a signed networking embedding methods, adopting a measure of polarization based on the signed random-walks. POLE first design a polarization measure for signed graphs and use matrix factorization optimized polarized similarity consistency.

Stance Detection Stance detection typically categorizes stances into three types: “support,” “oppose,” and “neutral,” where in this context, “support” can be viewed as a positive relationship and “oppose” as a negative relationship555Neutral could be modeled as either a positive relationship or unknown.. While a substantial amount of work on stance detection is based on Natural Language Processing (NLP). However, an increasing number of studies have found that social networks significantly influence people’s opinions on specific matters Pougué-Biyong et al. (2023). SEM Pougué-Biyong et al. (2023) jointly learns user and topic embeddings in signed social graphs with distinct edge types for each topic.

Recommendation System with Negative Feedback With the success of graph neural network models in the field of recommendation systems, these systems can be viewed as user-item interaction bipartite graphs. A substantial number of researchers have employed graph representation learning to model these bipartite graphs, as seen in works like PinSAGE Ying et al. (2018) and LightGCN He et al. (2020). However, in the context of recommendation systems, the negative feedback mechanism Xie et al. (2020) also refers to users expressing dissatisfaction with recommended products, manifested through actions like poor reviews, skipping recommended content, or reporting. The optimization of recommendation systems using negative feedback has garnered attention from researchers Tang et al. (2016a). Notably, Seo et al. (2022) introduced the SiReN model, which employs a signed graph neural network approach to model recommendation systems incorporating both positive and negative feedback. By conducting qualitative and quantitative experiments on the mechanism of negative feedback, Huang et al. (2023) found that negative feedback interactions (i.e., negative edges) could enhance the performance of recommendation systems (specifically, the prediction of positive edges) to a certain extent. They proposed a new Signed GNN model named SiGRec with a new SiC loss function to model signed graphs in recommendation systems.

Social Computing In addition to the applications mentioned above, due to the relevance of its links to human sentiment, SGRL has also been applied in various social computing fields. These include education Ni et al. (2023), communication He et al. (2022a), cryptocurrency, politics, and more. This broad applicability underscores the significance of understanding and leveraging the dynamics of signed networks in diverse social scenarios.

5.3 Tools

With the rise of graph representation learning, an increasing number of tools for modeling graph representations have been introduced. Among the most representative of these are Pytorch Geometric (PyG) and Deep Graph Library (DGL). However, these approaches often view signed graph merely as a type of relation, overlooking the social attributes and real-world contexts of signed networks. PyTorch Geometric Signed Directed (PyGSD) He et al. (2023) takes data classes, data loaders, as and data splitters into consideration, and design a specific tool for SGRL.

6 Future Direction

SGRL is a significant branch of graph representation learning. Despite notable advancements in this field, there still exists a plethora of research directions worthy of exploration.

Trustworthy research. GNNs focused on performance have shown potential drawbacks such as susceptibility to adversarial attacks, unexplainable bias against disadvantaged groups, and excessive resource consumption in edge computing settings. To mitigate these unintended consequences, the imperative is to construct proficient GNNs distinguished by their trustworthiness. For signed graph representation learning, current related research is limited to adversarial attack and defense Zhou et al. (2023); Zhang et al. (2023b). There is currently no relevant research in other crucial areas of trustworthy Graph Neural Networks (GNNs), such as interpretability, privacy, fairness, and accountability. Taking explainability research as an example, explainability research is to enhance understanding and trust in complex deep models by providing interpretable insights into their decision-making processes and predictions. Several explainability methods have been introduced to elucidate the underlying mechanisms of unsigned graph neural networks Yuan et al. (2022). However, there is currently no research on the interpretability of the SGNN model. Compared to explainability research on Graph Neural Networks (GNN), Signed GNN possesses two new features, new downstream task (i.e., link sign prediction), and new collective properties (i.e., balance theory and status theory). Therefore, considering the combination of the above two types of features, we believe that designing explainability methods for SGNN is a promising research direction.

Data-centric research. Data-centric AI is the discipline of systematically engineering the data used to build an AI system (by Andrew Ng), which has a broad research content, e.g., data collection, data labeling, data preparation, data augmentation, etc. Graph data augmentation Liu et al. (2022) is undoubtedly the most popular direction in the field currently. These graph data augmentation methods often rely on side information such as node labels and features, making it challenging to directly apply them to enhance signed graph data. However, for signed graphs, the current research efforts are still quite limited Zhang et al. (2023c). Sparse labels represent another prominent challenge in signed graph representation learning. This challenge is also expected to be addressed in data-centric research.

Large Language Models (LLMs). Large language models are currently a hot topic in research. In contrast to non-contextualized shallow textual embeddings, large language models (LLMs) exhibit extensive context-aware knowledge and superior semantic comprehension capabilities, achieved through pre-training on vast text corpora. The utilization of Large Language Models (LLMs) in Graph Representation Learning has seen some attempts in the past year Chen et al. (2023). However, without a doubt, research in this domain is still at a very early stage. The exploration of how to apply LLMs to signed graph representation learning is an intriguing area, and as of now, there is no existing research in this specific domain.

References

  • Bonchi et al. [2019] Francesco Bonchi, Edoardo Galimberti, Aristides Gionis, Bruno Ordozgoiti, and Giancarlo Ruffo. Discovering polarized communities in signed networks. In CIKM, 2019.
  • Cai et al. [2018] Hongyun Cai, Vincent W Zheng, and Kevin Chen-Chuan Chang. A comprehensive survey of graph embedding: Problems, techniques, and applications. TKDE, 2018.
  • Chakraborty et al. [2023] Roshni Chakraborty, Ritwika Das, and Joydeep Chandra. Siggan: Adversarial model for learning signed relationships in networks. TKDD, 2023.
  • Chen et al. [2018] Yiqi Chen, Tieyun Qian, Huan Liu, and Ke Sun. ”bridge”: Enhanced signed directed network embedding. In CIKM, 2018.
  • Chen et al. [2020] Xu Chen, Kun Xiong, Yongfeng Zhang, Long Xia, Dawei Yin, and Jimmy Xiangji Huang. Neural feature-aware recommendation with signed hypergraph convolutional network. TOIS, 2020.
  • Chen et al. [2023] Zhikai Chen, Haitao Mao, Hang Li, Wei Jin, Hongzhi Wen, Xiaochi Wei, Shuaiqiang Wang, Dawei Yin, Wenqi Fan, Hui Liu, et al. Exploring the potential of large language models (llms) in learning on graphs. arXiv preprint arXiv:2307.03393, 2023.
  • Chiang et al. [2012] Kai-Yang Chiang, Joyce Jiyoung Whang, and Inderjit S. Dhillon. Scalable clustering of signed networks using balance normalized cut. In CIKM, 2012.
  • Cucuringu et al. [2019] Mihai Cucuringu, Peter Davies, Aldo Glielmo, and Hemant Tyagi. SPONGE: A generalized eigenproblem for clustering signed networks. In AISTATS, 2019.
  • Cui et al. [2018] Peng Cui, Xiao Wang, Jian Pei, and Wenwu Zhu. A survey on network embedding. TKDE, 2018.
  • Derr et al. [2018] Tyler Derr, Yao Ma, and Jiliang Tang. Signed graph convolutional networks. In ICDM, 2018.
  • Hamilton et al. [2017] William L. Hamilton, Zhitao Ying, and Jure Leskovec. Inductive representation learning on large graphs. In NeurIPS, 2017.
  • He et al. [2020] Xiangnan He, Kuan Deng, Xiang Wang, Yan Li, Yong-Dong Zhang, and Meng Wang. Lightgcn: Simplifying and powering graph convolution network for recommendation. In SIGIR, 2020.
  • He et al. [2022a] Qiang He, Hongwei Du, and Ziwei Liang. Positive influence maximization in signed networks within a limited time. IEEE TCSS, 2022.
  • He et al. [2022b] Yixuan He, Gesine Reinert, Songchao Wang, and Mihai Cucuringu. Sssnet: Semi-supervised signed network clustering. In SDM, 2022.
  • He et al. [2023] Yixuan He, Xitong Zhang, Junjie Huang, Benedek Rozemberczki, Mihai Cucuringu, and Gesine Reinert. PyTorch Geometric Signed Directed: A Software Package on Graph Neural Networks for Signed and Directed Graphs. In LoG, 2023.
  • Heider [1946] Fritz Heider. Attitudes and cognitive organization. The Journal of psychology, 1946.
  • Huang et al. [2019] Junjie Huang, Huawei Shen, Liang Hou, and Xueqi Cheng. Signed graph attention networks. In ICANN, 2019.
  • Huang et al. [2021a] Junjie Huang, Huawei Shen, Qi Cao, Shuchang Tao, and Xueqi Cheng. Signed bipartite graph neural networks. In CIKM, 2021.
  • Huang et al. [2021b] Junjie Huang, Huawei Shen, Liang Hou, and Xueqi Cheng. SDGNN: learning node representation for signed directed networks. In AAAI, 2021.
  • Huang et al. [2022] Zexi Huang, Arlei Silva, and Ambuj Singh. Pole: Polarized embedding for signed networks. In WSDM, 2022.
  • Huang et al. [2023] Junjie Huang, Ruobing Xie, Qi Cao, Huawei Shen, Shaoliang Zhang, Feng Xia, and Xueqi Cheng. Negative can be positive: Signed graph neural networks for recommendation. IPM, 2023.
  • Islam et al. [2018] Mohammad Raihanul Islam, B Aditya Prakash, and Naren Ramakrishnan. Signet: Scalable embeddings for signed networks. In PAKDD. Springer, 2018.
  • Javari et al. [2020] Amin Javari, Tyler Derr, Pouya Esmailian, Jiliang Tang, and Kevin Chen-Chuan Chang. ROSE: role-based signed network embedding. In WWW, 2020.
  • Kim et al. [2018] Junghwan Kim, Haekyu Park, Ji-Eun Lee, and U Kang. SIDE: representation learning in signed directed networks. In WWW, 2018.
  • Kipf and Welling [2017] Thomas N. Kipf and Max Welling. Semi-supervised classification with graph convolutional networks. In ICLR, 2017.
  • Kumar et al. [2016] Srijan Kumar, Francesca Spezzano, VS Subrahmanian, and Christos Faloutsos. Edge weight prediction in weighted signed networks. In ICDM, 2016.
  • Lamb et al. [2020] Luís C. Lamb, Artur S. d’Avila Garcez, Marco Gori, Marcelo O. R. Prates, Pedro H. C. Avelar, and Moshe Y. Vardi. Graph neural networks meet neural-symbolic computing: A survey and perspective. In IJCAI, 2020.
  • Lee et al. [2020] Yeon-Chang Lee, Nayoun Seo, Kyungsik Han, and Sang-Wook Kim. ASiNE: Adversarial signed network embedding. In SIGIR, 2020.
  • Leskovec et al. [2010] Jure Leskovec, Daniel P. Huttenlocher, and Jon M. Kleinberg. Signed networks in social media. In CHI, 2010.
  • Li et al. [2019] Jinning Li, Yirui Gao, Xiaofeng Gao, Yan Shi, and Guihai Chen. Senti2pop: sentiment-aware topic popularity prediction on social media. In ICDM, 2019.
  • Li et al. [2020] Yu Li, Yuan Tian, Jiawei Zhang, and Yi Chang. Learning signed network embedding via graph attention. In AAAI, 2020.
  • Liu et al. [2021] Haoxin Liu, Ziwei Zhang, Peng Cui, Yafeng Zhang, Qiang Cui, Jiashuo Liu, and Wenwu Zhu. Signed graph neural network with latent groups. In KDD, 2021.
  • Liu et al. [2022] Songtao Liu, Rex Ying, Hanze Dong, Lanqing Li, Tingyang Xu, Yu Rong, Peilin Zhao, Junzhou Huang, and Dinghao Wu. Local augmentation for graph neural networks. In ICML, 2022.
  • Liu et al. [2023] Chuang Liu, Yibing Zhan, Jia Wu, Chang Li, Bo Du, Wenbin Hu, Tongliang Liu, and Dacheng Tao. Graph pooling for graph neural networks: Progress, challenges, and opportunities. In IJCAI, 2023.
  • Lu et al. [2019] Chunyu Lu, Pengfei Jiao, Hongtao Liu, Yaping Wang, Hongyan Xu, and Wenjun Wang. Ssne: status signed network embedding. In PAKDD, 2019.
  • Ni et al. [2023] Lin Ni, Sijie Wang, Zeyu Zhang, Xiaoxuan Li, Xianda Zheng, Paul Denny, and Jiamou Liu. Enhancing student performance prediction on learnersourced questions with sgnn-llm synergy. ArXiv preprint, 2023.
  • Pougué-Biyong et al. [2023] John Pougué-Biyong, Akshay Gupta, Aria Haghighi, and Ahmed El-Kishky. Learning stance embeddings from signed social graphs. In WSDM, 2023.
  • Seo et al. [2022] Changwon Seo, Kyeong-Joong Jeong, Sungsu Lim, and Won-Yong Shin. Siren: Sign-aware recommendation using graph neural networks. TNNLS, 2022.
  • Sharma et al. [2023] Kartik Sharma, Mohit Raghavendra, Yeon-Chang Lee, and Srijan Kumar. Representation learning in continuous-time dynamic signed networks. In CIKM, 2023.
  • Shen and Chung [2018] Xiao Shen and Fu-Lai Chung. Deep network embedding for graph representation learning in signed networks. IEEE Transactions on Cybernetics, 2018.
  • Shu et al. [2021] Lin Shu, Erxin Du, Yaomin Chang, Chuan Chen, Zibin Zheng, Xingxing Xing, and Shaofeng Shen. Sgcl: Contrastive representation learning for signed graphs. In CIKM, 2021.
  • Sun et al. [2020] Renjie Sun, Chen Chen, Xiaoyang Wang, Ying Zhang, and Xun Wang. Stable community detection in signed social networks. TKDE, (10), 2020.
  • Tang et al. [2016a] Jiliang Tang, Charu C. Aggarwal, and Huan Liu. Recommendations in signed social networks. In WWW, 2016.
  • Tang et al. [2016b] Jiliang Tang, Yi Chang, Charu Aggarwal, and Huan Liu. A survey of signed network mining in social media. ACM CSUR, 2016.
  • Velickovic et al. [2018] Petar Velickovic, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Liò, and Yoshua Bengio. Graph attention networks. In ICLR, 2018.
  • Wang et al. [2017] Suhang Wang, Jiliang Tang, Charu C. Aggarwal, Yi Chang, and Huan Liu. Signed network embedding in social media. In SDM, 2017.
  • Wang et al. [2018] Hongwei Wang, Fuzheng Zhang, Min Hou, Xing Xie, Minyi Guo, and Qi Liu. SHINE: signed heterogeneous information network embedding for sentiment link prediction. In WSDM, 2018.
  • Wu et al. [2020a] Shiwen Wu, Fei Sun, Wentao Zhang, Xu Xie, and Bin Cui. Graph neural networks in recommender systems: a survey. ACM CSUR, 2020.
  • Wu et al. [2020b] Zonghan Wu, Shirui Pan, Fengwen Chen, Guodong Long, Chengqi Zhang, and S Yu Philip. A comprehensive survey on graph neural networks. TNNLS, 2020.
  • Xiao et al. [2020] Han Xiao, Bruno Ordozgoiti, and Aristides Gionis. Searching for polarization in signed graphs: a local spectral approach. In WWW, 2020.
  • Xie et al. [2020] Ruobing Xie, Cheng Ling, Yalong Wang, Rui Wang, Feng Xia, and Leyu Lin. Deep feedback network for recommendation. In IJCAI, 2020.
  • Ying et al. [2018] Rex Ying, Ruining He, Kaifeng Chen, Pong Eksombatchai, William L. Hamilton, and Jure Leskovec. Graph convolutional neural networks for web-scale recommender systems. In KDD, 2018.
  • Yuan et al. [2017] Shuhan Yuan, Xintao Wu, and Yang Xiang. SNE: Signed Network Embedding. In PAKDD, 2017.
  • Yuan et al. [2022] Hao Yuan, Haiyang Yu, Shurui Gui, and Shuiwang Ji. Explainability in graph neural networks: A taxonomic survey. TPAMI, 2022.
  • Zhang et al. [2020] Ziwei Zhang, Peng Cui, and Wenwu Zhu. Deep learning on graphs: A survey. TKDE, 2020.
  • Zhang et al. [2023a] Zeyu Zhang, Jiamou Liu, Kaiqi Zhao, Song Yang, Xianda Zheng, and Yifei Wang. Contrastive learning for signed bipartite graphs. In SIGIR, 2023.
  • Zhang et al. [2023b] Zeyu Zhang, Jiamou Liu, Xianda Zheng, Yifei Wang, Pengqian Han, Yupan Wang, Kaiqi Zhao, and Zijian Zhang. Rsgnn: A model-agnostic approach for enhancing the robustness of signed graph neural networks. In Web Conference, 2023.
  • Zhang et al. [2023c] Zeyu Zhang, Shuyan Wan, Sijie Wang, Xianda Zheng, Xinrui Zhang, Kaiqi Zhao, Jiamou Liu, and Dong Hao. Sga: A graph augmentation method for signed graph neural networks. ArXiv preprint, 2023.
  • Zheng et al. [2015] Xiaolong Zheng, Daniel Zeng, and Fei-Yue Wang. Social balance in signed networks. Information Systems Frontiers, 2015.
  • Zhou et al. [2020] Jie Zhou, Ganqu Cui, Shengding Hu, Zhengyan Zhang, Cheng Yang, Zhiyuan Liu, Lifeng Wang, Changcheng Li, and Maosong Sun. Graph neural networks: A review of methods and applications. AI Open, 2020.
  • Zhou et al. [2023] Jialong Zhou, Yuni Lai, Jian Ren, and Kai Zhou. Black-box attacks against signed graph analysis via balance poisoning. ArXiv preprint, 2023.
  • Zitnik et al. [2018] Marinka Zitnik, Monica Agrawal, and Jure Leskovec. Modeling polypharmacy side effects with graph convolutional networks. Bioinformatics, 2018.