Deep Manifold Learning with Graph Mining
Abstract
Admittedly, Graph Convolution Network (GCN) has achieved excellent results on graph datasets such as social networks, citation networks, etc. However, softmax used as the decision layer in these frameworks is generally optimized with thousands of iterations via gradient descent. Furthermore, due to ignoring the inner distribution of the graph nodes, the decision layer might lead to an unsatisfactory performance in semi-supervised learning with less label support. To address the referred issues, we propose a novel graph deep model with a non-gradient decision layer for graph mining. Firstly, manifold learning is unified with label local-structure preservation to capture the topological information of the nodes. Moreover, owing to the non-gradient property, closed-form solutions is achieved to be employed as the decision layer for GCN. Particularly, a joint optimization method is designed for this graph model, which extremely accelerates the convergence of the model. Finally, extensive experiments show that the proposed model has achieved state-of-the-art performance compared to the current models.
Index Terms:
Deep graph learning, orthogonal manifold, closed-form solution, topological information.
1 Introduction
Classification,a classic task in machine learning, aims to classify the unlabeled data by learning the information from the labeled data. However, considering the expensive handcraft labeling, privacy, and security, it is difficult to obtain abundant labeled data for training classifiers. In recent years, semi-supervised algorithms trained with scarce labeled data make some progress in the practical application[1]. In general, the semi-supervised models could be divided into two categories, based on label information and based on the data distribution[2] .
Based on label information, these models mainly focus on learning high-quality features from the labeled data, such as co-training method [3, 4, 5, 6]. Besides, latent structure information often helps to enhance classification accuracy [7, 8, 9, 10, 11]. Among them, manifold learning is a classical method and can explore the latent topological information of the data [12]. The KPCA [13] integrates manifold learning into semi-supervised dimensionality reduction. Besides, a novel generalized power iteration (GPI) [14] is employed to solve the problem on the Stiefel manifold. Accordingly, the orthogonal regression manifold (OR) is the least square problem defined on the Stiefel manifold.
However, it is known that the classic convolution neural network mainly deals with regular data such as images and weakly handles irregular data tasks like graph node classification. Some researchers utilize spectral decomposition to explore the structure of the graph and expand the deep network into irregular data space [15, 16, 17, 18]. [15] offers flexibility to compose and evaluate different deep networks using classic machine learning algorithms. Besides, the pseudo-labels generated by label propagation are utilizing to train graph convolution networks[16]. Masked-GCN [17] only propagates some attributes of nodes to the neighbors, which jointly considers the distribution of local neighbors. Apart from that, [19] proposes Propagation-regularization to boost the performance of existing GNN models. Aiming to accelerate the speed of the model, [20] suggests a method called distant compatibility estimation.
Generally, softmax utilizes the deep feature to predict the label and is extensively employed as the decision layer in the graph networks. However, it ignores the distribution and latent structure of the irregular data during classification. Apart from that, this decision layer is optimized via gradient descent with thousands of iterations to obtain the approximate solutions. Besides, due to the lack of sufficient prior labels under the semi-supervised tasks, trivial solutions can not be avoided during the optimization.
To tackle the referred deficiency, we propose a novel deep model for graph knowledge learning with a non-gradient decision layer. Our major contributions are listed by the following items:
-
1)
Unifying the orthogonal manifold with label local-structure preservation to mine the topological information of the deep embeddings, the novel non-gradient graph decision layer is put forward.
-
2)
The theorems are designed to solve the proposed layer with an elegant analytical solution, which accelerates the convergence of the model.
-
3)
A novel deep graph convolution network is proposed and optimized jointly. Moreover, extensive experiments suggest that the model has achieved state-of-the-art results.
2 Notations and Background
2.1 Notations
In this paper, the boldface capital letters and boldface lower letters are employed to present matrics and vectors respectively. Suppose the as the -th row of the matrix . The element of is . Besides, the -norm of the matrix is defined as . The trace and transpose of matrix are denoted as and , respectively. means the the gradient of w.r.t. . And is a unit column vector with dimension c. For the semi-supervised classification, is the feature matrix with samples and dimensions. And we have labeled data points and unlabeled data samples where . Let denotes the labeled and unlabeled with classes. And is initialized according to . Apart from that, we denote as the -th layer hidden feature.
2.2 Background
The classic convolutional neural network is designed to process regular data like images. However, there exists numerous irregular data, such as citation networks and social networks. Owing to the distribution of these data is not shift-invariant, the traditional neural networks could not extract the latent topological information of these data. Aiming to tackle these issues, some researchers try to transfer traditional Laplacian kernel and convolution operators into graph data space by spectral decomposition based on the graph [21, 22, 23]. However, the spectral decomposition of the Laplacian matrix is very time-consuming. Defferrard et al [22] employ the Chebyshev polynomial to fit the results of spectral decomposition and design a Chebyshev convolution kernel as follows
(1) |
where is the normalized feature vector matrix, is the spectral radius, is Chebyshev coefficient vector, and the Chebyshev polynomial is defined as .
Based on this, Kipf et al [24] use the first-order expansion of the Chebyshev polynomial to fit the graph convolution kernel with . To prevent overfitting and reduce complexity, the authors introduce some tricks, , and the diagnonal element in degree matrix is . Finally, the classic GCN is defined as follows
(2) |
Apart from that, manifold learning [25] is a classic nonlinear algorithm for dimensionality reduction, representation learning, and classification tasks. It believes that the data nodes are generated from a low-dimensional manifold embedded in a high-dimensional space and this procedure is often described by a function with few underlying parameters. Based on this idea, manifold learning aims to uncover the function in order to explore the latent topological structure of the data.
3 Deep Manifold Learning with Graph Mining
In this section, we propose a novel graph deep model with a non-gradient decision layer for graph mining. The framework is illustrated in Fig. 1.
3.1 Problem Revisited: GCN with Softmax
As is known to us, Softmax has been widely used as a decision layer in lots of classic neural networks. Moreover, on some computer vision tasks such as image classification, these networks also have made excellent progress and can be generally formulated as
(3) |
where is a neural network such as CNN to extract the deep features, is the -th layer in the network, and is the weight matrix in the -th layer. The latent feature in the -th layer is defined as and the is equivalent to . is the prediction label. The latent feature has the same shape with . is the activation function. Aiming to predict the classes, Softmax maps and normalizes these deep features into real numbers in via the formula defined as
(4) |
The cross-entropy is employed as the loss function and optimized via gradient descent. Aiming to classify the irregular data such as citation network, [24] defines the graph convolution to extract the graph deep features and also classify them via softmax.
From the Eq. (4), softmax mainly forecasts the probability according to the value of the embedding. However, it ignores the latent topological distribution of the graph embedding. For example, in social networks, people ought to have a high degree of similarity to those with whom they socialize frequently. Therefore, introducing the softmax into GCN for irregular data classification may not get the same achievements as utilizing CNN for the regular data without this topological structure. More importantly, the softmax is optimized via gradient descent with thousands of iteration to obtain approximate solutions, whose time complexity is non-linear with the number of the convolutional layers. Aiming to improve the performance of the GCN, we propose a graph neural network with a non-gradient decision layer.
3.2 Graph-based Deep Model with a Non-Gradient Decision Layer
3.2.1 Unify Orthogonal Manifold and Label Local-structure Preservation
Owing to obtaining the closed-form result via the analytical method, the least square regression can also be a classic method to predict the label in machine learning. Apart from this, the softmax has low complexity and high interpretability. In semi-supervised learning, it can be defined as
(5) |
where is the feature matrix, is the projection matrix with classes, and is a indices matrix including one-hot labeled and unlabeled points. We can easily optimize the projection matrix via taking the derivative of Eq. (5) w.r.t and setting it to 0,
(6) |
Unfortunately, in the semi-supervised or unsupervised task, Eq. (6) may cause the trivial solution. The label rate is greatly small in semi-supervised learning, , . Under this circumstance, when , the trivial solution that may be triggered. Apart from the potential trivial solution, directly projecting the feature into the label space may ignore the original data structure.
In order to tackle these problems, we firstly introduce the orthogonal manifold, , , to learn the low-dimensional distribution of the graph data, which can avoid projecting the unlabeled graph embedding into the same class and improve the robustness of the model. Meanwhile, since the structural similarity and local information are often credible (especially in manifold learning), a label local-structure preservation is unified into the learned low-dimensional distribution. It guides the classification via label similarity measured by local structure and the decision layer can be defined as
(7) |
where the element of adjacent matrix is , is the one-hot label in , and is a trade-off parameter. The Eq. 7 can be directly solved with the proposed Theorem 1 and Theorem 2, which is non-gradient and proved in Section 4.
Theorem 1.
Given a feature matrix and a label matrix , the problem can be solved by , where and . is defined as and is generated in the orthogonal complement spaces of .
Theorem 2.
Given a feature matrix , a projection matrix and the labeled matrix a label matrix, the problem defined as
(8) |
can be solved by , where , and Laplacian
(9) |
3.2.2 A Novel Graph Deep Network
Because the orthogonal manifold projection is a linear transform, Eq. (7) has not good performance when facing non-linear problems. As is known to us, the kernel trick and neural network are widely employed to handle the nonlinearity. However, the kernel trick is weak in representation and severely sensitive to the choice of the kernel function. On the contrary, graph convolution not only has an excellent ability in representation learning for irregular data but also successfully mines the complex relationships and interdependence between graph nodes via spectral and spatial operators. Based on this, we finally put forward a novel GCN with a non-gradient decision layer like
(10) |
where the self-loop adjacent matrix is , and the graph kernel is . Moreover, the whole network is jointly optimized via Algorithm 1.
In this network, the non-gradient decision layer not only preserve and utilize the topological information of the deep graph embeddings but also can be solved with the closed-form solutions, which significantly improves the accuracy and accelerates the convergence.
The Merits of the Proposed Model: Compared with the classic softmax decision layer, the proposed model unifies the orthogonal manifold and label local-structure preservation to learn the distribution and topology of the graph nodes, which successfully avoids the trivial solutions and improves the robustness. Moreover, contrasting with optimized via gradient descent with thousands of iterations, our model can be solved with the analytical solutions via the proposed non-gradient theorems. Finally, we propose a novel deep graph convolution network and design a joint optimization strategy.
3.3 Time Complexity
In each iteration of the back propagation, the computaional complexity of graph convolution in Eq. (10) is , where is a sparse matrix and the amount of non-zero entries is . Assume that the proposed deep graph model converges after iterations. The time complexity of a non-gradient decision layer is , where is unlabel nodes. Therefore, the total complexity is . Considering that the solution of the non-gradient decision layer is independent with gradient and there is no need to optimize the classifier every iteration, the complexity can be approximate to .
4 Proof
4.1 Proof of Theorem 1
Since problem is difficult to solve directly, we introduce the matrix generated in the orthogonal complement spaces of according to [26]. The problem is equivalent to the following balanced problem
(11) |
To express more conveniently, we define the matrix and problem can be transformed into
(12) |
where .
Owing to the orthogonal embedding, the problem can be simplified into
(13) |
Notice that the first term in Eq. (13) is equal to a constant under the constraints. Therefore, the problem can be transformed into
(14) |
The following lemma [27] points out that the transformed problem Eq. (14) has a closed-form solution.
Lemma 1.
For , where , the problem
(15) |
can be solved by where , , , and .
According to Lemma 1, can be calculated by
(16) |
where . Therefore, the orthogonal matrix is obtained
(17) |
where means that the first columns of all the rows of is taken.
Dataset | Type | Nodes | Edges | Features | Classes | Label Rate |
---|---|---|---|---|---|---|
Cora | Citation network | 2,708 | 5,429 | 1,433 | 7 | 5.2% |
Citeseer | Citation network | 3,327 | 4,732 | 3,703 | 6 | 3.6% |
Pubmed | Citation network | 19,717 | 44,338 | 500 | 3 | 0.3% |
Coauthor-CS | Coauthor network | 18,333 | 81,894 | 6,805 | 15 | 1.6% |
Coauthor-Phy | Coauthor network | 34,496 | 247,962 | 8,415 | 5 | 57.9% |
4.2 Proof of Theorem 2
It is difficult to optimize the second term in problem (8) directly. Therefore, we transform it like
(18) |
Considering that , the Eq. (18) is formulated as
(19) |
where represents the diagonal element in . Furthermore, the inital problem is written as
(20) |
where the and are the label and unlabel graph feature, respectively.
The first parts in Eq. (20) can be transformed by
(21) |
And the can also be simplified by
(22) |
with .
Owing to a constraint on , the problem Eq. (23) is derivated w.r.t and set to 0. Then, we have
(24) |
4.3 Proof of Lemma 1
Having be decomposed with SVD, we can obtain the . Note that
(25) |
where . Clearly, such that . Hence, we have
(26) |
We can simply set . In other words,
(27) |
Consequently, the lemma is proved.








Input | CORA | CITESEER | PUBMED | Coauthor-CS | Coauthor-Phy | |
---|---|---|---|---|---|---|
DeepEmbedding | ,, | 59.00 1.72 | 59.60 2.29 | 71.10 1.37 | 70.93 0.55 | 85.77 0.08 |
GAE | ,, | 71.50 0.02 | 65.80 0.02 | 72.10 0.01 | 56.84 0.43 | 66.20 0.20 |
GraphEmbedding | , | 75.70 3.83 | 64.70 1.11 | 77.20 4.38 | 28.91 0.02 | 50.32 1.11 |
GCN | ,, | 81.50 0.50 | 70.30 0.50 | 78.33 0.70 | 78.12 3.42 | 92.64 3.80 |
DGI | , | 82.30 0.60 | 71.80 0.70 | 76.80 0.60 | 84.34 0.09 | 92.54 0.03 |
PFS-LS | , | 70.09 1.38 | 43.23 1.02 | 74.42 14.2 | 58.80 14.91 | 91.16 6.55 |
APPNP | ,, | 85.09 0.25 | 71.93 2.00 | 79.73 0.31 | 91.52 0.13 | 95.56 0.48 |
GCN-LPA | ,, | 77.28 0.16 | 70.72 0.16 | 73.63 0.40 | 87.63 0.38 | 95.80 0.03 |
MulGraph | ,, | 86.80 0.50 | 73.30 0.50 | 80.10 0.70 | 88.36 0.09 | 94.30 0.01 |
Ours | ,, | 85.65 0.76 | 74.80 0.41 | 80.87 0.17 | 91.60 0.23 | 95.89 0.09 |
5 Experiment
To verify the performance of the proposed model, in this section we will compare the proposed model with seven state-of-the-art methods on citation network benchmark datasets.
5.1 Benchmark Dataset Description
We employ citation networks and coauthor networks to evaluate the performance of models:
Citation Networks: The three classic citation networks involved in our experiments are Cora, Citeseer, and Pubmed, which are closely following the experimental setup in [18]. In these networks, the nodes are articles and the edge reflects the citations between these articles.
Coauthor Networks: The two co-authorship networks [29], Coauthor-CS and Coauthor-Phy, are also used for evaluation. In these networks, the nodes are authors and the edge suggests that two co-authored a paper. These features represent the keywords of the author’s papers. Besides, the coauthor networks are split according to [30].
The detail of these datasets is summarized in TABLE I.
5.2 Experiment Setting
The ratio of labeled data for training is divided according to TABLE I. Like the [18], we employ the extra 500 labeled nodes as a valid set to optimize the hyperparameters including the hidden units in each layer, weight coefficient in Eq. (10), dropout, and learning rate.





Before classification, the adjacent and feature matrix are accordingly (row-)normalized to the range of . The two graph convolution layers are used as a deep graph feature extractor. For the hidden units in each layer, dropout rate, learning rate, and weight coefficient , we choose values of parameters from , , and respectively via 4-fold cross-validation and record the best records. The initialization described in [31] is employed to initialize the network weights. We train the model for a maximum of 500 training iterations with batch gradient descent. After tuning the hyperparameters, we adopt the adam optimizer with a learning rate of 0.01 and optimize the orthogonal manifold classifier every 50 iterations. Accuracy is employed to evaluate the performance of the model.
5.3 Comparsion with Methods
We compare the proposed model with eight state-of-the-art semi-supervised models including Graph neural networks with personalized pagerank () [32], Unifying GCN and label propagation (-) [30], Multi-graph representation learning () [33], Deep graph infomax () [34],Parameter-Free Similarity of Label and Side Information (-) [35], Graph Convolution Network () [24], Variational Graph Auto-Encoder () [36], Semi-supervised learning with Graph Embedding ()[18] and Deep learning via semi-supervised embedding () [37]. Among them, - is a non-GNN model.
To be fair, we utilize the same experiment setting in section 5.2 train these comparison models and tune the correlated hyper-parameters. For classification models based graph, we utilize the self-loop adjacent to train as the same as in Eq. [10]. The models mentioned above are evaluated 10 times on the benchmark datasets. The accuracy and standard deviation are reported in TABLE II. Therefore, we could conclude that:
-
1)
The proposed model has achieved state-of-the-art accuracy with respect to the comparison methods. For example, on Citeseer and Pubmed dataset, we achieve and accuracy respectively. The scores are higher than the other methods, which indicates that the proposed model could mine the conceal relationship among points and work well on classification tasks with few labeled data.
-
2)
When dealing with the datasets with high dimensions, huge points, and low labeling ratios like Pubmed and Coauthor-CS, it has a relative improvement over previous state-of-the-art. Besides, the proposed model has a good performance on Coauthor-Phy which has nearly 2.5 million edges and complex topological relationships.
-
3)
The proposed model is more successful and reasonable to integrate the graph topological structure with the label information than some state-of-the-art models like GCN-LPA, APPNP, and MulGraph. Among them, GCN-LPA utilizes label propagation to learn the edge weights of graphs. Compared with these models, ours integrates manifold learning with label local-structure as a non-gradient classifier. In this way, the deep graph networks could extract meaningful features to contribute to the decision layer.
The proposed model and comparison models are all implemented with PyTorch 1.2.0 on Windows 10 PC.
5.4 Convergence & Visualization
Fig. 3 shows the results on each benchmark datasets. It suggests that the proposed model converges rapidly on each dataset with during the training step. The decision layer (defined in Eq. (7)) is optimized after every 100, 50, and 80 backpropagation on each dataset respectively, which contributes to the objective value dropped sharply and accelerates the training of the proposed model. Besides, we investigate the impacts of the dropout in graph network and parameter (defined in Eq. (10)) on the model. The dropout ratio and value are chosen like in Sec. 5.2. Convergence Epoch and Accuracy are utilized to evaluate the training speed and classification performance of the model. The result is shown in Fig. 4. From the results, we notice that the Convergence Epoch varies considerably with the changing. On the contrary, Accuracy is not sensitive to weight value. In conclusion, the mainly has an impact on the convergence rate of the model, and the proper dropout ratio could enhance the performance of the proposed model via discarding some neurons during the training.
Apart from that, we utilize the t-SNE [28] to reduce the deep graph embedding (defined in Eq. (10)) to 2D. Before visualize, these embeddings are not normalized and the dimension is not reduced via PCA and normalized. The perplexity of t-SNE is set to 30. And the visualization of the reduction results in Fig. 2. We notice that the topological data is not preserved well at the beginning of the training. When the algorithm has converged, the distribution of the deep graph embedding is evenly and could be easy to classify.
Cora | Citeseer | Pubmed | |
---|---|---|---|
GCN-Softmax | 78.99 0.61 | 70.39 0.58 | 78.39 0.61 |
GCN-LP | 79.70 0.52 | 71.52 0.28 | 78.67 0.47 |
GCN-OM | 83.91 0.23 | 73.71 0.16 | 79.82 0.72 |
OURS | 85.65 0.76 | 74.80 0.41 | 80.87 0.17 |
5.5 Ablation Study
We conduct an ablation study to evaluate how each part of the proposed non-gradient graph layer, Orthogonal Manifold (OM) and Local-structure Preservation (LP), contribute to overall model performances. Besides, GCN with a softmax classification strategy is employed as a baseline. The setting and results are summarized in Table III. We confirm that graph mining methods, orthogonal manifold learning and local-structure preservation, can both learn the relationship knowledge among the deep embedding. Besides, it is shown that the accuracy of GCN-OR and GCN-LP are both superior to traditional GCN-softmax. Moreover, the proposed deep model with a non-decision layer not only properly unifies the OM and LP parts but also successfully utilizes the topological structure and label information to enhance the performance.
6 Conclusion
In this paper, we propose a deep graph model with a non-gradient decision layer. Unifying the orthogonal manifold and label local-structure preservation, the deep graph model successfully learns the topological knowledge of the deep embedding. Furthermore, compared with softmax optimized via gradient descent, the non-gradient layer can be solved with the analytical solutions via the proposed theorems. On the benchmark datasets, the proposed model can achieve higher accuracy over the previous state-of-the-art works.
References
- [1] Xiaojin Zhu and Andrew B Goldberg, “Introduction to semi-supervised learning,” Synthesis lectures on artificial intelligence and machine learning, vol. 3, no. 1, pp. 1–130, 2009.
- [2] Xiaojin Jerry Zhu, “Semi-supervised learning literature survey,” Tech. Rep., University of Wisconsin-Madison Department of Computer Sciences, 2005.
- [3] Zhi-Hua Zhou and Ming Li, “Semi-supervised regression with co-training.,” in IJCAI, 2005, vol. 5, pp. 908–913.
- [4] Vikas Sindhwani, Partha Niyogi, and Mikhail Belkin, “A co-regularization approach to semi-supervised learning with multiple views,” in Proceedings of ICML workshop on learning with multiple views. Citeseer, 2005, vol. 2005, pp. 74–79.
- [5] Zhi-Hua Zhou and Ming Li, “Semisupervised regression with cotraining-style algorithms,” IEEE Transactions on Knowledge and Data Engineering, vol. 19, no. 11, pp. 1479–1493, 2007.
- [6] Ching-Hao Mao, Hahn-Ming Lee, Devi Parikh, Tsuhan Chen, and Si-Yu Huang, “Semi-supervised co-training and active learning based approach for multi-view intrusion detection,” in Proceedings of the 2009 ACM symposium on Applied Computing, 2009, pp. 2042–2048.
- [7] X. Zhu, “Learning from labeled and unlabeled data with label propagation,” Tech Report, 2002.
- [8] Tianhang Long, Junbin Gao, Mingyan Yang, Yongli Hu, and Baocai Yin, “Locality preserving projection via deep neural network,” in 2019 International Joint Conference on Neural Networks (IJCNN). IEEE, 2019, pp. 1–8.
- [9] Ahmet Iscen, Giorgos Tolias, Yannis Avrithis, and Ondrej Chum, “Label propagation for deep semi-supervised learning,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2019, pp. 5070–5079.
- [10] Mingchen Gao, Ziyue Xu, Le Lu, Aaron Wu, Isabella Nogues, Ronald M Summers, and Daniel J Mollura, “Segmentation label propagation using deep convolutional neural networks and dense conditional random field,” in 2016 IEEE 13th International Symposium on Biomedical Imaging (ISBI). IEEE, 2016, pp. 1265–1268.
- [11] Y. Yi, Y. Chen, J. Dai, X. Gui, C. Chen, Gang Lei, and W. Wang, “Semi-supervised ridge regression with adaptive graph-based label propagation,” Applied Sciences, vol. 8, no. 12, 2018.
- [12] Antonio Criminisi, Jamie Shotton, Ender Konukoglu, et al., “Decision forests: A unified framework for classification, regression, density estimation, manifold learning and semi-supervised learning,” 2012.
- [13] Ratthachat Chatpatanasiri and Boonserm Kijsirikul, “A unified semi-supervised dimensionality reduction framework for manifold learning,” Neurocomputing, vol. 73, no. 10-12, pp. 1631–1640, 2010.
- [14] “A generalized power iteration method for solving quadratic problem on the stiefel manifold,” Science China(Information Sciences), 2017.
- [15] Rahul Ragesh, Sundararajan Sellamanickam, Vijay Lingam, and Arun Iyer, “A graph convolutional network composition framework for semi-supervised classification,” 2020.
- [16] Hande Dong, Jiawei Chen, Fuli Feng, Xiangnan He, Shuxian Bi, Zhaolin Ding, and Peng Cui, “On the equivalence of decoupled graph convolution network and label propagation,” 2021.
- [17] Liang Yang, Fan Wu, Yingkui Wang, Junhua Gu, and Yuanfang Guo, “Masked graph convolutional network,” in Twenty-Eighth International Joint Conference on Artificial Intelligence IJCAI-19, 2019.
- [18] Zhilin Yang, William Cohen, and Ruslan Salakhudinov, “Revisiting semi-supervised learning with graph embeddings,” in International conference on machine learning. PMLR, 2016, pp. 40–48.
- [19] Han Yang, Kaili Ma, and James Cheng, “Rethinking graph regularization for graph neural networks,” 2020.
- [20] Krishna Kumar P., Paul Langton, and Wolfgang Gatterbauer, “Factorized graph representations for semi-supervised learning from sparse data,” 2020.
- [21] Joan Bruna, Wojciech Zaremba, Arthur Szlam, and Yann LeCun, “Spectral networks and locally connected networks on graphs,” arXiv preprint arXiv:1312.6203, 2013.
- [22] Michaël Defferrard, Xavier Bresson, and Pierre Vandergheynst, “Convolutional neural networks on graphs with fast localized spectral filtering,” in Advances in neural information processing systems, 2016, pp. 3844–3852.
- [23] David K Hammond, Pierre Vandergheynst, and Rémi Gribonval, “Wavelets on graphs via spectral graph theory,” Applied and Computational Harmonic Analysis, vol. 30, no. 2, pp. 129–150, 2011.
- [24] Thomas N Kipf and Max Welling, “Semi-supervised classification with graph convolutional networks,” arXiv preprint arXiv:1609.02907, 2016.
- [25] Lawrence Cayton, “Algorithms for manifold learning,” Univ. of California at San Diego Tech. Rep, vol. 12, no. 1-17, pp. 1, 2005.
- [26] Rui Zhang, Feiping Nie, and Xuelong Li, “Feature selection under regularized orthogonal least square regression with optimal scaling,” Neurocomputing, vol. 273, no. jan.17, pp. 547–553, 2017.
- [27] R. Zhang, H. Zhang, and X. Li, “Robust multi-task learning with flexible manifold constraint,” IEEE Transactions on Pattern Analysis and Machine Intelligence, pp. 1–1, 2020.
- [28] Van Der Maaten Laurens and Geoffrey Hinton, “Visualizing data using t-sne,” Journal of Machine Learning Research, vol. 9, no. 2605, pp. 2579–2605, 2008.
- [29] Oleksandr Shchur, Maximilian Mumme, Aleksandar Bojchevski, and Stephan Günnemann, “Pitfalls of graph neural network evaluation,” 2018.
- [30] Hongwei Wang and Jure Leskovec, “Unifying graph convolutional neural networks and label propagation,” 2020.
- [31] Xavier Glorot and Yoshua Bengio, “Understanding the difficulty of training deep feedforward neural networks,” in Proceedings of the thirteenth international conference on artificial intelligence and statistics, 2010, pp. 249–256.
- [32] Johannes Klicpera, Aleksandar Bojchevski, and Stephan Günnemann, “Predict then propagate: Graph neural networks meet personalized pagerank,” 2019.
- [33] Kaveh Hassani and Amir Hosein Khasahmadi, “Contrastive multi-view representation learning on graphs,” arXiv preprint arXiv:2006.05582, 2020.
- [34] Petar Velickovic, William Fedus, William L Hamilton, Pietro Lio, Yoshua Bengio, and R Devon Hjelm, “Deep graph infomax.,” in ICLR (Poster), 2019.
- [35] R. Zhang, F. Nie, and X. Li, “Semisupervised learning with parameter-free similarity of label and side information,” IEEE Transactions on Neural Networks and Learning Systems, vol. 30, no. 2, pp. 405–414, 2019.
- [36] Thomas N Kipf and Max Welling, “Variational graph auto-encoders,” arXiv preprint arXiv:1611.07308, 2016.
- [37] Jason Weston, Frédéric Ratle, Hossein Mobahi, and Ronan Collobert, “Deep learning via semi-supervised embedding,” in Neural networks: Tricks of the trade, pp. 639–655. Springer, 2012.