An Introduction to Robust Graph Convolutional Networks
Abstract.
Graph convolutional neural networks (GCNs) generalize tradition convolutional neural networks (CNNs) from low-dimensional regular graphs (e.g., image) to high dimensional irregular graphs (e.g., text documents on word embeddings). Due to inevitable faulty data collection instruments, deceptive data manipulation, or other system errors, the data might be error-contaminated. Even a small amount of error such as noise can compromise the ability of GCNs and render them inadmissible to a large extent. The key challenge is how to effectively and efficiently employ GCNs in the presence of erroneous data. In this paper, we propose a novel Robust Graph Convolutional Neural Networks for possible erroneous single-view or multi-view data where data may come from multiple sources. By incorporating an extra layers via Autoencoders into traditional graph convolutional networks, we characterize and handle typical error models explicitly. Experimental results on various real-world datasets demonstrate the superiority of the proposed model over the baseline methods and its robustness against different types of error.
1. Introduction
Traditional Convolutional Neural Networks (CNN) (LeCun et al., 1998) have shown promising capability to extract meaningful features in big datasets, where observations are represented by regular graphs (or grids) such as image (Cao et al., 2018), video (Kang et al., 2016) and speech (Abdel-Hamid et al., 2013). CNNs obtain the features by identifying the local shared properties across the data via localized convolutional filters or kernels. Although the preliminary CNN’s kernels are hand-engineered, recent CNNs learn them from the data.
Text documents on word embeddings, social networks or biological networks that can be represented as graphs lying on irregular or non-Euclidean domains has emerged as a topic of critical significance among the data mining and machine learning community. However, using CNNs for the graphs with irregular structure is not trivial as the convolution and pooling operators are only defined for regular graphs such as grids. To generalize CNNs for irregular graphs effectively and efficiently, Defferrard et al. (Defferrard et al., 2016) proposed a formulation of CNNs, named as Graph Convolutional Neural Networks (GCNs) in the domain of spectral graph theory, which is suitable for both regular and irregular graphs (Chung, 1997). Precisely, GCNs are based on spectral graph theoretical formulation of CNNs on graphs built on graph signal processing with filters with low computational complexity (Shuman et al., 2013).

Enjoying the baseball game from the third-base seats. Ichiro Suzuki preparing to bat.
A challenging problem may arise when the data is contaminated by error due to sensor failures, malicious tampering, or API limitation. Error could exist in either the data or label if any available. It could exhibit as noise which is a slight perturbation of random subset of entries in data, or feature-specific corruptions corresponding to perturbation of a set of random features (Liu et al., 2013). Fig. 1 illustrates these two types of error. With noise and feature-specific corruptions, although error exists only on portions of data, it can adversarially affect the capability of GCNs to learn from data. Consequently, they suffer from performance degradation when applied on erroneous data. For this reason, error-robust GCNs are highly desired.
Existing work on deep error-robust learning can be roughly classified into four categories: incorporating extra layers, modifying the loss function, cleaning up data, and enhancing deep learning model capability by using adversarial training. Adding extra layers to deep learning methods accounts for either error model e.g., noise transition layer or cleaning up data (Han et al., 2018; Wang et al., 2019), while modifying the loss function is done either by replacing it with error-tolerant loss or changing the loss with regularization biases (Li et al., 2016; Zhou and Paffenroth, 2017; Pu et al., 2019).
The basic idea of cleaning up the data is to first filter out those data instances that tend to be error-contaminated with high confidence and then trains the deep learning model with only the remaining ones (Jiang et al., 2018). This approach is often done in a two-step manner (sequentially). Adversarial examples leads deep learning models to learn superficial data statistics and causes significant risk for the models deployed in safety-critical systems such as user authentication. They are often created by injecting noise into clean data instances. Adversarial training improves robustness of the deep learning models against adversarial examples by training the models with them (Li et al., 2016).
To improve robustness of GCN against error in data, we propose a novel Robust Graph Convolutional Neural Networks (RGCN) based on GCN. The key idea is that unlike GCN, the proposed RGCN model uses extra layer in the form of robust non-linear deep autoencoders to deal with erroneous data elegantly. We employ two variants of autoencoders: denoising autoencoders and robust low-rank autoencoders. With the assumption of clean training data, the denoising autoencoders inject error into the data and learn from the erroneous data, while robust low-rank autoencoders assume possibly erroneous data and remove sparse error from it and obtain its clean low-rank approximation.
The decomposition of possibly erroneous data into clean data and sparse error in robust low-rank autoencoders facilitates robust recovery of low-rank clean component as well as extraction of sparse error component. To capture error as sparse part, the proposed RGCN model imposes norm on error component. With the popularity of multi-view data, where data may come from multiple sources, we also propose a novel generalization of RGCN for multi-view data (MVRGCN). For example, in Wikipedia, concept of cat may be represented by various views in the form of image (view 1), text (view 2), or even audio (view 3) and videos (view 4). Our main contributions can be summarized as follows:
-
•
To the best of our knowledge, the proposed RGCN and MVRGCN models are the first work that recovers clean data via robust low-rank autoencoders, while capturing error for erroneous data.
-
•
The proposed RGCN and MVRGCN models is the first method based on low-rank decomposition that isolates the clean low-rank approximation from error for erroneous data.
-
•
Through extensive experiments on real-world datasets, we show that the proposed RGCN and MVRGCN are superior to several state-of-the-art methods in supervised and semi-supervised learning and robust against error in text and image datasets.
2. Robust Graph Convolutional Neural Network
We begin with some necessary notations and concepts of Graph Convolutional Neural Networks (GCNs) and Robust Autoencoders. Table 1 lists basic symbols that will be used throughout the paper.
Symbol | Definition and description |
---|---|
lowercase letter represents a scale | |
boldface lowercase letter represents a vector | |
boldface uppercase letter represents a matrix | |
(Frobenius) norm of matrix | |
-norm of matrix | |
sum of the singular values of |
2.1. Graph Convolutional Neural Network
Graph Convolutional Neural Networks (GCNs) (Defferrard et al., 2016) is a generalization of traditional Convolutional Neural Networks (CNNs) that can be efficiently and effectively used for graphs with irregular structure. Examples include documents on word embeddings, social networks, and gene data networks. Different from traditional CNNs, GCNs employ localized graph convolution in the spectral domain via Laplacian matrix and graph Fourier transform (GFT). Normalized is defined as , where ( refers to number vertices in the graph), is the diagonal matrix with , and is the identity matrix.
Since is a real symmetric positive semidefinite matrix, it can be decomposed into , where is the matrix of eigenvectors with , and is the diagonal matrix of eigenvalues. Let be a signal defined on the vertices of the graph, where indicates the value of the signal at the -th vertex. The GFT is obtained via , and it converts signal to the spectral domain spanned via the Fourier basis . Based on these concepts, graph convolution can be formulated as follows (Defferrard et al., 2016):
(1) |
where the parameter is a vector of Fourier coefficients, and is the filter which is a function of . For fast filtering, can be approximated by Chebyshev polynomials of order as follows (Defferrard et al., 2016):
(2) |
where the parameter indicates a vector of Chebyshev coefficients and refers to the Chebyshev polynomial of order evaluated at . By substituting Eq. (2) into Eq. (1) results in , where . We can use the recurrence relation to compute , if we denote , and . For learning filter, the output feature map of the sample can be obtained as follows (Defferrard et al., 2016):
(3) |
where denotes the input feature maps and are vectors of Chebyshev coefficients refer to the layer’s trainable parameters. Like pooling in traditional CNNs, pooling for GCNs requires neighborhoods on graphs, where similar vertices are grouped together. Likewise, multi-scale clustering of the graph can be used for pooling for multiple layers in GCNs. The two final layers in GCN are fully connected layer with an regularization on the weights and softmax regression. The loss function for GCN include cross entropy with an regularization.
2.2. Robust Autoencoders
An autoencoder is a neural network which is trained to attempt to copy its input to its output. Precisely, it learns a mapping from the input to itself through a pair of encoding and decoding phases as follows:
(4) |
where is the input data where denotes number of data points and represents number of features, is the encoding function which maps input data to the hidden layer, is a decoding function which maps from the hidden layer to the output layer, and is the reconstructed input (or recovered input). The objective function for autoencoder can be formulated as follows:
(5) |
where is -norm. Autoencoder can be used for dimensionality reduction or feature learning when the identity mapping is not desired. That can be achieved either by regularization or hidden layers that learn low-dimensional and non-linear representation of input data. An autoencoder with more than one hidden layer is referred to as a deep autoencoder. Each additional hidden layer is a pair of encoder and decoder. A denoising autoencoder instead minimizes the following loss function:


(6) |
where is a copy of that has been contaminated by some type of error. The denoising autoencoder is an autoencoder that receives erroneous data as input and is trained to predict the original, uncontaminated data point as its output i.e., reconstructed component. For this reason, the denoising autoencoder has shown to be useful for improving robustness as well as generalizability of the model.
Motivated by robust principle component analysis (RPCA) (Candès et al., 2011), an robust low-rank autoencoder outputs a low-rank clean approximation of input data as well as sparse error component (Zhou and Paffenroth, 2017). The key idea of RPCA is to decompose the possibly erroneous input data into low-rank approximation part and sparse error component. By explicitly accounting for sparse error component to model error in data, the quality of low-rank representation can be significantly improved. Based on this motivation, the loss function for robust low-rank autoencoder is formulated as rank minimization as follows:
(7) |
where represents low-rank clean approximation or representation of , refers to the rank of , captures error in the possibly erroneous input data , and is trade-off parameter. Since and -norm are non-convex, the objective function in Eq. (7) is an instance of NP-hard problem. One natural way is to replace with the trace norm and -norm with norm. The resulted objective function (or loss function) is as follows:
(8) |
The trace norm is the convex envelope of the rank. As a result, minimization of the trace norm is equivalent to low-rank structure. The -norm of is well-known to be a convex surrogate of . Algorithm 1 shows the optimization procedure for the challenging objective function Eq. (8) using Alternating Direction Method of Multipliers (ADMM) (Zhou and Paffenroth, 2017). The ADMM minimizes one part of the loss function while the others are fixed. More precisely, to obtain , we fix other variables such as . We train autoencoder part using backpropagation. To approximate , we keep other variables fixed. can be then solved using proximal operator for -norm defined as follows (Lin et al., 2010):
(9) |
Input:
Parameter: ,
Output: ,

2.3. Problem Formulation
We state the problem for error-robust GCN as follows:
Problem. In the setting of supervised or semi-supervised learning, given distinct possibly erroneous training data points (or samples), the goal is to learn a classifier that predicts label for unlabeled data points effectively and efficiently using GCN such that it is robust against error in data. GCN can effectively capture the nonlinearity of data points and possess strong capability to exploit graph characteristics. The training set is represented as , where ( denotes number of unlabeled data points and indicates number of labeled data points), and denotes the number of features. For supervised learning, , while for semi-supervised learning, .
2.4. Model Architecture
To alleviate sensitivity of GCN against error in data, we utilize robust autoencoder. With that aim, we introduce three architectures for robust GCN (RGCN) and multi-view RGCN (MVRGCN) as follows:
Architecture 1. In this architecture, GCN receives low-rank clean component of robust low-rank deep autoencoder as input data. The robust low-rank deep autoencoder eliminates error from the possibly erroneous data and returns it as low-rank clean component. Note that the low-rank clean approximation has the same dimensionality as possibly erroneous input data. The loss function for this proposed robust GCN (RGCN) has two main terms: loss for GCN and loss for robust low-rank deep autoencoder. Fig. 2 shows the proposed architecture.
Architecture 2. In this architecture, after cleaning up the input data clean training data, GCN receives reconstructed component of denoising deep autoencoder as input data. Fig. 2 illustrates the proposed architecture.
Architecture 3. We also propose a novel generalization of RGCN for multi-view data (MVRGCN) where data are represented by multiple views. Each view may be described by an arbitrary type and number of features. With the aim of exploiting complementary and consistent information from multiple views rather than relying on the individual view, we use a separate RCGN with respect to each view and then combine the learned feature matrices using pooling to establish common feature matrix across all views. The pooling layer could be max-pooling, average-pooling, or mixed-pooling that is a linear combination of max-pooling and average-pooling with trade-off parameter i.e., mixed_pooling = max_pooling + average_pooling. The schematic representation of the proposed MVRGCN is shown in Fig. 3.
2.5. Model Optimization
In all of the three proposed architectures, robust deep autoencoders and GCN are trained in a joint manner. For Architecture 1, we optimize the proposed model through Adaptive Moment Estimation (ADAM) (Kingma and Ba, 2014) over shuffled mini-batches. The optimization for Architecture 2 requires joint ADAM and ADMM on shuffled mini-batches. For Architecture 3, we train through ADAM and/or ADMM over shuffled mini-batches depending on the type of robust low-rank deep autoencoders.
3. Experimental Evaluation
To evaluate the effectiveness of the proposed RGCN models, we conduct experiments on real-world datasets and compare RGCN with the following baselines: (1) Softmax. (2) Fully Connected (FCk) denotes a fully connected layer with hidden units. (3) Support Vector Machine (SVM). (4) Multinomial Naive Bayes (MNB). (5) Robust Deep Autoenoder (RDAE) denotes robust deep autoencoders with regularization and softmax (Zhou and Paffenroth, 2017). (6) GCNk denotes GCN with graph convolutional layer with feature maps (Defferrard et al., 2016).
For multi-view datasets, we compare the proposed MVRGCN model with the following baselines: (1) Best Single View (BSV) feeds the most informative view to Softmax. (2) Concatenation (Concat.) - Softmax concatenates all views and then feeds it into Softmax. (3) Concatenation (Concat.) - Multinomial Naive Bayes (MNB) concatenates all views and then feeds it into Multinomial Naive Bayes. (4) Concatenation (Concat.) - GCN concatenates all views and then feeds it into GCN. (5) Concatenation (Concat.) - RGCN concatenates all views and then gives it as input to RGCN using robust low-rank deep autoencoders.
Similar to (Defferrard et al., 2016), we use Accuracy as performance evaluation metric. Each experiment is repeated for three times, and the mean of each metric in each dataset is reported. We set learning rate for optimization to , learning rate decay , and momentum of . The batch size is set to be . We apply linear search for graph convolution (or filter) size from for Chebyshev, Fourier and Spline, and only report the result for the best option. The pooling layer inside GCN is of size . We only report the results for the proposed RGCN model with denoising deep autoencoders on clean dataset becasue we could not successfully clean up the erroneous datasets.
Dataset | # training dp. | # testing dp. | # views | # classes |
---|---|---|---|---|
20News | 10171 | 7532 | 1 | 20 |
MNIST | 60000 | 10000 | 1 | 10 |
Fox | 1000 | 523 | 2 | 4 |
We implement the proposed RGCN and MVRGCN models to investigate their effectiveness: RGCNk(RLDAE) denotes Architecture 1 with graph convolutional layer with feature maps, RGCNk(DDAE) which represents Architecture 2 with graph convolutional layer with feature maps. MVGCNk denotes the proposed GCN model for multi-view dataset with graph convolutional layers with feature maps. MVRGCNk(RLDAE) denotes multi-view GCN with robust low-rank deep autoencoders with graph convolutional layers with feature maps, while MVRGCNk(DDAE) indicates multi-view GCN with denoising autoencoders with graph convolutional layers with feature maps. Statistics of the real-world datasets are summarized in Table 2 (training dp. denotes training data points and testing dp. indicates testing data points).
3.1. Text Categorization on 20News
Similar to (Defferrard et al., 2016), the preprocessing steps consist of removing short documents and infrequent words. We then extract 10000 most common words from the unique words across all remaining documents. Each document is modeled as a bag of words and normalized across words. The feature graph is a 16 nearest neighbours, where vertices are those 10000 words, and edges represent the similarity between the words computed based on their corresponding word2vec embeddings using the following formula:
(10) |
where denotes edge weight between vertices and , and and indicate word2vec embeddings for vertices and , respectively. denotes standard deviation. We obtain pre-trained word2vec word embeddings from GoogleNew-vectors-negative300. Table 3 presents the comparison results on this dataset. The robust autoencoders is of hidden layers with 5000 and 350 units. Although the proposed RGCN models are not superior to MNB on clean 20News (first column), they outperform the fully connected networks which has much more trainable parameters and is thus of higher space complexity.
To evaluate the robustness of the proposed RGCN methods on noise, we add masking noise (feature specific corruption) with various levels on 20News training data points (denoted as E20News0.2 and E20News0.4). In masking noise, a set of random features are assigned value . In our experiments, we set . Table 3 also presents the results on erroneous 20News (second and third columns). The proposed RGCN model with robust low-rank deep autoencoders outperforms the baselines which demonstrates its robustness against error of type of masking noise.
Method | 20News | E20News0.2 | E20News0.4 |
---|---|---|---|
Softmax | 66.28 | 61.45 | 58.12 |
FC2500 | 64.64 | 59.08 | 55.67 |
FC2500-FC500 | 65.76 | 54.87 | 53.12 |
SVM | 65.90 | 62.07 | 60.00 |
MNB | 68.51 | 64.04 | 62.00 |
RDAE | 65.70 | 62.90 | 61.05 |
GCN32 | 68.26 | 63.30 | 60.40 |
RGCN32(RLDAE) | 67.90 | 66.21 | 65.31 |
RGCN32(DDAE) | 67.01 | NA | NA |
3.2. Image Classification on MNIST
Although the main motivation of the proposed RGCN models is related to graphs with irregular structure, we apply them on image datasets which can be represented as graphs with grid (or regular) structure. In MNIST dataset, each data point in the dataset is a 2D grid of size . The feature graph is constructed as an 8 nearest neighbours graph of the 2D grid. The edge weight is computed by the formula in Eq. (10).
Table 4 presents the results on this dataset. The proposed RGCN model with robust low-rank deep autoencoders outperforms the baselines on clean MNIST (first column). The robust deep autoencoders is of hidden layers with 392 and 196 units. We investigate the robustness of the proposed RCGN models against error by adding Gaussian noise to MNIST training data points. Gaussian noise is one of the popular noise models for image datasets. The mean of Gaussian distribution is set to , while standard deviation (or noise level) is chosen from denoted as EMNIST0.01 and EMNIST0.02, respectively. Table 4 also presents the results for erroneous MNIST dataset. The proposed RGCN model with robust low-rank deep autoencoders is consistently superior to the baselines and therefore robust against this error type.
Method | MNIST | EMNIST0.02 | EMNIST0.04 |
---|---|---|---|
Softmax | 89.78 | 88.09 | 85.14 |
SVM | 98.02 | 95.00 | 93.01 |
MNB | 83.31 | 80.00 | 76.50 |
RDAE | 92.00 | 89.00 | 87.43 |
GCN10 | 97.47 | 95.42 | 93.04 |
RGCN10(RLDAE) | 98.00 | 97.03 | 96.00 |
RGCN10(DDAE) | 98.03 | NA | NA |
3.3. Semi-Supervised Classification on Fox
With the aim of evaluating the effectiveness of the proposed MVRCGN model, we conduct experiments on FOX111https://sites.google.com/site/qianmingjie/home/datasets/ which is a small dataset and has two views (view1: text, view2: image). Text view contains text about news, while image view consists of one image about the corresponding news. The preprocessing for text view consists of removing infrequent words, computing TFIDF for 2711 top words. For image view, we normalize the images.
Since number of features in each view is different, the input data is a similarity graph, where vertices are data points and edge weight is computed by cosine similarity for text view and Euclidean distance for image view. The feature graph for text view is constructed based on Euclidean distance between word2vec embeddings of those 2711 top words. We use Eq. (10) to compute edge weight for text view. For image view, the feature graph is a grid of the same size as number of pixels in each image. View fusion in the proposed MVGCN and MVRGCN models are done using max-pooling. Table 5 presents the results on Fox dataset. The proposed MVGCN model is superior to the baselines on clean Fox (first column).
Method | Fox | EFox0.2 | EFox0.4 |
---|---|---|---|
BSV | 65.19 | 59.87 | 56.02 |
Concat.-Softmax | 60.83 | 55.64 | 53.00 |
Concat.-MNB. | 65.40 | 61.56 | 58.20 |
Concat.-GCN32 | 68.60 | 60.00 | 56.45 |
Concat.-RGCN32(RLDAE) | 66.01 | 61.98 | 59.70 |
MVGCN32 | 72.76 | 66.00 | 59.00 |
MVRGCN32(RLDAE) | 70.10 | 68.02 | 64.50 |
MVRGCN32(DDAE). | 70.00 | NA | NA |
To investigate robustness of the proposed MVRGCN model against error in data, we inject masking noise with various levels to Fox dataset denoted as EFox0.2 and EFox0.4, respectively. Table 5 also presents the results on erroneous Fox. The proposed MVRGCN model with robust low-rank deep autoencoders outperforms the baselines and is thus robust against that error type compared to the other approaches. The proposed MVRGCN model with robust low-rank deep autoendoders is superior to the proposed RCGN model with robust low-rank deep autoencoders whose input data is concatenation of all views. This is mainly because the proposed MVRGCN model exploits complementary and consistent information from multiple views elegantly.
4. Related Work
Improving robustness of machine learning algorithms has received considerable attention recently (Najafi et al., 2017, 2019; Li et al., 2016; Pu et al., 2019). Existing methods for improving robustness of the deep learning models can be roughly classified into four categories. The first class is to modify the typical loss function so as to make it error-tolerant (Li et al., 2016; Zhou and Paffenroth, 2017; Pu et al., 2019; Vincent et al., 2010, 2010; Xie et al., 2012). The proposed RGCN and MVRGCN models belong to this category. There are two broad research directions under this category: (1) training with erroneous input data, (2) training with augmented clean input data. Under the first direction, Zhou and Paffenroth proposed robust low-rank deep autoencoders based on robust principle component analysis for anomaly detection (Zhou and Paffenroth, 2017). The idea is to incorporate low-rank decomposition of anomalous data into low-rank clean component and sparse error component in deep autoencoders. The main focus of that work is on anomaly detection. Different from (Zhou and Paffenroth, 2017), we feed low-rank clean component of robust deep autoencoders as input data to GCN for robustness improvement.
As another work under training with erroneous input data, Pu et al. devised robust autoencoders based on robust principle component analysis via disentanglement of two autoencoders to obtain low-rank and sparse components (Pu et al., 2019). The evaluation was done for image denoising and music/voice/video separation. Different from (Pu et al., 2019), we learn both low-rank clean and sparse error components with one robust deep autoencoders for robust supervised and semi-supervised classification. The similarity between our proposed models and (Zhou and Paffenroth, 2017; Pu et al., 2019) is to incorporate the idea of low-rank decomposition in robust principle component analysis in deep autoencoders to learn clean low-rank component.
For training with augmented clean input by modifying loss function, the prominent approach is to use denoising autoencoders (Vincent et al., 2010; Xie et al., 2012; Goodfellow et al., 2016). Likewise, we used denoising autoencoders in order to robustify GCN, but as we stated earlier, this type of autoencoder requires clean training input data. Therefore, given possibly erroneous input data, filtering is indeed necessary.
In the second category, extra layer is plugged into the deep learning model (Han et al., 2018; Wang et al., 2019; Goldberger and Ben-Reuven, 2017). The proposed RGCN and MVRGCN models belong to this category. One natural existing method here is to add noise layer as first hidden layer to the deep learning model, but it requires clean training data points and labels. On the other hand, there are existing approaches this category that are based on the assumption that input data points are clean, but labels are corrupted (Goldberger and Ben-Reuven, 2017; Han et al., 2018; Wang et al., 2019). For example, Wang et al. proposed a method for sentence-level sentiment classification given clean sentences but noisy labels (Wang et al., 2019). The key idea is to add noise model layer that learns noise transition matrix from the given noisy labels. Different from (Wang et al., 2019), our proposed models can handle erroneous documents (or sentences) in data, but assumes clean labels. Also, our proposed models can be successfully applied to both text and image datasets.
In the third category, erroneous data is filtered out and the deep learning model is trained with the remaining (or filtered out) data points that should be clean with high confidence (Jiang et al., 2018). There are therefore two sequential steps in this category for model training: (1) data filtering (2) model training with clean data points. The proposed RGCN and MVRGCN models with denoising deep autoencoders belong to this category. Jiang et el. presented a novel method for training deep CNNs on corrupted labels (Jiang et al., 2018). Different from (Jiang et al., 2018), our proposed models can handle erroneous input data, but clean labels. However, please note that we will investigate the robustness of the proposed models on erroneous data with denoising autoencoders in our future work.
The fourth category of approaches uses adversarial examples (Li et al., 2016; Jin et al., 2015; Zhu et al., 2019). Adversarial examples leads deep learning models to learn superficial data statistics and causes significant risk for the models deployed in safety-critical systems such as user authentication. They are often created by injecting noise into clean data points. Adversarial training improves robustness of the deep learning models against adversarial examples by training the models with them. Different from this category, the proposed RGCN and MVRGCN models do not explicitly use adversarial training.
5. Conclusion and Future Work
In this paper, we developed a novel robust model for single-view and multi-view data via graph convolutional neural networks and robust deep autoencoders. The proposed RGCN and MVRGCN models have several advantages over traditional graph convolutional neural networks and other existing deep learning and non-deep learning models. First, they handle typical types of error with various levels elegantly. Second, an optimization approach via ADAM or combination of ADAM and ADMM shows to converge on several well-known datasets and their erroneous variants well. Compared to the traditional graph convolutional neural networks, some of existing deep learning models such as fully connected networks, and non-deep learning models such as SVM, the proposed RGCN and MVRGCN models achieved better performance on three real-world datasets with typical error models.
For future work, we plan to apply the proposed RGCN and MVRGCN models on large multi-view datasets such as Reuters. For Fox dataset, we will investigate the effectiveness of the proposed models with word embeddings. Evaluating the robustness of the proposed RGCN and MVRGCN models against outlier-contaminated datasets, and conducting experiments on various architectures for robust deep autoendoers when using two separate deep autoencoders to capture clean low-rank and sparse components are considered as other directions for the future work. It is worth mentioning that the results for Fox dataset does not outperform state-of-the-art methods based on Markov chains e.g., (Najafi et al., 2017). My intuition is that those results are achievable with training the models for more epochs and trying other architectures for robust deep autoencoders.
References
- (1)
- Abdel-Hamid et al. (2013) Ossama Abdel-Hamid, Li Deng, and Dong Yu. 2013. Exploring convolutional neural network structures and optimization techniques for speech recognition.. In INTERSPEECH. ISCA, 3366–3370.
- Candès et al. (2011) Emmanuel J. Candès, Xiaodong Li, Yi Ma, and John Wright. 2011. Robust Principal Component Analysis? J. ACM 58, 3 (2011), 11:1–11:37.
- Cao et al. (2018) X. Cao, F. Zhou, L. Xu, D. Meng, Z. Xu, and J. Paisley. 2018. Hyperspectral Image Classification With Markov Random Fields and a Convolutional Neural Network. IEEE Transactions on Image Processing 27, 5 (2018), 2354–2367.
- Chung (1997) F. R. K. Chung. 1997. Spectral Graph Theory.
- Defferrard et al. (2016) Michaël Defferrard, Xavier Bresson, and Pierre Vandergheynst. 2016. Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering. In Proceedings of the 30th International Conference on Neural Information Processing Systems (NIPS’16). 3844–3852.
- Goldberger and Ben-Reuven (2017) Jacob Goldberger and Ehud Ben-Reuven. 2017. Training deep neural-networks using a noise adaptation layer. In 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings.
- Goodfellow et al. (2016) Ian Goodfellow, Yoshua Bengio, and Aaron Courville. 2016. Deep Learning. The MIT Press.
- Han et al. (2018) Bo Han, Jiangchao Yao, Gang Niu, Mingyuan Zhou, Ivor W. Tsang, Ya Zhang, and Masashi Sugiyama. 2018. Masking: A New Perspective of Noisy Supervision. In Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018, 3-8 December 2018, Montréal, Canada. 5841–5851.
- Jiang et al. (2018) Lu Jiang, Zhengyuan Zhou, Thomas Leung, Li-Jia Li, and Li Fei-Fei. 2018. MentorNet: Learning Data-Driven Curriculum for Very Deep Neural Networks on Corrupted Labels. In Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholmsmässan, Stockholm, Sweden, July 10-15, 2018. 2309–2318.
- Jin et al. (2015) Jonghoon Jin, Aysegul Dundar, and Eugenio Culurciello. 2015. Robust Convolutional Neural Networks under Adversarial Noise. ArXiv abs/1511.06306 (2015).
- Kang et al. (2016) Kai Kang, Wanli Ouyang, Hongsheng Li, and Xiaogang Wang. 2016. Object Detection From Video Tubelets With Convolutional Neural Networks. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
- Kingma and Ba (2014) Diederik P. Kingma and Jimmy Ba. 2014. Adam: A Method for Stochastic Optimization. cite arxiv:1412.6980Comment: Published as a conference paper at the 3rd International Conference for Learning Representations, San Diego, 2015.
- LeCun et al. (1998) Yann LeCun, Léon Bottou, Yoshua Bengio, and Patrick Haffner. 1998. Gradient-Based Learning Applied to Document Recognition. In Proceedings of the IEEE, Vol. 86. 2278–2324.
- Li et al. (2016) Yitong Li, Trevor Cohn, and Timothy Baldwin. 2016. Learning Robust Representations of Text. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing. 1979–1985.
- Lin et al. (2010) Zhouchen Lin, Minming Chen, and Yi Ma. 2010. The Augmented Lagrange Multiplier Method for Exact Recovery of Corrupted Low-Rank Matrices.
- Liu et al. (2013) Guangcan Liu, Zhouchen Lin, Shuicheng Yan, Ju Sun, Yong Yu, and Yi Ma. 2013. Robust Recovery of Subspace Structures by Low-Rank Representation. IEEE Trans. Pattern Anal. Mach. Intell. 35, 1 (2013), 171–184.
- Najafi et al. (2017) Mehrnaz Najafi, Lifang He, and Philip S. Yu. 2017. Error-robust multi-view clustering. 2017 IEEE International Conference on Big Data (Big Data) (2017), 736–745.
- Najafi et al. (2019) Mehrnaz Najafi, Lifang He, and Philip S. Yu. 2019. Outlier-Robust Multi-Aspect Streaming Tensor Completion and Factorization. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI 2019. 3187–3194.
- Pu et al. (2019) Jie Pu, Yannis Panagakis, and Maja Pantic. 2019. Learning Low Rank and Sparse Models via Robust Autoencoders. ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (2019), 3192–3196.
- Shuman et al. (2013) D. I. Shuman, S. K. Narang, P. Frossard, A. Ortega, and P. Vandergheynst. 2013. The emerging field of signal processing on graphs: Extending high-dimensional data analysis to networks and other irregular domains. IEEE Signal Processing Magazine 30, 3 (2013), 83–98.
- Vincent et al. (2010) Pascal Vincent, Hugo Larochelle, Isabelle Lajoie, Yoshua Bengio, and Pierre-Antoine Manzagol. 2010. Stacked Denoising Autoencoders: Learning Useful Representations in a Deep Network with a Local Denoising Criterion. J. Mach. Learn. Res. 11 (2010), 3371–3408.
- Wang et al. (2019) Hao Wang, Bing Liu, Chaozhuo Li, Yan Yang, and Tianrui Li. 2019. Learning with Noisy Labels for Sentence-level Sentiment Classification. CoRR abs/1909.00124 (2019).
- Xie et al. (2012) Junyuan Xie, Linli Xu, and Enhong Chen. 2012. Image Denoising and Inpainting with Deep Neural Networks. In Proceedings of the 25th International Conference on Neural Information Processing Systems - Volume 1 (NIPS’12). 341–349.
- Zhou and Paffenroth (2017) Chong Zhou and Randy C. Paffenroth. 2017. Anomaly Detection with Robust Deep Autoencoders. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD ’17). 665–674.
- Zhu et al. (2019) Dingyuan Zhu, Ziwei Zhang, Peng Cui, and Wenwu Zhu. 2019. Robust Graph Convolutional Networks Against Adversarial Attacks. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (KDD ’19). 1399–1407.