Partially Shared Semi-supervised Deep Matrix Factorization with Multi-view Data
Abstract
Since many real-world data can be described from multiple views, multi-view learning has attracted considerable attention. Various methods have been proposed and successfully applied to multi-view learning, typically based on matrix factorization models. Recently, it is extended to the deep structure to exploit the hierarchical information of multi-view data, but the view-specific features and the label information are seldom considered. To address these concerns, we present a partially shared semi-supervised deep matrix factorization model (PSDMF). By integrating the partially shared deep decomposition structure, graph regularization and the semi-supervised regression model, PSDMF can learn a compact and discriminative representation through eliminating the effects of uncorrelated information. In addition, we develop an efficient iterative updating algorithm for PSDMF. Extensive experiments on five benchmark datasets demonstrate that PSDMF can achieve better performance than the state-of-the-art multi-view learning approaches. The MATLAB source code is available at https://github.com/libertyhhn/PartiallySharedDMF.
Index Terms:
Multi-view learning; Deep matrix factorization; Semi-supervised learning; Partially shared structure.I Introduction
In practical applications, real-world data can be described from different views, that is the so called multi-view data. For instance, an image can be described by several characteristics, e.g., shape, color, texture and so on. Because multi-view representation learning can exploit the implicit-dependent structure of multiple views and improve the performance of the learning tasks. In recent years, multi-view representation learning has attracted increasing research attention in machine learning [1, 2, 3].
Over the past decade, lots of researches on multi-view learning have emerged. In particular, Non-negative Matrix Factorization (NMF) [4, 5] as one of the most popular high-dimensional data processing algorithms, have been widely used for clustering multi-view data [7, 6]. NMF-based multi-view clustering methods have shown to generate superior clustering results which are easy to interpret [8, 9]. Considering the relations between view-specific (uncorrelated) and common information (correlated), partially shared NMF-based multi-view learning approaches are proposed [10, 11].
Nevertheless, most existing methods are single layer structures, which are hard to extract hierarchical structural information [12] of multi-view data. With the development of deep learning, Wang et al. proposed DCCA to extract the hidden hierarchical information in 2-view data [13]. Zhao et al. [14] extended single-view deep matrix factorization [15] into multi-view and proposed graph regularized deep multi-view clustering, which can eliminate the interference in multi-view data. Huang et al. [16] designed a novel robust deep multi-view clustering model to learn the hierarchical semantics without hyperparameters. For clarity, the general deep matrix factorization framework for multi-view data as illustrated in Fig. 1. However, most existing deep matrix factorization multi-view methods are merely considering the common information and ignoring the effect of view-specific information in each individual view. Besides, they are formulated as an unsupervised learning problem and inapplicable when partially labeled data is available. Actually, researchers have found that integrating such label information can produce a considerable improvement in learning performance [19, 18, 17].
To this end, we propose a novel deep multi-view clustering methods called Partially Shared Semi-supervised Deep Matrix Factorization (PSDMF). In our method, correlated and uncorrelated features of multi-view data are both considered by partially shared approaches, in which the latent representation of each view is divided into the common part and the view-specific part. A robust sparse regression term with norm is adopted to integrate the partial label information of the labeled data. Besides, to respect the intrinsic geometric relationship and avoid the parameters problem in different views, we apply graph regularization and auto-weighted strategy to PSDMF. An efficient iterative updating algorithm with pre-training scheme is designed for PSDMF. We summarize our major contributions as follows:
-
•
This paper proposes an semi-supervised deep multi-view clustering model to improve the performance of traditional unsupervised deep matrix factorization methods by using label regression learning approach.
-
•
To respect the common information and view-specific features of multi-view data, we propose a partially shared deep matrix factorization method to jointly exploit two kinds of information and learn a comprehensive final layer representation of multi-view data.
-
•
Local invariant graph regularization and the auto-weighted strategy are introduced to preserve the intrinsic geometric structure in each view and further boost the quality of output representation.
The rest of this paper is organized as follows. In Section II, we give a brief review of related works and describe the proposed PSDMF. Section III present an efficient algorithm to solve the optimization problem. In Section IV, we report the experimental results on five real-world datasets. Finally, Section V concludes this paper. Table. I summarizes the general notations in this article for reader’s convenience.
Notation | Description |
---|---|
The number of views | |
The number of layers | |
The layer sizes | |
The number of samples | |
The number of labeled samples | |
The number of unlabeled samples | |
The dimension of view-specific encoding matrix | |
The dimension of shared encoding matrix | |
The data matrix of the -th view | |
The hidden matrix in the -th layer of -th view | |
The partially shared factor latent representation matrix | |
The -th layer partially shared factor of the -th view | |
The regression weight matrix |
II Methodology
II-A Overview of Deep Matrix Factorization
Semi-NMF is not only a useful data dimensionality reduction technique but also beneficial to data clustering [20]. Motivated by deep neural network structures, Trigeorgis et al. [15] proposed a multi-layer structure Semi-NMF called Deep Semi-NMF to exploit data complex hierarchical information with implicit lower-level hidden attributes. The Deep Semi-NMF decomposes dataset hierarchically, and the process can be formulated as:
(1) | ||||
where is the numbers of layers, is the -th layer hidden matrix, denotes the -th layer low-dimensional representation matrix and we use the notation to state that a matrix contains only non-negative elements. The loss function of Deep Semi-NMF is:
(2) |
The Semi-NMF and Deep Semi-NMF mentioned above can be regarded as single view algorithms. For multi-view data, Zhao et al. [14] presented a new Deep Semi-NMF framework with graph regularization (DMVC) which can eliminate the negative interference in the multi-source data and obtain an effective consensus representation in the final layer. Given data consisting of views, denoted by , the loss function of DMVC is:
(3) | ||||
s.t. |
where denotes the weighting coefficient of the -th view and is a important hyperparameter to control the weights distribution. is the graph Laplacian of the graph for view and is used to adjust the contribution of the graph constrains. The details of how to construct a graph matrix will be discussed in the next subsection. However, as an unsupervised method, DMVC cannot make use of partial prior knowledge of data (e.g. labels). Besides, DMVC only concerns common representation of multi-view data and ignores the view-specific features. In this paper, we propose a novel semi-supervised deep matrix factorization model to address such challenging problems.
II-B Partially Shared Semi-supervised Deep Matrix Factorization
To make full use of priori knowledge and learn low-dimensional factors with powerful discrimination, motivated by recently proposed label regression learning technique [10, 17], we combine the following -norm regularized regression in our model:
(4) |
where denotes the labeled part of representation matrix and is the unlabeled part as shown in Fig. 2 (i.e., ). is the regression coefficient matrix and the -norm is used to enforce it sparse in rows. Consider we have a priori knowledge that those samples share the same label, a binary weight matrix is constructed by following rule:
(5) |
The effectiveness of graph regularization technique has been shown in recent research work [28, 14] and it is able to keep the geometric structure of data when the dimension is changed. Similar to the DMVC, we construct a local graph Laplacian matrix to preserve the local geometrical structure of the each view . As introduced in [14], the graph binary weight matrix is constructed in k-nearest neighbor (k-NN) fashion. Formally, the regularization term is calculated:
(6) | ||||
where . denotes the graph Laplacian matrix for each view data.
Different from the existing deep matrix factorization multi-view methods, we use partially shared strategy to jointly exploit view-specific and common features. The partially shared latent representation matrix and . The common factor ratio . The final -th layer partially shared factor is divided into four parts: the labeled and unlabeled view-specific encoding matrix (i.e. and ), the labeled and unlabeled shared encoding matrix (i.e. and ). We propose a general partially shared deep matrix factorization framework as follows:
(7) | |||
where and denote the hidden view-specific matrix and common matrix, respectively. To simplify the problem, we use , , by combining Eq.(4), Eq.(6) and Eq.(7), we get the cost function of PSDMF as:
(8) | ||||
where is the weighting coefficient for the -th view and it has a great influence on the effect of the model. In DMVC [14], the weights distribution smooth is determined by parameter , but needs to be searched in a large range manually, which makes it to adjust. To avoid this problem, inspired by [25, 26], we use the auto-weighted strategy that obtains the value of based on the distance between the data and the decomposition matrices, as follows:
(9) |
III Optimization
To expedite the approximation of the matrices in our proposed model, we conduct the pre-training [24] by decomposing each view by minimizing the Semi-NMF , where and . Then the factor matrix is further decomposed as by minimizing the , where , and denote the dimensions for layer 1 and layer 2. The process will be repeated till all of the layers are pre-trained. The optimization of Semi-NMF can be derived following a similar process as described in [20]. To save the space, we omit the updating rules here.
Update rule for hidden matrix : By fixing and , we minimize the objective function (8) with respect to . Letting the partial derivative , we can obtain
(10) |
where and .
Update rule for regression weight matrix : Following [27], the derivative of objective function (8) with respect to is as follows:
(11) |
where is a diagonal matrix with . According to the optimization theory, set , then we can obtain the update rule for is
(12) |
Update rule for partially shared factor matrix : As illustrated in Fig. 2, can be divided into four parts: and . Similarly, each data matrix is divided into two parts and ; each Laplacian Graph matrix is divided into two parts and .
For the constraint , we introduce the Lagrangian multiplier as follows:
(13) | ||||
where . Accordingly, is divided into two parts and . For the convenience of writing, we denote that , .
The zero gradient condition of with respect to and , respectively, we have
(14) | ||||
where . Following a similar proof to [14], using the Karush-Kuhn-Tucker condition, we can formulate the updating rules for :
(15) | ||||
where and denote a matrix that all the negative elements are replaced by and all the positive elements are replaced by , respectively. That is,
(16) |
Datasets | DMVC | GMC | lLSMC | LMVSC | DICS | MvSL | PSLF | GPSNMF | Ours | |
---|---|---|---|---|---|---|---|---|---|---|
Extended | ACC | 50.340.07 | 43.380.00 | 53.411.29 | 36.620.00 | 47.543.58 | 19.641.81 | 43.794.90 | 69.568.04 | 87.383.45 |
Yale B | NMI | 49.970.14 | 44.900.00 | 53.400.64 | 28.090.00 | 50.055.11 | 11.822.88 | 32.144.85 | 60.449.12 | 83.532.44 |
Purity | 50.490.07 | 43.690.00 | 53.451.26 | 42.770.00 | 51.393.82 | 20.521.66 | 44.104.36 | 69.568.04 | 87.383.45 | |
Prokaryotic | ACC | 53.502.23 | 49.550.00 | 52.316.01 | 57.530.00 | 56.0711.76 | 35.682.51 | 46.496.63 | 63.452.59 | 66.832.94 |
NMI | 3.280.66 | 19.340.00 | 16.498.59 | 13.370.00 | 12.639.88 | 1.050.41 | 1.691.02 | 18.465.48 | 21.084.21 | |
Purity | 57.310.65 | 58.440.00 | 63.084.15 | 62.940.00 | 60.145.16 | 57.030.07 | 57.460.73 | 63.452.59 | 66.532.94 | |
Caltech101 | ACC | 54.600.43 | 69.200.00 | 51.982.61 | 69.470.00 | 50.9510.61 | 46.424.14 | 76.883.09 | 89.772.54 | 90.243.08 |
-7 | NMI | 37.605.79 | 65.950.00 | 53.222.22 | 46.680.00 | 53.087.71 | 51.592.81 | 40.947.75 | 74.534.10 | 75.525.36 |
Purity | 79.554.64 | 88.470.00 | 85.611.64 | 77.070.00 | 85.283.70 | 83.731.49 | 77.564.12 | 90.382.33 | 91.122.03 | |
Caltech101 | ACC | 51.721.72 | 45.640.00 | 45.034.16 | 42.080.00 | 35.753.83 | 42.012.86 | 55.7510.35 | 77.333.34 | 79.743.73 |
-20 | NMI | 52.181.49 | 48.090.00 | 61.211.14 | 49.640.00 | 55.992.58 | 60.481.45 | 44.855.77 | 67.483.52 | 71.243.06 |
Purity | 70.230.92 | 55.490.00 | 76.950.79 | 50.840.00 | 69.322.47 | 76.161.17 | 62.554.94 | 77.812.60 | 79.742.84 | |
MSRCV1 | ACC | 40.762.37 | 74.760.00 | 69.911.79 | 64.760.00 | 51.916.61 | 78.075.83 | 82.964.51 | 72.174.03 | 86.245.69 |
NMI | 27.603.17 | 70.090.00 | 62.372.71 | 58.810.00 | 50.4811.29 | 68.324.69 | 75.043.49 | 67.414.34 | 76.143.42 | |
Purity | 45.142.41 | 79.050.00 | 72.861.38 | 69.050.00 | 58.098.68 | 78.525.73 | 82.964.51 | 74.924.44 | 86.244.45 |
Until now, we have all the update rules done. Updates iterations repeatedly until convergence. The overall optimization process of PSDMF is outlined in Algorithm 1, where the ”Semi-NMF” procedure performs the pre-train as described earlier. Once the optimal partially shared latent representation matrix and regression weight matrix are obtained, cluster label where .
IV Experiments
To comparatively study the performance of PSDMF, we consider several state-of-the-art baselines as stated bellow:
Multi-view clustering via deep matrix factorization (DMVC) [14], graph-based multi-view clustering (GMC) [21], latent multi-view subspace clustering (lLSMC) [23], large-scale multi-view subspace clustering (LMVSC) [22], multi-view discriminative learning via joint non-negative matrix factorization (DICS) [18], graph-regularized multi-view semantic subspace learning (MvSL) [19], partially shared latent factor learning (PSLF) [10] and semi-supervised multi-view clustering with graph-regularized partially shared non-negative matrix factorization (GPSNMF) [17]. Among them, four methods are unsupervised (i.e., DMVC, GMC, lLSMC, LMVSC) and the others are semi-supervised.
IV-A Datasets and Evaluation Metric
We perform experiments on five benchmark datasets: Extended Yale B, Prokaryotic, MSRCV1, Caltech101-7 and 20. More detailed information about these data is shown in Table III.
Dataset | Dimension of Views | Sizes | Classes |
---|---|---|---|
Extended Yale B | 2500/3304/6750 | 650 | 10 |
Prokaryotic | 393/3/438 | 551 | 4 |
Caltech101-7/20 | 48/40/254/1984/512/928 | 1474/2386 | 7/20 |
MSRCV1 | 24/576/512/256/254 | 210 | 7 |
For evaluation metrics, we use Accuracy (ACC), Normalized Mutual Information (NMI) and Purity to comprehensively measure the clustering performance in our experiments. The formal definitions of these metrics, omitted here to save space, can be found in [9].
IV-B Experiment Setup
Following [17], for PSLF, GPSNMF and PSDMF, the dimension of partially shared latent representation and the common factor ratio are set to 100 and 0.5, respectively. Thus, and . In addition, they all predicate the labels with regression matrix . Other than these three algorithms, all the above methods are evaluated by the classic clustering algorithm, K-means. In terms of semi-supervised methods (i.e. DICS, MvSL, PSLF, GPSNMF and PSDMF), we denote as the proportions of the labeled data. Following [14], the layer sizes of deep structure models (i.e. DMVC, PSDMF) are set as .
For all compared methods, 10 repeated times are run to reduce the randomness caused by initialization, and report their average scores and standard deviations. As to the compared methods, source codes are obtained from the authors’ websites, and we set their parameters to the optimal value if they have.
IV-C Performance Evaluation
The performances of multi-view clustering is shown in Table II. From the experiment results, we have drawn some conclusions as following: In general, semi-supervised algorithms are superior to unsupervised algorithms. For example, our PSDMF significantly improves the performance compared with DMVC all over the five datasets, even when only label information is used. For the dataset Extended Yale B, we raise the performance bar by around in ACC, in NMI and in Purity. On average, we improve the state-of-the-art GPSNMF by more than , which means multi-layer structure methods have clustering advantages over single-layer methods. Through the partially shared deep representation, PSDMF eliminates the influence of undesirable factors by considering both view-specific and common features, and keeps the important information in the final representation layer. In different datasets experiments, our model achieves better performances than the state-of-the-art methods, which demonstrate the robustness of ours.
IV-D Parameters Sensitivity
There are three nonnegative essential parameters (i.e. ) in our proposed model. controls the smoothness of the partially shared factor. and controls the sparsity of the weight matrix and balances the relation between the regression term and data reconstruction, respectively. Following [10, 17], usually obtains better performance. As is shown in Fig. 3, we test the sensitivity of the parameters and in terms of clustering results, which are tuned in the range of and , respectively. We can observe that the parameters affect the experimental results in different datasets, which indicates that and play an important role in PSDMF. In practice, we fix and as default in our experiments.
V conclusion
In this paper, we introduced a novel semi-supervised deep matrix factorization model for multi-view learning, called PSDMF, which is able to learn a comprehensive partially shared latent final-layer representation for multi-view data. Through the partially shared multi-layer matrix factorization structure, our method is capable of exploiting view-specific and common features among different views, simultaneously. Benefitting from the label regression term, it can incorporate information from the labeled data. Furthermore, we utilize graph regularization and auto-weighted strategy to preserve the intrinsic geometric structure of the data. An iterative optimization algorithm is developed to deal with the PSDMF. Experimental results show that the proposed model achieves superior performances.
Acknowledgment
This work was supported by the National Natural Science Foundation of China under Grant Nos. 61722304, 61801133, and 61803096, in part by the Guangdong Science and Technology Foundation under Grant Nos. 2019B010118001, 2019B010121001, and 2019B010154002, National Key Research and Development Project, China under Grant No. 2018YFB1802400.
References
- [1] S. Sun, “A survey of multi-view machine learning,” Neural computing and applications, vol. 23, no. 7–8, pp. 2031–2038, 2013
- [2] Y. Li, M. Yang, Z. Zhang, “A survey of multi-view representation learning,” IEEE Trans. Knowl. Data Eng., vol. 31, no. 10, pp. 1863–1883, 2018.
- [3] R. Zhang, F. Nie, X. Li and X. Wei, “Feature selection with multi-view data: A survey,” Information Fusion, vol. 50, pp. 158–167, 2019.
- [4] D. D. Lee, H. S. Seung, “Algorithms for non-negative matrix factorization,” in NIPS, pp. 556–562, 2001.
- [5] Z. Yang, Y. Xiang, K. Xie and Y. Lai, “Adaptive method for nonsmooth nonnegative matrix factorization,” IEEE Trans. Neural Netw. Learn. Syst., vol. 28, no. 4, pp. 948–960, 2017.
- [6] X. Zhang, L. Zong, X. Liu and H. Yu, “Constrained NMF-based multi-view clustering on unmapped data,” in AAAI, pp. 3174–3180, 2015.
- [7] J. Liu, C. Wang, J. Gao and J. Han, “Multi-view clustering via joint nonnegative matrix factorization,” in SDM, pp. 252–260, 2013.
- [8] L. Zong, X. Zhang, L. Zhao, H. Yu and Q. Zhao, “Multi-view clustering via multi-manifold regularized non-negative matrix factorization,” Neural Netw., vol. 88, pp. 74–89, 2017.
- [9] Z. Yang, N. Liang, W. Yan, Z. Li and S. Xie, “Uniform distribution non-negative matrix factorization for multiview clustering,” IEEE Trans. Cybern., in press, 2020.
- [10] J. Liu, Y. Jiang, Z. Li, Z. H. Zhou and H. Lu, “Partially shared latent factor learning with multiview data,” IEEE Trans. Neural Netw. Learn. Syst., vol. 26, no. 6, pp. 1233–1246, 2015.
- [11] L. Zhao, T. Yang, J. Zhang, Z. Chen, Y. Yang and Z. J. Wang, ”Co-learning non-negative correlated and uncorrelated features for multi-view data,” IEEE Trans. Neural Netw. Learn. Syst., in press, 2020.
- [12] Y. Meng, R. Shang, F. Shang, L. Jiao, S. Yang and R. Stolkin, “Semi-supervised graph regularized deep NMF with bi-orthogonal constraints for data representation,” IEEE Trans. Neural Netw. Learn. Syst., in press, 2019.
- [13] W. Wang, R. Arora, K. Livescu and J. Bilmes, “On deep multi-view representation learning,” in ICML, pp. 1083–1092, 2015.
- [14] H. Zhao, Z. Ding and Y. Fu, “Multi-view clustering via deep matrix factorization,” in AAAI, pp. 2921–2927, 2017.
- [15] G. Trigeorgis, K. Bousmalis, S. Zafeiriou, and B. W. Schuller, “A deep matrix factorization method for learning attribute representations,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 39, no. 3, pp. 417–429, 2017.
- [16] S. Huang, Z. Kang and Z. Xu, “Auto-weighted multi-view clustering via deep matrix decomposition,” Pattern Recognition, vol. 97, no. 107015, 2020.
- [17] N. Liang, Z. Yang, Z. Li, S. Xie and C. Su, “Semi-supervised multi-view clustering with Graph-regularized Partially Shared Non-negative Matrix Factorization,” Knowledge-Based Systems, vol. 190, no. 105185, 2020.
- [18] Z. Zhang, Z. Qin, P. Li, Q. Yang and J. Shao, “Multi-view discriminative learning via joint non-negative matrix factorization,” in Int. Conf. Database Systems for Advanced Applications, pp. 542–557, 2018.
- [19] J. Peng, P. Luo, Z. Guan and J. Fan, ”Graph-regularized multi-view semantic subspace learning,” Int. Jour. Mach. Learn. Cybern., vol. 10, no. 5, pp. 879–895, 2019.
- [20] C. H. Ding, T. Li, and M. I. Jordan, “Convex and semi-nonnegative matrix factorizations,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 32, no. 1, pp. 45–55, 2010.
- [21] H. Wang, Y. Yang and B. Liu, “GMC: Graph-based multi-view clustering,” IEEE Trans. Knowl. Data Eng., vol. 32, no. 6, pp. 1116–1129, 2019.
- [22] Z. Kang, W. Zhou, Z. Zhao, J. Shao, M. Han and Z. Xu, “Large-scale multi-view subspace clustering in linear time,” in AAAI, 2020.
- [23] C. Zhang, H. Fu, Q. Hu, X. Cao, Y. Xie, D. Tao and D. Xu, “Generalized latent multi-view subspace clustering,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 42, no. 1, pp. 86–99, 2020.
- [24] W. Zhao, C. Xu, Z. Guan and Y. Liu, “Multiview concept learning via deep matrix factorization,” IEEE Trans. Neural Netw. Learn. Syst., in press, 2020.
- [25] F. Nie, G. Cai, J. Li and X. Li, “Auto-weighted multi-view learning for image clustering and semi-supervised classification,” IEEE Trans. Image Process., vol. 27, no. 3, pp. 1501–1511, 2018.
- [26] Z. Kang, X. Lu, J. Yi and Z. Xu, “Self-weighted multiple kernel learning for graph-based clustering and semi-supervised classification,” in IJCAI, pp. 2312–2318, 2018.
- [27] F. Nie, X. Cai, H. Huang and C. Ding, “Efficient and robust feature selection via joint l2, 1-norms minimization,” in NIPS, pp. 1813–1821, 2010.
- [28] D. Cai, X. He, J. Han and T.S. Huang, “Graph regularized nonnegative matrix factorization for data representation,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 33 no. 8, pp. 1548–1560, 2011.