Generalized Clustering and Multi-Manifold Learning with
Geometric Structure Preservation
Abstract
Though manifold-based clustering has become a popular research topic, we observe that one important factor has been omitted by these works, namely that the defined clustering loss may corrupt the local and global structure of the latent space. In this paper, we propose a novel Generalized Clustering and Multi-manifold Learning (GCML) framework with geometric structure preservation for generalized data, i.e., not limited to 2-D image data and has a wide range of applications in speech, text, and biology domains. In the proposed framework, manifold clustering is done in the latent space guided by a clustering loss. To overcome the problem that the clustering-oriented loss may deteriorate the geometric structure of the latent space, an isometric loss is proposed for preserving intra-manifold structure locally and a ranking loss for inter-manifold structure globally. Extensive experimental results have shown that GCML exhibits superior performance to counterparts in terms of qualitative visualizations and quantitative metrics, which demonstrates the effectiveness of preserving geometric structure. Code has been made available at: https://github.com/LirongWu/GCML.
1 Introduction
Clustering, a fundamental tool for data analysis and visualization, has been an essential research topic in data science. This paper focuses on generalized clustering, which takes vector data as input and is applicable to data with various dimensions, not limited to 2-D image data. Conventional clustering algorithms such as -Means [11], Gaussian Mixture Models [1], and Spectral Clustering [19] perform clustering based on distance or similarity measures. However, handcrafted distance or similarity measures are rarely reliable for large-scale high-dimensional data, making it increasingly challenging to achieve effective clustering. An intuitive solution is to transform the data from the high-dimensional input space to the low-dimensional latent space and then to cluster the data in the latent space. This can be achieved by applying manifold-based dimensionality reduction techniques, such as t-SNE [10], and UMAP [13]. However, since these methods are not specifically designed for clustering tasks, some of their properties may be contrary to our expectations, e.g., two data points from different manifolds that are close in the input space will be closer in the latent space learned by UMAP. Therefore, the first question here is how to perform multi-manifold learning for dimensionality reduction that favors clustering?
The two main points for the multi-manifold learning are Point (1) preserving the local geometric structure within each manifold and Point (2) ensuring the discriminability between different manifolds. Most previous works seem to start with the assumption that data labels are known, and then design the algorithm in a supervised manner. However, it is challenging to decouple complex crossover relations and ensure discriminability between different manifolds, especially in unsupervised settings. One natural strategy is to achieve Point (2) through clustering to get pseudo-labels and then performing single-manifold learning for each manifold. However, the clustering-oriented loss may deteriorate the geometric structure of the latent space111This claim was first made by IDEC [4], but they did not provide any experiment or analysis to support it. In this paper, however, we show that the geometric structure of the latent space is indeed corrupted by extensive visualizations (Fig. 3 and Fig. 4) and statistical analysis (Fig. 6)., and hence clustering is somewhat contrary to Point (1) (this will be detailed in Sec. 3.3). Therefore, it is important to alleviate this contradiction so that clustering helps both Point (1) and Point (2). Thus, the second question here is how to cluster data that favors multi-manifold learning?
In this paper, we propose to jointly perform generalized clustering and multi-manifold learning with geometric structure preservation. Inspired by [24], the clustering centers are defined as a set of learnable parameters, and we use a clustering loss to simultaneously guide the separation of data points from different manifolds and the learning of the clustering centers. To prevent clustering loss from deteriorating the latent space, an isometric loss and a ranking loss are proposed to preserve the intra-manifold local structure and inter-manifold global structure. Finally, we achieve the following three goals related to clustering, geometric structure preservation, and multi-manifold learning: (1) Clustering helps to ensure inter-manifold discriminability (Point 2); (2) Local structure preservation (Point 1) can be compatible with clustering; (3) Geometric structure preservation helps to cluster. The contributions are summarized as:
-
•
Proposing to combine generalized deep clustering and multi-manifold learning into a unified framework with local and global structure preservation.
-
•
Setting the clustering centers as a set of learnable parameters and achieve global structure preservation in a faster, more efficient, and easier to optimize manner by applying ranking loss to the clustering centers.
-
•
Pointing out contradictions between two optimization goals of clustering and local structure preservation and proposing an elegant training strategy to alleviate it.

2 Related Work
Clustering Analysis. As a fundamental tool in machine learning, clustering has been widely applied in various domains. One branch of classical clustering is -Means and Gaussian Mixture Models (GMM), which are fast, easy to understand, and can be applied to a large number of problems. However, limited by Euclidean measures, their performance on high-dimensional data is often unsatisfactory. Spectral clustering and its variants (such as SC-Ncut [1]) extend clustering to high-dimensional data by allowing more flexible similarity measures. However, limited by the computational efficiency of the full Laplace matrix, spectral clustering is challenging to extend to large-scale datasets.
Deep Clustering. The success of deep learning has contributed to the growth of deep clustering [2, 9]. One branch of deep learning performs clustering after learning low-dimensional embeddings. For example, [21] uses autoencoder to learn low-dimensional features and then runs -Means to get clustering results (AE+-Means). Instead, N2D [12] applies UMAP to find the best clusterable manifold of the learned embeddings, and then runs -Means to discover higher-quality clusters. The other category of algorithms tries to optimize clustering and dimensionality reduction jointly. The closest work to us is DEC [24], which learns a mapping from the input space to a low-dimensional latent space through iteratively optimizing clustering-oriented objective. As a modified version of DEC, while IDEC claims to preserve the local structure of data, their contribution is nothing more than adding a reconstruction loss. Besides, JULE [26] unifies representation learning with clustering, which can be considered as a neural extension of hierarchical clustering. Instead, SpectralNet [18] directly embeds the input into the Laplacian eigenspace in which clustering is performed. DSC [27] devises a dual autoencoder to embed data into latent space, and then deep spectral clustering is applied to obtain label assignments. Moreover, DDC [16] proposes to use a density-based clustering algorithm to initialize cluster centers and then perform image cluster discovery.
To avoid any possible misunderstanding, we would like to highlight that generalized deep clustering and visual self-supervised learning (SSL) are two different research fields. SSL typically uses more powerful CNN architecture (applicable only to 2-D image data) and applies sophisticated techniques such as contrastive learning [6], data augmentation, and clustering [28, 22] for better performance on large-scale datasets, such as ImageNet. For example, ASPC-DA [5] combines data augmentation with self-paced learning to encourage the learned embeddings to be cluster-oriented. Besides, ClusterGAN [14] trains a generative adversarial network jointly with a clustering-specific loss to achieve clustering in the latent space. The generalized deep clustering, on the other hand, uses a generalized MLP architecture (applicable to all kinds of data with various dimensions, not limited to 2-D image data) and has a very wide range of applications in images, text, and biology domains.
Manifold Learning. The manifold assumption hypothesizes that a low-dimensional manifold is embedded in a high-dimensional space, and the manifold learning aims to achieve dimensionality reduction via a nonlinear mapping that preserves the geometric structure. Isomap, as a classical algorithm of single-manifold learning, aims to seek a global optimal subspace that best preserves the geodesic distance between data points [20]. In contrast, some algorithms, such as the Locally Linear Embedding (LLE) [17], are more concerned with the preservation of local neighborhood information. Furthermore, multi-manifold learning has been proposed to obtain intrinsic properties of different manifolds. [25] proposes a supervised discriminant isomap where data points are partitioned into different manifolds according to label information. Similarly, [29] proposes a semi-supervised learning framework that applies the labeled and unlabeled training samples to perform the joint learning of local-preserving features. In most previous work, it is assumed that the label is known or partially known, which significantly simplifies the problem. However, it is challenging to decouple multiple overlapping manifolds in unsupervised settings, and that is what this paper aims to explore.
3 Proposed Method
Consider a dataset with samples, and each sample is sampled from different manifolds . Assume that each category in the dataset lies in a compact low-dimensional manifold, and the number of manifolds is prior knowledge. Define two nonlinear mapping and , where is the embedding of in the latent space, and is the reconstruction of . The -th cluster center is denoted as , where is defined as a set of learnable parameters. We aim to find optimal parameters and so that the embeddings can achieve clustering with local and global structure preservation. To this end, a denoising autoencoder shown in Fig. 1 is first pre-trained in an unsupervised manner to learn an initial latent space. Denoising autoencoder aims to optimize the self-reconstruction loss , where the is a copy of with Gaussian noise added, that is, . Then the autoencoder is fine-tuned by optimizing the following clustering-oriented loss and structure-oriented losses . Since the clustering should be performed on features of clean data, instead of noised data that is used in denoising autoencoder, the clean data is used for fine-tuning.
3.1 Clustering-oriented Loss
First, the cluster centers in the latent space are initialized (the initialization method will be introduced in Sec. 4.1). Then the similarity between the embedded point and cluster centers is measured by Student’s -distribution, as follows
(1) |
The auxiliary target distribution is further designed to help manipulate the latent space, defined as:
(2) |
where is the normalized cluster frequency, used to balance the size of different clusters. Then the encoder is optimized by the following objective:
(3) |
The gradient of with respect to each learnable cluster center can be computed as:
(4) |
where facilitates the aggregation of data points within the same manifold, while data points from different manifolds are kept away from each other. However, we found that the clustering-oriented losses may deteriorate the geometric structure of the latent space. To prevent this deterioration, we introduces three other structure-oriented losses to preserve the local and global manifold structures.
3.2 Structure-oriented Loss
Intra-manifold Isometry Loss. The intra-manifold local structure is preserved by optimizing the objective as:
(5) |
where represents the neighborhood of data point in the latent space , and the NN is applied to determine the neighborhood. is an indicator function, and is a manifold determination function that returns the manifold where sample is located, that is, . Then we can derive manifolds by . The loss constrains the isometry within each manifold.
Inter-manifold Ranking Loss. The inter-manifold global structure is preserved by optimizing the objective as:
(6) |
where is defined as the ground-truth centers of different manifolds in the input space with . The parameter determines the extent to which different manifolds move away from each other. The larger is, the further away the different manifolds are from each other. The derivation for the gradient of with respect to each learnable cluster center is placed in Appendix A.2. Note that is optimized in an iterative manner, rather than by initializing once and then separating different clusters based only on initialization results. Additionally, contrary to us, the conventional methods for dealing with inter-manifold separation typically impose push-away constraints on all data points from different manifolds [29, 25], defined as:
(7) |
The main differences between and are as follows: (1) imposes constraints on embedding , which indirectly affects the network parameters . In contrast, imposes constrains directly on parameters in the form of regularization item. (2) is easier to optimize, faster to process, and more accurate. is imposed on all data points from different manifolds, which involves point-to-point relationships. This means that each point may be subject to the push-away force from other manifolds, but at the same time, each point has to meet the isometry constraint with its neighboring points. Under these two constraints, optimization is difficult and it is easy to fall into a local optimal solution. In contrast, is imposed directly on the clustering centers, involving only cluster-to-cluster relationships, which avoids the above problem and makes it easier to optimize. (3) The parameter introduced in allows us to control the extent of separation between manifolds.
Alignment Loss. Global ranking loss is imposed directly on , so optimizing only updates rather the encoder’s parameter . However, the optimization of not only relies on , but is also constrained by , which ensures that data points remain roughly distributed around cluster centers and do not deviate significantly from them during the optimization process. Alignment loss , as an auxiliary term, aims to help align learnable centers with the ground-truth centers and make this binding stronger, defined as
(8) |
where are defined as . The derivation for the gradient of with respect to center is placed in Appendix A.2. As shown in Fig. 1, three structure-oriented losses , form a closed loop between input , embeddings , and cluster centers .
3.3 Training Strategy
Contradiction. The contradiction between clustering and local structure preservation is analyzed from the forces analysis perspective. As shown in Fig. 2 (a), we assume that there exists a data point (red point) and its three nearest neighbors (blue points) around a cluster center (gray point). When clustering and local structure preserving are optimized simultaneously, it’s easy to fall into a local optimum, where the data point is in steady-state, and the resultant force from its three nearest neighbors is equal in magnitude and opposite to the gravitational forces of the cluster.
Alternating Training. To solve the above problem and integrate the goals of clustering and local structure preservation into a unified framework, we take an alternating training strategy. Within each epoch, we first optimize and in a mini-batch, with joint loss defined as
(9) |
Then at each epoch, we jointly optimize isometry loss and on the whole dataset, defined as
(10) |
Weight Continuation. At different stages of training, we have different requirements for clustering and structure preservation. At the beginning of training, to successfully decouple the overlapping manifolds, we hope that the will dominate and will be auxiliary. When the margin between different manifolds is sufficiently large, the weight for can be gradually reduced, while the weight for can be gradually increased, focusing on the preservation of the local isometry. The whole algorithm is summarized in Algorithm 1.



Three-stage Explanation. The training process can be roughly divided into three stages, as shown in Fig. 2 (b), to explain the training strategy more vividly. Also, we provide the learning curves of key losses , , on the MNIST-test dataset in Fig. 2 (c). At first, four different manifolds overlap. At Stage 1, dominates, thus data points within each manifold converge towards cluster centers to form spheres, but the local structure of manifolds is corrupted. At Stage 2, dominates, thus different manifolds in the latent space move away from each other to increase the manifold margin and enhance the discriminability. At stage 3, these manifolds gradually recover their original local structure from the spherical shape with dominating. Note that the above losses may coexist rather than being completely independent at different stages, but the role played by different losses varies due to the alternating training and weight continuation strategies.
Algorithms | MNIST-full | MNIST-test | USPS | Fashion-MNIST | REUTERS-10K | HAR | pendigits | |
---|---|---|---|---|---|---|---|---|
-Means [1] | 0.532/0.500 | 0.546/0.501 | 0.668/0.601 | 0.474/0.512 | 0.599/0.375* | 0.599/0.588 | 0.666/0.681 | |
SC-Ncut [1] | 0.656/0.731 | 0.660/0.704 | 0.649/0.794 | 0.508/0.575 | 0.658/0.401* | 0.538/0.741 | 0.724/0.784 | |
GMM [1] | 0.389/0.333 | 0.464/0.465 | 0.562/0.540 | 0.463/0.514 | 0.613/0.388* | 0.585/0.648 | 0.673/0.682 | |
AE+-Means [21] | 0.818/0.747 | 0.815/0.784* | 0.662/0.693 | 0.566/0.585* | 0.721/0.432* | 0.674/0.670* | 0.713/0.733* | |
DEC [24] | 0.903/0.854* | 0.885/0.851* | 0.889/0.873* | 0.554/0.576* | 0.773/0.528* | 0.759/0.695* | 0.679/0.671* | |
IDEC [4] | 0.918/0.868* | 0.876/0.817* | 0.893/0.876* | 0.572/0.601* | 0.785/0.541* | 0.786/0.718* | 0.739/0.757* | |
VaDE [7] | 0.945/0.876 | 0.287/0.287 | 0.566/0.512 | 0.578/0.630 | 0.795/0.556* | 0.801/0.720* | 0.762/0.743* | |
DEPICT [3] | 0.965/0.917 | 0.963/0.915 | 0.899/0.906 | 0.392/0.392 | - | - | - | |
JULE [26] | 0.964/0.913 | 0.961/0.915 | 0.950/0.913 | 0.563/0.608 | - | - | - | |
DSC [27] | 0.978/0.941 | 0.980/0.946 | 0.869/0.857 | 0.662/0.645 | - | - | - | |
ASPC-DA [5] | 0.988/0.966 | 0.973/0.936 | 0.982/0.951 | 0.591/0.654 | - | - | - | |
ASPC (w/o DA) [5] | 0.931/0.886* | 0.813/0.792* | 0.768/0.803* | 0.576/0.632* | 0.692/0.418* | 0.769/0.682* | 0.769/0.751* | |
N2D [12] | 0.969/0.928* | 0.954/0.897* | 0.954/0.898* | 0.671/0.678* | 0.784/0.536* | 0.807/0.721* | 0.847/0.808* | |
GCML (ours) | 0.980/0.946 | 0.972/0.930 | 0.958/0.902 | 0.710/0.685 | 0.836/0.590 | 0.844/0.762 | 0.855/0.814 |
4 Experiments
4.1 Experimental Setups
In this section, the effectiveness of GCML is evaluated in 7 benchmark datasets: MNIST-full, MNIST-test, USPS, Fashion-MNIST, REUTERS-10K, HAR, and Pendigits on which GCML is compared with 12 other methods in 8 evaluation metrics including metrics specifically designed for clustering and multi-manifold learning. The brief descriptions of the datasets are given in Appendix A.1.
Evaluation Metrics. Two standard evaluation metrics: Accuracy (ACC) and Normalized Mutual Information (NMI) are used to evaluate clustering performance. Besides, six evaluation metrics are adopted in this paper to evaluate the performance of multi-manifold learning, including Relative Rank Error (RRE), Trustworthiness (Trust), Continuity (Cont), Root Mean Reconstruction Error (RMRE), Locally Geometric Distortion (LGD) and Cluster Rank Accuracy (CRA). Limited by space, their precise definitions are available in Appendix A.3.
Parameters Settings. The encoder structure is -500-500-500-2000-10 where is the dimension of the input data, and the decoder is its mirror. After pretraining, to initialize the learnable clustering centers, the t-SNE is applied to find the best clustable manifold in the latent space , and then the -Means algorithm is run to obtain the label assignments for each data point. The centers of each category in the latent space are set as initial cluster centers . Besides, Adam optimizer with learning rate =0.001 is used, the batch size is set to 256, the epoch is set to 300, the parameter for nearest neighbor is set to 5, and the parameter is set to 3 for all datasets. Sensitivity analysis for parameters and is available in Appendix A.4. As described in Sec. 3.3, the weight continuation is applied to train the model. The weight parameter for decreases linearly from 0.1 to 0 within epoch 0-150. In contrast, the weight parameter for increases linearly from 0 to 1.0 within epoch 0-150. In this paper, each set of experiments is run 5 times with different random seeds, and the results are averaged into the final performance metrics.


4.2 Evaluation of Clustering
Quantitative Comparison. The metrics ACC/NMI of different methods are reported in Tab. 1. For those methods whose results are not reported or experimental settings are not clear, we run the released code with the same provided hyperparameters and mark them with (*). Moreover, we mark those methods that are only applicable to 2-D image data as (-) on the vector dataset. While ASPC-DA achieves the best performance on three datasets (MNIST-test, MNIST-full, and USPS), its performance gains do not come directly from clustering, but from sophisticated modules such as data augmentation and self-paced learning. Once these modules are removed, there is large performance degradation. For example, with Data Augmentation (DA) removed, ASPC achieves less competitive performance, e.g., an accuracy of 0.931 (vs 0.988) on MNIST-full, 0.813 (vs 0.973) on MNIST-test and 0.768 (vs 0.982) on USPS. Since ASPC-DA is based on the MLP architecture, its image-based augmentation cannot be applied directly to vector data, which explains why ASPC has no performance advantage (even compared to DEC and IDEC) on the vector datasets, such as REUTERS-10K and HAR datasets.
In a fair comparison (without considering ASPC-DA and marking its results in Tab. 1 as gray color), we find that GCML outperforms -Means, GNN, and SC-Ncut by a significant margin and surpasses the other nine compared DNN-based algorithms on all datasets except MNIST-test. Nevertheless, even with the MNIST-test dataset, GCML still ranks second, outperforming the third by 0.9%. In particular, we obtain the best performance on the Fashion-MNIST, REUTERS-10K, and HAR datasets, and more notably, our clustering accuracy exceeds the current SOTA method by 3.9%, 4.1%, and 3.8%, respectively.
Clustering Visualization. The visualization of GCML with comparison methods is shown in Fig. 3 (visualized using UMAP). Among all methods, only DEC, IDEC, and GCML can hold clear boundaries between different clusters, while the cluster boundaries of the other methods are indistinguishable. Though DEC and IDEC successfully separate different clusters, they group many data points from different classes into the same cluster. Most importantly, due to the use of the clustering-oriented loss, the embedding learned by these algorithms (especially DSC) tend to form spheres and disrupt the original topological structure. Instead, GCML overcomes the above problems and achieves almost perfect separation between different clusters while preserving the local and global structure.


The embeddings of the latent space during the training process are visualized in Fig. 4 for explaining how both clustering and structure-preserving are achieved. We can see that the different clusters initialized by the pre-trained autoencoder are closely adjacent. In the early stage of training, with clustering loss and global ranking loss , different manifolds are separated from each other, but each manifold loses its local structure, and all of them degenerate into spheres. As the training progresses, the weight for gradually decreases, while the weight for increases and the optimization is gradually focused from global to local, with each manifold gradually recovering its original geometric structure from the sphere. These visualizations show that clustering-oriented loss does deteriorate the geometric structure of the latent space, and the designed structure-oriented losses help to recover it.
Generalizability Evaluation. Tab. 2 demonstrates that a learned GCML can generalize well to unseen data with high clustering accuracy. Taking MNIST-full as an example, GCML was trained using 50,000 training samples and then tested on the remaining 20,000 testing samples using the learned model. In terms of the metrics ACC and MNI, GCML is optimal for both training and testing samples. More importantly, there is hardly any degradation in the performance of GCML on the testing samples compared to the training samples, while all other methods show a significant drop in performance, e.g., DEC from 84.1% to 74.8%. This demonstrates the importance of geometric structure preservation for good generalizability. The visualization results of the testing samples are shown in Fig. 5; even for testing samples, GCML still shows distinguishable inter-cluster discriminability, while other methods couple different clusters together, which shows GCML’s great generalizability.
Algorithms | training samples | testing samples |
---|---|---|
AE+-Means | 0.815/0.736 | 0.751/0.711 |
DEC | 0.841/0.773 | 0.748/0.704 |
IDEC | 0.845/0.860 | 0.826/0.842 |
JULE | 0.958/0.907 | 0.921/0.895 |
DSC | 0.975/0.939 | 0.969/0.921 |
N2D | 0.974/0.930 | 0.965/0.911 |
GCML (ours) | 0.978/0.941 | 0.978/0.941 |
Datasets | Algorithms | RRE | Trust | Cont | -RMSE | LGD | CRA | MLP | RFC | SVM | LR |
---|---|---|---|---|---|---|---|---|---|---|---|
MNIST-full | DEC | 0.09988 | 0.84499 | 0.94805 | 44.8535 | 4.37986 | 0.28 | 0.8647 | 0.8706 | 0.8707 | 0.8566 |
IDEC | 0.00984 | 0.99821 | 0.97936 | 24.5803 | 1.71484 | 0.33 | 0.9797 | 0.9737 | 0.9852 | 0.9650 | |
JULE | 0.02657 | 0.93675 | 0.98321 | 28.3412 | 2.12955 | 0.27 | 0.9802 | 0.9825 | 0.9787 | 0.9743 | |
DSC | 0.09785 | 0.87315 | 0.92508 | 6.98098 | 1.19886 | 0.23 | 0.9622 | 0.9501 | 0.9837 | 0.9752 | |
N2D | 0.01002 | 0.99243 | 0.98466 | 5.7162 | 0.69946 | 0.21 | 0.9796 | 0.9803 | 0.9799 | 0.9792 | |
GCML (ours) | 0.00567 | 0.99978 | 0.98716 | 5.4986 | 0.69168 | 1.00 | 0.9851 | 0.9874 | 0.9869 | 0.9841 | |
Fashion-MNIST | DEC | 0.04787 | 0.93896 | 0.95450 | 39.3274 | 3.87731 | 0.37 | 0.6268 | 0.9853 | 0.6377 | 0.6245 |
IDEC | 0.01089 | 0.99683 | 0.97797 | 25.4024 | 1.91385 | 0.27 | 0.8367 | 0.9918 | 0.8607 | 0.7514 | |
JULE | 0.03013 | 0.97732 | 0.97923 | 15.2213 | 1.43642 | 0.43 | 0.8541 | 0.9892 | 0.8566 | 0.7723 | |
DSC | 0.05168 | 0.95013 | 0.96121 | 17.2201 | 1.42091 | 0.36 | 0.8084 | 0.9823 | 0.8618 | 0.7676 | |
N2D | 0.00894 | 0.99062 | 0.98054 | 14.49079 | 1.28180 | 0.26 | 0.8412 | 0.9493 | 0.8230 | 0.7753 | |
GCML (ours) | 0.00836 | 0.99868 | 0.98203 | 13.3788 | 1.33893 | 1.00 | 0.8642 | 0.9942 | 0.8468 | 0.7768 |
Datasets | Methods | ACC/NMI | RRE | Trust | Cont | -RMSE | LGD | CRA |
---|---|---|---|---|---|---|---|---|
w/o SL | 0.976/0.939 | 0.0093 | 0.9967 | 0.9816 | 24.589 | 1.6747 | 0.32 | |
w/o CL | 0.814/0.736 | 0.0004 | 0.9998 | 0.9990 | 7.458 | 0.0487 | 1.00 | |
w/o WC | 0.977/0.943 | 0.0065 | 0.9987 | 0.9860 | 5.576 | 0.6968 | 0.98 | |
w/o AT | 0.978/0.944 | 0.0069 | 0.9986 | 0.9851 | 5.617 | 0.7037 | 0.96 | |
MNIST-full | full model | 0.980/0.946 | 0.0056 | 0.9997 | 0.9871 | 5.498 | 0.6916 | 1.00 |
w/o SL | 0.706/0.682 | 0.0108 | 0.9964 | 0.9781 | 25.954 | 1.8936 | 0.30 | |
w/o CL | 0.576/0.569 | 0.0004 | 0.9994 | 0.9995 | 7.654 | 0.0523 | 1.00 | |
w/o WC | 0.702/0.695 | 0.0084 | 0.9972 | 0.9814 | 13.238 | 1.3474 | 1.00 | |
w/o AT | 0.708/0.694 | 0.0097 | 0.9975 | 0.9798 | 13.354 | 1.3611 | 1.00 | |
Fashion-MNIST | full model | 0.710/0.685 | 0.0083 | 0.9986 | 0.9820 | 13.378 | 1.3389 | 1.00 |
4.3 Evaluation of Multi-Manifold Learning
Quantitative Metrics. Though many previous works have claimed that they brought clustering and dimensionality reduction into a unified framework, unfortunately, they all lacked an analysis of the effectiveness of the learned embeddings. In this paper, we compare GCML with the other five methods in six quantitative metrics on seven datasets. Limited by space, only the results of MNIST-full and Fashion-MNIST are provided on the left side of Tab. 3 and more results are in Appendix A.5. The results show that GCML outperforms all other methods, especially in the CRA metric, which is not only the best on all datasets but reaches 1.0, which means that the “rank” between different manifolds in the latent space is completely preserved and proves the effectiveness of the global ranking loss .
Statistical Analysis. The statistical analysis is performed to show the extent to which local and global structure is preserved in the latent space for each algorithm. Taking MNIST-full as an example, the statistical analysis of the global rank-preservation is shown in Fig. 6 (a)-(f). For the -th cluster, if the rank (in terms of Euclidean distance) between it and the -th cluster is preserved from the input space to the latent space, then the grid in the -th row and -th column is marked as blue, otherwise yellow. As shown in the figure, only GCML can fully preserve the global rank between different clusters, while all other methods fail.
Moreover, we perform statistical analysis for the local isometry property of each algorithm. For each sample , it forms a number of point pairs with its neighborhood samples . We compute the differences in the distance of these point pairs from the input space to the latent space , and plot them as a histogram. As shown in Fig. 6 (g), the curve of GCML are distributed on both sides of the 0 value, with maximum peak height and minimum peak-bottom width, respectively, which indicates that GCML achieves the best local isometry. Though IDEC [4] claims that they can preserve the local structure well, their results are still far from ours.
Downstream Tasks. Numerous deep clustering algorithms have recently claimed to obtain meaningful low-dimensional embeddings; however, they have not analyzed and experimented with the “meaningful” ones. Therefore, we are interested in whether these proposed methods can really learn manifold embeddings that are useful for downstream tasks. Four different classifiers, including a linear classifier (Logistic Regression; LR), two nonlinear classifiers (MLP, SVM), and a tree-based classifier (Random Forest Classifier; RFC) are used as downstream tasks, all of which use default parameters and default implementations in sklearn [15] for a fair comparison. The learned embeddings are frozen and used as input for training. The classification accuracy evaluated on the test set serves as a metric to evaluate the effectiveness of learned embeddings. Limited by space, only the results of MNIST-full and Fashion-MNIST are provided on the right side of Tab. 3 and more results are in Appendix A.5. The results show that GCML outperforms the other methods overall on all seven datasets, with MLP, RFC, and LR as downstream tasks.
4.4 Ablation Study
Tab. 4 evaluates the effectiveness of the proposed loss terms and training strategies with 5 sets of experiments: the model without (A) Structure-oriented Loss (SL); (B) Clustering-oriented Loss (CL); (C) Weight Continuation (WC); (D) Alternating Training (AT), and (E) the full model. Limited by space, only the results MNIST-full and Fashion-MNIST are provided and more results are in Appendix A.6. After analyzing the results, we can conclude: (1) CL is the most important factor for obtaining excellent clustering performance, the lack of which leads to unsuccessful clustering, hence the numbers in the table are not meaningful and marked in gray color. (2) SL not only brings subtle improvements in clustering performance but greatly improves the performance of multi-manifold learning. (3) The training strategies (WC and AT) both improve the performance of clustering and multi-manifold learning to some extent, especially on metrics such as RRE, Trust, CRA, etc.
5 Conclusion
The proposed GCML framework imposes clustering-oriented and structure-oriented constraints to optimize the latent space for simultaneously performing clustering and multi-manifold learning with geometric structure preservation. Extensive experiments demonstrate that GCML is not only comparable to the SOTA clustering algorithms but learns effective manifold embeddings, which is beyond the capability of those algorithms that only care about clustering accuracy. Finally, GCML is appliable to generalized data with various dimensions, not limited to 2-D image data.
References
- [1] Christopher M Bishop. Pattern recognition and machine learning. springer, 2006.
- [2] Zhangyang Gao, Haitao Lin, Stan Li, et al. Clustering based on graph of density topology. arXiv preprint arXiv:2009.11612, 2020.
- [3] Kamran Ghasedi Dizaji, Amirhossein Herandi, Cheng Deng, Weidong Cai, and Heng Huang. Deep clustering via joint convolutional autoencoder embedding and relative entropy minimization. In Proceedings of the IEEE international conference on computer vision, pages 5736–5745, 2017.
- [4] Xifeng Guo, Long Gao, Xinwang Liu, and Jianping Yin. Improved deep embedded clustering with local structure preservation. In IJCAI, pages 1753–1759, 2017.
- [5] Xifeng Guo, Xinwang Liu, En Zhu, Xinzhong Zhu, Miaomiao Li, Xin Xu, and Jianping Yin. Adaptive self-paced deep clustering with data augmentation. IEEE Transactions on Knowledge and Data Engineering, 2019.
- [6] Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9729–9738, 2020.
- [7] Zhuxi Jiang, Yin Zheng, Huachun Tan, Bangsheng Tang, and Hanning Zhou. Variational deep embedding: An unsupervised and generative approach to clustering. arXiv preprint arXiv:1611.05148, 2016.
- [8] Yann LeCun, Léon Bottou, Yoshua Bengio, and Patrick Haffner. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278–2324, 1998.
- [9] Stan Z Li, Lirong Wu, and Zelin Zang. Consistent representation learning for high dimensional data analysis. arXiv preprint arXiv:2012.00481, 2020.
- [10] Laurens van der Maaten and Geoffrey Hinton. Visualizing data using t-sne. Journal of machine learning research, 9(Nov):2579–2605, 2008.
- [11] J MacQueen. Some methods for classification and analysis of multivariate observations. In Proc. 5th Berkeley Symposium on Math., Stat., and Prob, page 281, 1965.
- [12] Ryan McConville, Raul Santos-Rodriguez, Robert J Piechocki, and Ian Craddock. N2d:(not too) deep clustering via clustering the local manifold of an autoencoded embedding. arXiv preprint arXiv:1908.05968, 2019.
- [13] Leland McInnes, John Healy, and James Melville. Umap: Uniform manifold approximation and projection for dimension reduction. arXiv preprint arXiv:1802.03426, 2018.
- [14] Sudipto Mukherjee, Himanshu Asnani, Eugene Lin, and Sreeram Kannan. Clustergan: Latent space clustering in generative adversarial networks. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 4610–4617, 2019.
- [15] F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay. Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825–2830, 2011.
- [16] Yazhou Ren, Ni Wang, Mingxia Li, and Zenglin Xu. Deep density-based image clustering. Knowledge-Based Systems, 197:105841, 2020.
- [17] Sam T Roweis and Lawrence K Saul. Nonlinear dimensionality reduction by locally linear embedding. science, 290(5500):2323–2326, 2000.
- [18] Uri Shaham, Kelly Stanton, Henry Li, Boaz Nadler, Ronen Basri, and Yuval Kluger. Spectralnet: Spectral clustering using deep neural networks. arXiv preprint arXiv:1801.01587, 2018.
- [19] Jianbo Shi and Jitendra Malik. Normalized cuts and image segmentation. IEEE Transactions on pattern analysis and machine intelligence, 22(8):888–905, 2000.
- [20] Joshua B Tenenbaum, Vin De Silva, and John C Langford. A global geometric framework for nonlinear dimensionality reduction. science, 290(5500):2319–2323, 2000.
- [21] Fei Tian, Bin Gao, Qing Cui, Enhong Chen, and Tie-Yan Liu. Learning deep representations for graph clustering. In Aaai, volume 14, pages 1293–1299. Citeseer, 2014.
- [22] Wouter Van Gansbeke, Simon Vandenhende, Stamatios Georgoulis, Marc Proesmans, and Luc Van Gool. Scan: Learning to classify images without labels. In Proceedings of the European Conference on Computer Vision (ECCV), 2020.
- [23] Han Xiao, Kashif Rasul, and Roland Vollgraf. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. arXiv preprint arXiv:1708.07747, 2017.
- [24] Junyuan Xie, Ross Girshick, and Ali Farhadi. Unsupervised deep embedding for clustering analysis. In International conference on machine learning, pages 478–487, 2016.
- [25] Bo Yang, Ming Xiang, and Yupei Zhang. Multi-manifold discriminant isomap for visualization and classification. Pattern Recognition, 55:215–230, 2016.
- [26] Jianwei Yang, Devi Parikh, and Dhruv Batra. Joint unsupervised learning of deep representations and image clusters. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 5147–5156, 2016.
- [27] Xu Yang, Cheng Deng, Feng Zheng, Junchi Yan, and Wei Liu. Deep spectral clustering using dual autoencoder network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 4066–4075, 2019.
- [28] Xiaohang Zhan, Jiahao Xie, Ziwei Liu, Yew-Soon Ong, and Chen Change Loy. Online deep clustering for unsupervised representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6688–6697, 2020.
- [29] Yan Zhang, Zhao Zhang, Jie Qin, Li Zhang, Bing Li, and Fanzhang Li. Semi-supervised local multi-manifold isomap by linear embedding for feature extraction. Pattern Recognition, 76:662–678, 2018.
Appendix
A.1 Datasets
To show that our method works well with various kinds of datasets, we choose the following seven image and vector datasets. A brief description is given in Tab. A1.
(1). MNIST-full [8]: The MNIST-full dataset consists of 70000 handwritten digits of 28 28 pixels. Each gray image is reshaped to a 784-dimensional vector.
(2). MNIST-test [8]: The MNIST-test is the testing part of MNIST dataset, which contains a total of 10000 samples.
(3). USPS: The USPS dataset is composed of 9298 gray-scale handwritten digit images with a size of 1616 pixels.
(4). Fashion-MNIST [23]: This Fashion-MNIST dataset has the same number of images and the same image size as MNIST-full, but it is fairly more complicated. Instead of digits, it consists of various types of fashion products.
(5). REUTERS-10K: REUTERS-10K is a subset of REUTERS with 10000 samples. Four root categories (corporate/industrial, government/social, markets, and economics) are used as labels, and the tf-idf features on the 2000 most frequent words are computed.
(6). HAR: HAR is a time-series dataset consisting of 10299 sensor samples from a smartphone. It was collected from 30 people performing six activities: walking, walking upstairs, walking downstairs, sitting, standing, and laying.
(7). Pendigits: HAR is a dataset consisting of 10992 samples from pressure-sensitive plates with ten different digits, each represented by 8 coordinates of the stylus.
Dataset | Samples | Categories | Data Size |
---|---|---|---|
MNIST-full | 70000 | 10 | 28281 |
MNIST-test | 10000 | 10 | 28281 |
USPS | 9298 | 10 | 16161 |
Fashion-MNIST | 70000 | 10 | 28281 |
REUTERS-10K | 10000 | 4 | 2000 |
HAR | 10299 | 6 | 561 |
Pendigits | 10992 | 10 | 16 |
A.2 Gradient Derivation
In the paper, we have emphasized time and again that is a set of learnable parameters, which means that we can optimize it while optimizing the network parameter . In Eq. (4) of the paper, we have presented the gradient of with respect to . In addition to , both and are involving . Hence, the detailed derivations for the gradient of and with respect to are also provided. The gradient of with respect to each cluster center can be computed as:
The Euclidean metric is used for both the input space and the hidden layer space, i.e., . In addition, the symbols are somewhat abused for clear derivation, representing with . Accordingly, the above equation can be further derived as follows:
The gradient of with respect to each learnalbe cluster center can be computed as:
Parameters | ACC/NMI | RRE | Trust | Cont | -RMSE | LGD | CRA |
---|---|---|---|---|---|---|---|
=1, =3 | 0.975/0.936 | 0.0125 | 0.9944 | 0.9756 | 5.757 | 0.8868 | 1.00 |
=3, =3 | 0.973/0.931 | 0.0114 | 0.9970 | 0.9757 | 5.805 | 0.9207 | 1.00 |
=5, =3 | 0.972/0.930 | 0.0109 | 0.9981 | 0.9761 | 5.800 | 0.9339 | 1.00 |
=8, =3 | 0.972/0.929 | 0.0104 | 0.9989 | 0.9765 | 5.810 | 0.9476 | 1.00 |
=10, =3 | 0.972/0.929 | 0.0105 | 0.9990 | 0.9764 | 5.704 | 0.9487 | 1.00 |
=5, =1 | 0.967/0.912 | 0.0068 | 0.9993 | 0.9845 | 5.409 | 0.2524 | 1.00 |
=5, =3 | 0.972/0.930 | 0.0109 | 0.9981 | 0.9761 | 5.800 | 0.9339 | 1.00 |
=5, =5 | 0.972/0.929 | 0.0146 | 0.9964 | 0.9691 | 15.0653 | 1.5719 | 1.00 |
=5, =8 | 0.972/0.929 | 0.0190 | 0.9943 | 0.9615 | 29.4607 | 2.5410 | 1.00 |
=5, =10 | 0.972/0.929 | 0.0195 | 0.9951 | 0.9597 | 37.7661 | 3.1434 | 1.00 |
A.3 Definitions of Performance Metrics
The following notations are used for the definitions:
-
: the pairwise distance between and in input space ;
-
: the pairwise distance between and in latent space ;
-
: the set of indices to the -nearest neighbor (NN) of in input space ;
-
: the set of indices to the -nearest neighbor (NN) of in latent space ;
-
: the rank of the closeness (in Euclidean distance) of to in input space ;
-
: the rank of the closeness (in Euclidean distance) of to in latent space .
The eight evaluation metrics are defined below:
-
(1)
ACC (Accuracy) measures the accuracy of clustering:
where and are the true and predicted labels for data point , respectively, and is all possible one-to-one mappings between clusters and label categories.
-
(2)
NMI (Normalized Mutual Information) NMI calculates the normalized measure of similarity between two labels of the same data
where is the mutual information between the real label and predicted label , and represents their entropy.
-
(3)
RRE (Relative Rank Change) measures the average of changes in neighbor ranking between two spaces and :
where and are the lower and upper bounds of the -NN.
where is the normalizing term, defined as
-
(4)
Trust (Trustworthiness) measures to what extent the nearest neighbors of a point are preserved when going from the input space to the latent space:
where and are the bounds of the number of nearest neighbors.
-
(5)
Cont (Continuity) is defined analogously to , but checks to what extent neighbors are preserved when going from the latent space to the input space:
where and are the bounds of the number of nearest neighbors.
-
(6)
-RMSE (Root Mean Square Error) measures to what extent the two distributions of distances coincide:
-
(7)
LGD (Locally Geometric Distortion) measures how much corresponding distances between neighboring points differ in two metric spaces and is the primary metric for isometry, defined as:
where and are the lower and upper bounds of the -NN.
-
(8)
CRA (Cluster Rank Accuracy) measures the changes in ranks of cluster centers from the input space and to the latent space :
where is the number of clusters, is the cluster center of the th cluster in the input space , is the cluster center of the th cluster in the latent space , denotes the rank of the closeness (in terms of Euclidean distance) of to in space in the input space , and denotes the rank of the closeness (in terms of Euclidean distance) of to in space .
A.4 Parameter Sensitivity
We also evaluated the sensitivity of parameters and on the MNIST-test dataset and the results are shown in Tab. A2. It is found that the clustering performance is not sensitive to parameters and , and some combinations of and even produce better clustering performance than the metrics reported in the main paper. However, the effect of and on multi-manifold learning is more pronounced, and different combinations of and may increase or decrease performance. In general, this paper focuses on the design of algorithms and implementations, and no hyperparameter search is performed to find the best performance metrics.
A.5 Evaluation of Multi-Manifold Learning
GCML is compared with other methods on six manifold-related quantitative metrics to demonstrate its performance advantage for multi-manifold learning, and the complete results on seven datasets are shown on the left side of Tab. A3. In addition, the right side of Tab. A3 compares GCML with other methods on all datasets to see if these methods can really learn embeddings that are useful for downstream tasks. As shown in the table, GCML outperforms all the other methods with MLP, RFC, and LR as downstream tasks.
A.6 More Ablation Study Experiments
The results of the ablation study on the MNIST-full dataset have been presented in Tab. 4 in Sec 4.4. Here, we provide complete ablation study results on all datasets in Tab. A4. The conclusion is similar (note that the clustering performance of the model without clustering-oriented losses is very poorly, so the “best” metric numbers are not meaningful and are marked in gray color): (1) CL is very important for obtaining good clustering. (2) SL is beneficial for both clustering and multi-manifold learning. (3) The training strategies (WC and AT) contribute to improving performance metrics such as RRE, Trust, Cont, and CRA.
Datasets | Algorithms | RRE | Trust | Cont | -RMSE | LGD | CRA | MLP | RFC | SVM | LR |
---|---|---|---|---|---|---|---|---|---|---|---|
MNIST-full | DEC | 0.09988 | 0.84499 | 0.94805 | 44.8535 | 4.37986 | 0.28 | 0.8647 | 0.8706 | 0.8707 | 0.8566 |
IDEC | 0.00984 | 0.99821 | 0.97936 | 24.5803 | 1.71484 | 0.33 | 0.9797 | 0.9737 | 0.9852 | 0.9650 | |
JULE | 0.02657 | 0.93675 | 0.98321 | 28.3412 | 2.12955 | 0.27 | 0.9802 | 0.9825 | 0.9787 | 0.9743 | |
DSC | 0.09785 | 0.87315 | 0.92508 | 6.98098 | 1.19886 | 0.23 | 0.9622 | 0.9501 | 0.9837 | 0.9752 | |
N2D | 0.01002 | 0.99243 | 0.98466 | 5.7162 | 0.69946 | 0.21 | 0.9796 | 0.9803 | 0.9799 | 0.9792 | |
GCML (ours) | 0.00567 | 0.99978 | 0.98716 | 5.4986 | 0.69168 | 1.00 | 0.9851 | 0.9874 | 0.9869 | 0.9841 | |
MNIST-test | DEC | 0.12800 | 0.81841 | 0.91767 | 14.6113 | 2.29499 | 0.19 | 0.8525 | 0.8605 | 0.8725 | 0.8685 |
IDEC | 0.01505 | 0.99403 | 0.97082 | 7.4599 | 1.08350 | 0.38 | 0.9740 | 0.9725 | 0.9845 | 0.9655 | |
JULE | 0.04122 | 0.92971 | 0.97208 | 9.4768 | 1.17176 | 0.42 | 0.9775 | 0.9845 | 0.9800 | 0.9825 | |
DSC | 0.10728 | 0.85498 | 0.92254 | 7.1689 | 1.19239 | 0.26 | 0.9535 | 0.9740 | 0.9825 | 0.9795 | |
N2D | 0.01565 | 0.98764 | 0.97572 | 5.0120 | 0.97454 | 0.33 | 0.9715 | 0.9760 | 0.9725 | 0.9725 | |
GCML (ours) | 0.01090 | 0.99811 | 0.97612 | 5.8000 | 0.93394 | 1.00 | 0.9855 | 0.9875 | 0.9865 | 0.9855 | |
USPS | DEC | 0.07911 | 0.88871 | 0.94628 | 16.4355 | 1.77848 | 0.31 | 0.8289 | 0.8668 | 0.8289 | 0.8294 |
IDEC | 0.01043 | 0.99726 | 0.97960 | 13.0573 | 1.11689 | 0.30 | 0.9482 | 0.9556 | 0.9656 | 0.9125 | |
JULE | 0.02972 | 0.98763 | 0.98810 | 14.6324 | 1.43426 | 0.33 | 0.9576 | 0.9617 | 0.9703 | 0.9476 | |
DSC | 0.06319 | 0.9151 | 0.93988 | 8.4412 | 1.02131 | 0.27 | 0.9351 | 0.9572 | 0.9612 | 0.9342 | |
N2D | 0.01337 | 0.98769 | 0.98135 | 8.1961 | 0.54967 | 0.37 | 0.9569 | 0.9569 | 0.9569 | 0.9541 | |
GCML (ours) | 0.00577 | 0.99979 | 0.98701 | 6.4980 | 0.53180 | 1.00 | 0.9656 | 0.9651 | 0.9604 | 0.9551 | |
Fashion-MNIST | DEC | 0.04787 | 0.93896 | 0.95450 | 39.3274 | 3.87731 | 0.37 | 0.6268 | 0.9853 | 0.6377 | 0.6245 |
IDEC | 0.01089 | 0.99683 | 0.97797 | 25.4024 | 1.91385 | 0.27 | 0.8367 | 0.9918 | 0.8607 | 0.7514 | |
JULE | 0.03013 | 0.97732 | 0.97923 | 15.2213 | 1.43642 | 0.43 | 0.8541 | 0.9892 | 0.8566 | 0.7723 | |
DSC | 0.05168 | 0.95013 | 0.96121 | 17.2201 | 1.42091 | 0.36 | 0.8084 | 0.9823 | 0.8618 | 0.7676 | |
N2D | 0.00894 | 0.99062 | 0.98054 | 14.49079 | 1.28180 | 0.26 | 0.8412 | 0.9493 | 0.8230 | 0.7753 | |
GCML (ours) | 0.00836 | 0.99868 | 0.98203 | 13.3788 | 1.33893 | 1.00 | 0.8642 | 0.9942 | 0.8468 | 0.7768 | |
REUTERS-10K | DEC | 0.26192 | 0.65518 | 0.80477 | 40.4671 | 4.00423 | 0.63 | 0.7985 | 0.7880 | 0.8105 | 0.7450 |
IDEC | 0.05981 | 0.95840 | 0.90550 | 43.9556 | 2.01365 | 0.75 | 0.9225 | 0.8930 | 0.9280 | 0.7705 | |
JULE | - | - | - | - | - | - | - | - | - | - | |
DSC | - | - | - | - | - | - | - | - | - | - | |
N2D | 0.03827 | 0.97385 | 0.93412 | 36.1042 | 1.69013 | 0.31 | 0.9205 | 0.9080 | 0.9240 | 0.8335 | |
GCML (ours) | 0.03206 | 0.98380 | 0.93802 | 34.5478 | 2.72096 | 1.00 | 0.9360 | 0.9185 | 0.9390 | 0.8475 | |
HAR | DEC | 0.09060 | 0.89097 | 0.91766 | 10.0222 | 1.58691 | 0.30 | 0.7696 | 0.7847 | 0.7628 | 0.7634 |
IDEC | 0.01031 | 0.99433 | 0.98132 | 9.9155 | 0.93736 | 0.39 | 0.8973 | 0.9031 | 0.9041 | 0.8822 | |
JULE | - | - | - | - | - | - | - | - | - | - | |
DSC | - | - | - | - | - | - | - | - | - | - | |
N2D | 0.00841 | 0.99281 | 0.97695 | 8.2326 | 0.64296 | 0.33 | 0.9138 | 0.9083 | 0.9174 | 0.8799 | |
GCML (ours) | 0.00665 | 0.99895 | 0.98634 | 15.2876 | 0.46189 | 1.00 | 0.9235 | 0.9193 | 0.9293 | 0.8996 | |
pendigits | DEC | 0.03932 | 0.95149 | 0.96300 | 21.6608 | 1.32970 | 0.24 | 0.7215 | 0.8432 | 0.7376 | 0.7429 |
IDEC | 0.00384 | 0.99879 | 0.99255 | 19.7243 | 0.86137 | 0.28 | 0.8870 | 0.9461 | 0.9595 | 0.8636 | |
JULE | - | - | - | - | - | - | - | - | - | - | |
DSC | - | - | - | - | - | - | - | - | - | - | |
N2D | 0.00262 | 0.99919 | 0.99473 | 20.7052 | 0.76941 | 0.42 | 0.9131 | 0.9669 | 0.9528 | 0.8516 | |
GCML (ours) | 0.00091 | 0.99994 | 0.99808 | 2.5184 | 0.08223 | 1.00 | 0.9538 | 0.9705 | 0.9532 | 0.8953 |
Datasets | Methods | ACC/NMI | RRE | Trust | Cont | -RMSE | LGD | CRA |
---|---|---|---|---|---|---|---|---|
w/o SL | 0.976/0.939 | 0.0093 | 0.9967 | 0.9816 | 24.589 | 1.6747 | 0.32 | |
w/o CL | 0.814/0.736 | 0.0004 | 0.9998 | 0.9990 | 7.458 | 0.0487 | 1.00 | |
w/o WC | 0.977/0.943 | 0.0065 | 0.9987 | 0.9860 | 5.576 | 0.6968 | 0.98 | |
w/o AT | 0.978/0.944 | 0.0069 | 0.9986 | 0.9851 | 5.617 | 0.7037 | 0.96 | |
MNIST-full | full model | 0.980/0.946 | 0.0056 | 0.9997 | 0.9871 | 5.498 | 0.6916 | 1.00 |
w/o SL | 0.973/0.932 | 0.0146 | 0.9928 | 0.9727 | 7.701 | 1.0578 | 0.31 | |
w/o CL | 0.773/0.747 | 0.0020 | 0.9994 | 0.9954 | 7.229 | 0.0809 | 1.00 | |
w/o WC | 0.956/0.904 | 0.0132 | 0.9955 | 0.9735 | 5.470 | 0.9364 | 1.00 | |
w/o AT | 0.970/0.929 | 0.0118 | 0.9974 | 0.9747 | 5.567 | 0.9404 | 1.00 | |
MNIST-test | full model | 0.972/0.930 | 0.0109 | 0.9981 | 0.9761 | 5.800 | 0.9339 | 1.00 |
w/o SL | 0.957/0.902 | 0.0095 | 0.9967 | 0.9812 | 14.609 | 0.9847 | 0.29 | |
w/o CL | 0.664/0.658 | 0.0020 | 0.9996 | 0.9952 | 2.934 | 0.0687 | 1.0 | |
w/o WC | 0.956/0.896 | 0.0060 | 0.9991 | 0.9868 | 6.572 | 0.5335 | 1.00 | |
w/o AT | 0.947/0.885 | 0.0080 | 0.9979 | 0.9833 | 5.960 | 0.4967 | 1.00 | |
USPS | full model | 0.958/0.902 | 0.0057 | 0.9997 | 0.9870 | 6.498 | 0.5318 | 1.00 |
w/o SL | 0.706/0.682 | 0.0108 | 0.9964 | 0.9781 | 25.954 | 1.8936 | 0.30 | |
w/o CL | 0.576/0.569 | 0.0004 | 0.9994 | 0.9995 | 7.654 | 0.0523 | 1.00 | |
w/o WC | 0.702/0.695 | 0.0084 | 0.9972 | 0.9814 | 13.238 | 1.3474 | 1.00 | |
w/o AT | 0.708/0.694 | 0.0097 | 0.9975 | 0.9798 | 13.354 | 1.3611 | 1.00 | |
Fashion-MNIST | full model | 0.710/0.685 | 0.0083 | 0.9986 | 0.9820 | 13.378 | 1.3389 | 1.00 |
w/o SL | 0.819/0.564 | 0.0529 | 0.9610 | 0.9185 | 44.481 | 1.9090 | 0.38 | |
w/o CL | 0.542/0.279 | 0.0277 | 0.9868 | 0.9456 | 37.018 | 2.2294 | 1.00 | |
w/o WC | 0.830/0.583 | 0.0420 | 0.9667 | 0.9361 | 35.302 | 2.8286 | 1.00 | |
w/o AT | 0.825/0.563 | 0.0440 | 0.9650 | 0.9330 | 39.275 | 2.9146 | 1.00 | |
REUTERS-10K | full model | 0.836/0.590 | 0.0320 | 0.9838 | 0.9380 | 34.547 | 2.7209 | 1.00 |
w/o SL | 0.835/0.746 | 0.0116 | 0.9944 | 0.9792 | 8.168 | 0.8882 | 0.33 | |
w/o CL | 0.744/0.615 | 0.0024 | 0.9986 | 0.9948 | 15.060 | 0.2193 | 1.00 | |
w/o WC | 0.786/0.701 | 0.0130 | 0.9950 | 0.9756 | 15.398 | 0.6171 | 1.00 | |
w/o AT | 0.834/0.745 | 0.0089 | 0.9965 | 0.9835 | 15.726 | 0.4734 | 1.00 | |
HAR | full model | 0.844/0.862 | 0.0066 | 0.9989 | 0.9863 | 15.287 | 0.4618 | 1.00 |
w/o SL | 0.843/0.803 | 0.0027 | 0.9989 | 0.9948 | 20.692 | 0.7587 | 0.31 | |
w/o CL | 0.773/0.705 | 0.0004 | 0.9998 | 0.9992 | 2.177 | 0.0385 | 1.00 | |
w/o WC | 0.840/0.790 | 0.0014 | 0.9995 | 0.9972 | 2.690 | 0.1124 | 1.00 | |
w/o AT | 0.842/0.804 | 0.0018 | 0.9991 | 0.9965 | 2.756 | 0.1148 | 1.00 | |
pendigits | full model | 0.855/0.814 | 0.0009 | 0.9999 | 0.9981 | 2.518 | 0.0822 | 1.00 |