This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Tucker-O-Minus Decomposition for Multi-view Tensor Subspace Clustering

Yingcong Lu, Yipeng Liu, Zhen Long, Zhangxin Chen, Ce Zhu All the authors are with the School of Information and Communication Engineering, University of Electronic Science and Technology of China (UESTC), Chengdu, 611731, China. E-mail: [email protected]
Abstract

With powerful ability to exploit latent structure of self-representation information, different tensor decompositions have been employed into low rank multi-view clustering (LRMVC) models for achieving significant performance. However, current approaches suffer from a series of problems related to those tensor decomposition, such as the unbalanced matricization scheme, rotation sensitivity, deficient correlations capture and so forth. All these will lead to LRMVC having insufficient access to global information, which is contrary to the target of multi-view clustering. To alleviate these problems, we propose a new tensor decomposition called Tucker-O-Minus Decomposition (TOMD) for multi-view clustering. Specifically, based on the Tucker format, we additionally employ the O-minus structure, which consists of a circle with an efficient bridge linking two weekly correlated factors. In this way, the core tensor in Tucker format is replaced by the O-minus architecture with a more balanced structure, and the enhanced capacity of capturing the global low rank information will be achieved. The proposed TOMD also provides more compact and powerful representation abilities for the self-representation tensor, simultaneously. The alternating direction method of multipliers is used to solve the proposed model TOMD-MVC. Numerical experiments on six benchmark data sets demonstrate the superiority of our proposed method in terms of F-score, precision, recall, normalized mutual information, adjusted rand index, and accuracy.

1 Introduction

Tensor, a higher order generation of vector and matrix, provides a natural representation for high order data. For example, ORL multi-view data set [54] is a 3rd-order tensor 𝒳I×Cv×V\mathcal{X}^{I\times C_{v}\times V} (v=1,,V)(v=1,\cdots,V), where II is the number of samples, CvC_{v} is the feature size of the vv-th view, and VV is the number of views. Tensor decomposition allows us to simultaneously explore all dimensions to obtain more latent information, which has attracted attention in a series of fields, e.g., image processing [8, 32], machine learning [15, 16] and signal processing [5, 1]. Tensor-based multi-view clustering (MVC) is one of them, which separates multi-view data into clusters by exploiting their latent information.

Most tensor MVC methods are based on the assumption that their self-representation tensors are low rank [53]. For example, Chen et al. [7] combine the low-rank tensor graph and the subspace clustering into a unified model in multi-view clustering, which applies the Tucker decomposition [41, 42] to explore the low-rank information of representation tensor. With the emergence of a new type of decomposition for 3rd-order tensors [21, 20], multiple multi-view subspace clustering methods based on the tensor singular value decomposition (t-SVD) have been proposed [45, 46, 58]. In addition, tensor networks provide a more compact and flexible representation for higher order tensor than traditional tensor decomposition [27, 30, 31, 33]. In this way, Yu et al. [51] have proposed a novel non-negative tensor ring (NTR) decomposition and graph-regularized NTR (GNTR) for the non-negative multi-way representation learning with satisfying performance in clustering tasks.

Although the aforementioned methods have achieved promising clustering performance, there are still several problems. Since the Tucker rank is related to the unbalanced mode-nn unfolding matrices [2], capturing the global information of the self-representation tensor by simply processing Tucker decomposition may be difficult in LRMVC. Besides, the t-SVD suffers from rotation sensitivity and the low rank information can not be fully discovered in the 3rd mode [10]. Thus, the t-SVD based methods always need to utilize the rotation operation of the self-representation tensor to explore the correlations across different views. Naturally, the acquisition of the correlations among samples will be inadequate. Furthermore, tensor ring (TR) has shown better performance in exploring the low-rank information with the well-balanced matricization scheme. However, the interaction of neighboring modes in TR is stronger than that of two modes separated by a large distance, whose correlations are also have been ignored.

Considering the above problems, we propose the Tucker-O-Minus decomposition for multi-view subspace clustering. For employing the additional factors and correlations to construct a more compact and balanced tensor network, a simple way seems to replace the core factor of the Tucker structure with the TR format. However, as we mentioned above, the TR architecture suffers from a deficiency in exploring the correlations between the weakly-connected modes, which will result in the loss of essential clustering information in LRMVC. To this end, we propose the O-minus structure for LRMVC in the following three considerations. Firstly, from the perspective of two-point correlations in high-energy physics [3], appending moderate lines with effective vertexes between two slightly correlated factors will strengthen their relationships. Similarly, we try to add the ”bridge” with a tensor factor based on the TR structure to better capture the low rank information. Nonetheless, more links will generate more loops, which will cause huge difficulties in tensor contraction and the computational complexity burden of the tensor networks [9]. Simultaneously, since the number of samples IVI\gg V in LRMVC, compared with higher-order information across different views, the information related to the correlations of instances requires more connections to be fully discovered. Accordingly, to efficiently explore the low rank information in LRMVC, the correlations related to samples (i.e., I1I_{1}, and I3I_{3} in Fig. 1) are further strengthened by the special ”bridge”. This unique internal architecture is similar to ”\ominus”, thus namely O-Minus. And the whole architecture called Tucker-O-Minus decomposition is illustrated in Fig. 1. In this way, the low rank based multi-view clustering problem can be successfully solved with satisfying results. The main contributions of this paper are summarized as follows:

  1. 1.

    We propose the Tucker-O-Minus decomposition. Different from the existing tensor networks, it allows more valid interaction among nonadjacent factors and obtains a better low rank representation for high-dimensional data. Numerical experimental results on gray image reconstruction show that the performance of TOMD in exploring low rank information is superior to others.

  2. 2.

    We apply TOMD to a unified multi-view clustering framework. The proposed model, namely TOMD-MVC, utilizes the TOMD to capture the low-rank properties of self-representation tensor from multi-view data. The alternating direction method of multipliers (ADMM) is applied to solve the optimization model.

  3. 3.

    We conduct extensive experiments on six real-world multi-view data sets to demonstrate the performance of our method. Compare with state-of-the-art models, TOMD-MVC achieves highly competitive or even better performance in terms of six evaluation metrics.

The remainder of this paper is organized as follows. Section 2 gives the used notations, mathematical backgrounds, and review the related works. The proposed tensor network is presented in details in Section 3. Section 4 gives the multi-view clustering method based on the newly proposed tensor network. The experimental results are demonstrated and analyzed in Section 5. Section 6 draws the conclusion finally.

Refer to caption
Figure 1: The graphical illustration of TOMD for a 4th-order tensor 𝒳I1×I2×I3×I4\mathcal{X}\in\mathbb{R}^{I_{1}\times I_{2}\times I_{3}\times I_{4}}, where 𝐔(i)Ii×Ri\mathbf{U}^{(i)}\in\mathbb{R}^{I_{i}\times R_{i}} (i=1,,4)(i=1,\cdots,4) are factor matrices, 𝒢(n)\mathcal{G}^{(n)}, n=1,,5n=1,\cdots,5, are the sub-tensors.

2 Notations, Preliminaries, and Related Works

2.1 Notations

We give a brief introduction about notations and preliminaries in this section. Throughout this paper, we use lower case letters, bold lower case letters, bold upper case letters, and calligraphic letters to denote scalars, vectors, matrices and tensors, respectively, e.g., aa, 𝐚\mathbf{a}, 𝐀\mathbf{A} and 𝒜\mathcal{A}. Some other frequently used notations are listed in Table 1, where 𝒜\mathcal{A} is a 3rd-order tensor and 𝐀\mathbf{A} is a matrix.

Table 1: Summary of notations in this paper.
Denotation Definition
𝐀i\mathbf{A}_{i} 𝐀i=𝒜(:,:,i)\mathbf{A}_{i}=\mathcal{A}(:,:,i)
ai1,i2,i3a_{i_{1},i_{2},i_{3}} The (i1,i2,i3)(i_{1},i_{2},i_{3})-th entry of tensor 𝒜\mathcal{A}
tr(𝐀)\text{tr}(\mathbf{A}) tr(𝐀)=iai,i\text{tr}(\mathbf{A})=\sum_{i}a_{i,i}
𝐀F\|\mathbf{A}\|_{\mathrm{F}} 𝐀F=i1i2ai1,i22\|\mathbf{A}\|_{\mathrm{F}}=\sqrt{\sum_{i_{1}i_{2}}a_{i_{1},i_{2}}^{2}}
𝐀\|\mathbf{A}\|_{\infty} 𝐀=maxi1i2|𝐀i1,i2|\|\mathbf{A}\|_{\infty}=\max_{i_{1}i_{2}}\left|\mathbf{A}_{i_{1},i_{2}}\right|
𝐀2,1\|\mathbf{A}\|_{2,1} 𝐀2,1=j𝐀(:,j)2\|\mathbf{A}\|_{2,1}=\sum_{j}\|\mathbf{A}(:,j)\|_{2}
𝒜F\|\mathcal{A}\|_{\mathrm{F}} 𝒜F=i1i2i3|ai1,i2,i3|2\|\mathcal{A}\|_{\mathrm{F}}=\sqrt{\sum_{i_{1}i_{2}i_{3}}\left|a_{i_{1},i_{2},i_{3}}\right|^{2}}

2.2 Preliminaries

Definition 1.

(Mode-nn unfolding[22] For tensor 𝒜I1××IN\mathcal{A}\in\mathbb{R}^{I_{1}\times\cdots\times I_{N}}, its matricization along the nn-th mode is denoted as 𝐀(n)In×I1I2In1In+1IN\mathbf{A}_{(n)}\in\mathbb{R}^{I_{n}\times I_{1}I_{2}\cdots I_{n-1}I_{n+1}\cdots I_{N}}.

Definition 2.

(nn-unfolding[9] Given an NN-way tensor 𝒜I1××IN\mathcal{A}\in\mathbb{R}^{I_{1}\times\cdots\times I_{N}}, the definition of its nn-unfolding is expressed as 𝐀nI1In×In+1IN\mathbf{A}_{\langle n\rangle}\in\mathbb{R}^{I_{1}\cdots I_{n}\times I_{n+1}\cdots I_{N}}.

Definition 3.

(Mode-nn product[22] The mode-nn product of 𝒜I1××IN\mathcal{A}\in\mathbb{R}^{I_{1}\times\cdots\times I_{N}} and matrix 𝐁J×In\mathbf{B}\in\mathbb{R}^{J\times I_{n}} is defined as

𝒳=𝒜×n𝐁I1××In1×J×In+1××IN.\mathcal{X}=\mathcal{A}\times_{n}\mathbf{B}\in\mathbb{R}^{I_{1}\times\cdots\times I_{n-1}\times J\times I_{n+1}\times\cdots\times I_{N}}. (1)
Definition 4.

(Tensor Network Contraction[38] Given a tensor network composed of NN sub-tensors {𝒢(n)}(n=1,,N)\{\mathcal{G}^{(n)}\}(n=1,\cdots,N), then 𝒢\mathcal{G} is denoted as the tensor network contraction result of {𝒢(n)}(n=1,,N)\{\mathcal{G}^{(n)}\}(n=1,\cdots,N), whose general mathematical equation can be written as

𝒢=\displaystyle\mathcal{G}= n=1Nr1n,r2n,R1n,R2n,{d1,d2D1,D2,n=1N𝒢(n)({r1n,r2n,d1,d2})}\displaystyle\sum_{n=1}^{N}\sum_{r_{1}^{n},r_{2}^{n},\cdots}^{R_{1}^{n},R_{2}^{n},\cdots}\{\sum_{d_{1},d_{2}\cdots}^{D_{1},D_{2},\cdots}\prod_{n=1}^{N}\mathcal{G}^{(n)}(\{r_{1}^{n},r_{2}^{n}\cdots,d_{1},d_{2}\cdots\})\} (2)
=\displaystyle= TC({𝒢(n)}n=1N),\displaystyle\operatorname{TC}(\{\mathcal{G}^{(n)}\}_{n=1}^{N}),

where {D1,D2,}\{D_{1},D_{2},\cdots\} are geometrical indexes, each of which is shared normally two tensors and will be contracted, ={R11,R21,,R1N,R2N,}\mathbb{R}=\{R_{1}^{1},R_{2}^{1},\cdots,R_{1}^{N},R_{2}^{N},\cdots\} are open bounds, each of which only belongs to one tensor. After contracting all the geometrical indexes, the 𝒢\mathcal{G} represents a PP-th order tensor with PP the total number of the open indexes \mathbb{R}.

2.3 Related Works

This section reviews some closely related works, including low-rank approximation based multi-view subspace clustering, and multi-view graph clustering.

2.3.1 Low-rank Approximation Based Multi-view Subspace Clustering

Subspace learning based multi-view clustering methods assume that each sample can be represented as a linear combination of all the samples [40].

Chen et al. [7] utilize the Tucker decomposition to obtain the ”clean” representation tensor from view specific matrices. He et al. [12] introduce a tensor-based multi-linear multi-view clustering (MMC) method. Under the consideration of Canonical Polyadic Decomposition (CPD) [19], the MMC is able to explore the higher-order interaction information among the multiple views. Recently, tensor singular value decomposition (t-SVD) based tensor nuclear norm [21, 20] shows satisfying performance in capturing low rank information for 3rd-order tensor. Thus, many t-SVD based works [48, 45, 46] are proposed for better capturing the high-order correlation hidden in 3rd-order tensor from multi-view data. To preserve the consensus across different views, Xie et al. [48] proposed the t-SVD-MSC model, which adopts the t-SVD based tensor nuclear norm to LRMVC. Consider that those self-representation based subspace clustering method will lead to high computation complexity, the essential tensor learning for multi-view spectral clustering (ETLMSC) [45] with Markov chain based spectral clustering and the unified graph and low-rank tensor learning (UGLTL) [46] have been proposed. Zheng et al. [59] develop a bi-linear factorization based clustering method with the orthogonality and low-rank constraint to effectively explore the consensus information of multi-view data. Work in [47] develops the tensorized bipartite graph learning for multi-view clustering (TBGL), which considers the inter-view and intra-view similarities with tensor Schatten pp-norm and 1,2\ell_{1,2}-norm penalty, respectively. Yu et al. [51] propose NTR decomposition and the GNTR decomposition, which yield better performance than the state-of-the-art tensor based methods in clustering.

2.3.2 Multi-view Graph Clustering

Graph learning based multi-view clustering obtains a fusion graph across all views and uses the standard clustering algorithm to produce the final clusters [50]. Given a multi-view dataset, which has VV multiple views {𝐗vCv×N}\left\{\mathbf{X}_{v}\in\mathbb{R}^{C_{v}\times N}\right\} (v=1,,Vv=1,\cdots,V), a general multi-view graph clustering optimization model can be formulated as follows:

min𝐌iv=1Vj=1N12(𝐗i)v(𝐗j)v22mi,j+λj=1Nmi,j2\displaystyle\min\limits_{\mathbf{M}_{i}}\sum\limits_{v=1}^{V}\sum\limits_{j=1}^{N}\frac{1}{2}\left\|(\mathbf{X}_{i})_{v}-(\mathbf{X}_{j})_{v}\right\|_{2}^{2}m_{i,j}+\lambda\sum\limits_{j=1}^{N}m_{i,j}^{2} (3)
s. t. 𝐌iT𝟏=1,0mi,j1,\displaystyle\text{ s.~{}t.~{}}\mathbf{M}_{i}^{\mathrm{T}}\mathbf{1}=1,0\leq{m}_{i,j}\leq 1,

where 𝐌N×N\mathbf{M}\in\mathbb{R}^{N\times N} is the affinity matrix, 𝐌i\mathbf{M}_{i} is the ii-th column of matrix 𝐌\mathbf{M}, (𝐗i)v(\mathbf{X}_{i})_{v} denotes the ii-th column of 𝐗v\mathbf{X}_{v}, and λ\lambda is the trade-off parameter.

By defining the graph Laplacian matrix 𝐋=𝐃(𝐌+𝐌T)/2\mathbf{L}=\mathbf{D}-\left(\mathbf{M}+\mathbf{M}^{\mathrm{T}}\right)/2, where 𝐃\mathbf{D} is a diagonal matrix whose ii-th diagonal entry is j(mi,j+mj,i)/2\sum_{j}\left(m_{i,j}+m_{j,i}\right)/2, (3) can be rewritten as

min𝐌v=1Vtr(𝐗𝐋𝐗T)+λ𝐌F2\displaystyle\min_{\mathbf{M}}\sum\limits_{v=1}^{V}\operatorname{tr}\left(\mathbf{X}\mathbf{L}\mathbf{X}^{\mathrm{T}}\right)+\lambda\|\mathbf{M}\|_{\text{F}}^{2} (4)
s. t. 𝐌T𝟏=𝟏,𝟎𝐌𝟏.\displaystyle\text{ s. t. }\mathbf{M}^{\mathrm{T}}\mathbf{1}=\mathbf{1},\mathbf{0}\leq\mathbf{M}\leq\mathbf{1}.

Based on (4), works in [36, 13, 37, 61] use a parameter-free multiple graph clustering model to automatically learn the optimal weight for graphs. Moreover, Nie et al. introduce a multi-view learning model with adaptive neighbors [35], which simultaneously performs clustering and local structure learning. Wang et al. introduce a general graph-based multi-view clustering (GMC) method [43], which does not only obtain the final clusters directly but also constructs the graph of each view and the fusion graph jointly, thus they can help each other mutually.

3 Tucker-O-Minus Decomposition

In this section, the mathematical formulation of Tucker-O-Minus decomposition with its corresponding solution is given in details.

3.1 Decomposition Formulation

Given a 4th-order tensor 𝒳I1×I2×I3×I4\mathcal{X}\in\mathbb{R}^{I_{1}\times I_{2}\times I_{3}\times I_{4}}, the mathematical formulation of the TOMD is as follows:

𝒳=\displaystyle\mathcal{X}= {r1,,r4R1,,R4(d1,,d6D1,,D6𝒢(1)(d4,r1,d1,d5)𝒢(2)(d1,r2,d2)\displaystyle\{\sum_{r_{1},\cdots,r_{4}}^{R_{1},\cdots,R_{4}}(\sum_{d_{1},\cdots,d_{6}}^{D_{1},\cdots,D_{6}}\mathcal{G}^{(1)}(d_{4},r_{1},d_{1},d_{5})\mathcal{G}^{(2)}(d_{1},r_{2},d_{2}) (5)
𝒢(3)(d2,r3,d3,d6)𝒢(4)(d3,r4,d4)𝐆(5)(d5,d6))}\displaystyle\mathcal{G}^{(3)}(d_{2},r_{3},d_{3},d_{6})\mathcal{G}^{(4)}(d_{3},r_{4},d_{4})\mathbf{G}^{(5)}(d_{5},d_{6}))\}
×1𝐔(1)×2𝐔(2)×3𝐔(3)×4𝐔(4)\displaystyle\times_{1}\mathbf{U}^{(1)}\times_{2}\mathbf{U}^{(2)}\times_{3}\mathbf{U}^{(3)}\times_{4}\mathbf{U}^{(4)}
=\displaystyle= TC({𝒢(n)}n=15)×1𝐔(1)×2𝐔(2)×3𝐔(3)×4𝐔(4),\displaystyle\operatorname{TC}(\{\mathcal{G}^{(n)}\}_{n=1}^{5})\times_{1}\mathbf{U}^{(1)}\times_{2}\mathbf{U}^{(2)}\times_{3}\mathbf{U}^{(3)}\times_{4}\mathbf{U}^{(4)},

where {D1,,D6}\{D_{1},\cdots,D_{6}\} are geometrical indexes, each of which is shared two tensors and should be contracted. 𝐔(i)Ii×Ri(i=1,,4)\mathbf{U}^{(i)}\in\mathbb{R}^{I_{i}\times R_{i}}(i=1,\cdots,4) are factor matrices, 𝒢(n)\mathcal{G}^{(n)}, n=1,,5n=1,\cdots,5, are the factor tensors. Specifically, 𝒢(1)D4×R1×D1×D5\mathcal{G}^{(1)}\in\mathbb{R}^{D_{4}\times R_{1}\times D_{1}\times D_{5}}, 𝒢(2)D1×R2×D2\mathcal{G}^{(2)}\in\mathbb{R}^{D_{1}\times R_{2}\times D_{2}}, 𝒢(3)D2×R3×D3×D6\mathcal{G}^{(3)}\in\mathbb{R}^{D_{2}\times R_{3}\times D_{3}\times D_{6}}, 𝒢(4)D3×R4×D4\mathcal{G}^{(4)}\in\mathbb{R}^{D_{3}\times R_{4}\times D_{4}}, and 𝐆(5)D5×D6\mathbf{G}^{(5)}\in\mathbb{R}^{D_{5}\times D_{6}}. [R1,,R4,D1,,D6]T[R_{1},\cdots,R_{4},D_{1},\cdots,D_{6}]^{\text{T}} is denoted as the rank of TOMD.

From the graphical illustration of TOMD in Fig. 1, it can be clearly seen that the overall architecture of our proposed tensor network is relatively compact and balanced. In this way, based on the Tucker format, O-minus structure is further applied to enhance the relationships among the adjacent and non-adjacent modes of 𝒳\mathcal{X}, simultaneously. In addition, the computational complexity of tensor network contraction can be controlled.

3.2 ALS algorithm

To compute the Tucker-O-Minus decomposition, we employ the alternation least squares (ALS) algorithm, which has been widely used for CPD, Tucker decomposition and TR decomposition [4, 23, 57]. Supposing that we have a 4th-order tensor 𝒳I1×I2×I3×I4\mathcal{X}\in\mathbb{R}^{I_{1}\times I_{2}\times I_{3}\times I_{4}}, the optimization model for approximation a tensor 𝒳\mathcal{X} by the proposed TOMD can be formulated as follows:

min{𝒢(n)}n=15{𝐔(n)}n=14𝒳TC({𝒢(n)}n=15)×1𝐔(1)×2×4𝐔(4)F.\min\limits_{\begin{subarray}{c}\{{\mathcal{G}}^{(n)}\}_{n=1}^{5}\\ \{{\mathbf{U}}^{(n)}\}_{n=1}^{4}\end{subarray}}\|\mathcal{X}-\operatorname{TC}(\{\mathcal{G}^{(n)}\}_{n=1}^{5})\times_{1}\mathbf{U}^{(1)}\times_{2}\cdots\times_{4}\mathbf{U}^{(4)}\|_{\mathrm{F}}. (6)
Theorem 1.

Given a Tucker-O-Minus decomposition 𝒳=TC({𝒢(n)}n=15)×1𝐔(1)×2𝐔(2)×3𝐔(3)×4𝐔(4)\mathcal{X}=\operatorname{TC}(\{\mathcal{G}^{(n)}\}_{n=1}^{5})\times_{1}\mathbf{U}^{(1)}\times_{2}\mathbf{U}^{(2)}\times_{3}\mathbf{U}^{(3)}\times_{4}\mathbf{U}^{(4)}, the corresponding mode-nn unfolding based matrix equivalence of (5) can be written as 𝐗(n)=𝐔(n)(𝐀(n))\mathbf{X}_{(n)}=\mathbf{U}^{(n)}(\mathbf{A}_{(n)}), or 𝐗(n)=𝐔(n)(𝐆(2)(n))(𝐀n)\mathbf{X}_{(n)}=\mathbf{U}^{(n)}(\mathbf{G}_{(2)}^{(n)})(\mathbf{A}^{\neq n}), where 𝐀n=𝐀3n(n=1or3)\mathbf{A}^{\neq n}=\mathbf{A}_{\langle 3\rangle}^{\neq n}~{}(n=1~{}\text{or}~{}3), and 𝐀n=𝐀2n(n=2or4)\mathbf{A}^{\neq n}=\mathbf{A}_{\langle 2\rangle}^{\neq n}~{}(n=2~{}\text{or}~{}4). Moreover, we have 𝒜=𝒢×1𝐔(1)×n1𝐔(n1)×n+1𝐔(n+1)×4𝐔(4)\mathcal{A}=\mathcal{G}\times_{1}\mathbf{U}^{(1)}\cdots\times_{n-1}\mathbf{U}^{(n-1)}\times_{n+1}\mathbf{U}^{(n+1)}\cdots\times_{4}\mathbf{U}^{(4)}, and 𝒜=𝒢×1𝐔(1)×n1𝐔(n1)×n+1𝐔(n+1)×4𝐔(4)\mathcal{A}=\mathcal{G}\times_{1}\mathbf{U}^{(1)}\cdots\times_{n-1}\mathbf{U}^{(n-1)}\times_{n+1}\mathbf{U}^{(n+1)}\cdots\times_{4}\mathbf{U}^{(4)}, respectively, where 𝒢=TC({𝒢(n)}n=15)\mathcal{G}=\operatorname{TC}(\{\mathcal{G}^{(n)}\}_{n=1}^{5}) and 𝒢n=TC({𝒢(n1)}n1n5)\mathcal{G}^{\neq n}=\operatorname{TC}(\{\mathcal{G}^{(n_{1})}\}_{n_{1}\neq n}^{5}). 𝐀2n\mathbf{A}_{\langle 2\rangle}^{\neq n} and 𝐀3n\mathbf{A}_{\langle 3\rangle}^{\neq n} are the nn-unfolding matrices of 𝒜n\mathcal{A}^{\neq n}.

Theorem 2.

Suppose that we have a Tucker-O-Minus decomposition 𝒳=TC({𝒢(n)}n=15)×1𝐔(1)×2𝐔(2)×3𝐔(3)×4𝐔(4)\mathcal{X}=\operatorname{TC}(\{\mathcal{G}^{(n)}\}_{n=1}^{5})\times_{1}\mathbf{U}^{(1)}\times_{2}\mathbf{U}^{(2)}\times_{3}\mathbf{U}^{(3)}\times_{4}\mathbf{U}^{(4)}, and its vectorized form can be formulated as 𝐱=𝐠(5)𝐀~2\mathbf{x}=\mathbf{g}^{(5)}\tilde{\mathbf{A}}_{\langle 2\rangle}, where 𝐱1×I1I2I3I4\mathbf{x}\in\mathbb{R}^{1\times I_{1}I_{2}I_{3}I_{4}}, 𝐠(5)1×D5D6\mathbf{g}^{(5)}\in\mathbb{R}^{1\times D_{5}D_{6}} are the vectorized form of 𝒳\mathcal{X} and 𝐆(5)\mathbf{G}^{(5)}, respectively. Furthermore, 𝒜~\tilde{\mathcal{A}} can be calculated by 𝒜~=𝒢5×3𝐔(1)×4×6𝐔(4)D5×D6×I1××I4\tilde{\mathcal{A}}=\mathcal{G}^{\neq 5}\times_{3}\mathbf{U}^{(1)}\times_{4}\cdots\times_{6}\mathbf{U}^{(4)}\in\mathbb{R}^{D_{5}\times D_{6}\times I_{1}\times\cdots\times I_{4}}, where 𝒢5=TC({𝒢(n)}n=14)D5×D6×R1×R2×R3×R4\mathcal{G}^{\neq 5}=\operatorname{TC}(\{\mathcal{G}^{(n)}\}_{n=1}^{4})\in\mathbb{R}^{D_{5}\times D_{6}\times R_{1}\times R_{2}\times R_{3}\times R_{4}}. 𝐀~2\tilde{\mathbf{A}}_{\langle 2\rangle} is the nn-unfolding of 𝒜~\tilde{\mathcal{A}}.

From Fig. 1, Theorem 1 and Theorem 2, the optimization problem (6) can be changed into a few subproblems by alternating least squares method. In detail, fixing all but one factor, the problem can reduce to three least-squares problems:

  1. 1.

    Subproblem with respect to 𝐔(n)(n=1,,4)\mathbf{U}^{(n)}(n=1,\cdots,4):

    min𝐔(n)𝐗(n)𝐔(n)(𝐀(n))F.\min\limits_{\mathbf{U}^{(n)}}\|\mathbf{X}_{(n)}-\mathbf{U}^{(n)}(\mathbf{A}_{(n)})\|_{\text{F}}. (7)
  2. 2.

    Subproblem with respect to 𝒢(n)(n=1,,4)\mathcal{G}^{(n)}(n=1,\cdots,4):

    min𝐆(2)(n)𝐗(n)𝐔(n)(𝐆(2)(n))(𝐀n)F.\min\limits_{\mathbf{G}_{(2)}^{(n)}}\|\mathbf{X}_{(n)}-\mathbf{U}^{(n)}(\mathbf{G}_{(2)}^{(n)})(\mathbf{A}^{\neq n})\|_{\text{F}}. (8)
  3. 3.

    Subproblem with respect to 𝒢(5)\mathcal{G}^{(5)}:

    min𝐠(5)𝐱(𝐠(5))𝐀~2F.\min\limits_{\mathbf{g}^{(5)}}\|\mathbf{x}-(\mathbf{g}^{(5)})\tilde{\mathbf{A}}_{\langle 2\rangle}\|_{\text{F}}. (9)

Moreover, 9 factors (i.e., {𝒢(n)}n=15,{𝐔(n)}n=14\{\mathcal{G}^{(n)}\}_{n=1}^{5},\{{\mathbf{U}}^{(n)}\}_{n=1}^{4}) can be initialized using the randomization method or truncated SVD with specific values of ranks. In our experiments, we choose the latter way to initialize all factors, and the stopping condition is 𝒳last𝒳newF/𝒳lastFtolals,\|\mathcal{X}_{\text{last}}-\mathcal{X}_{\text{new}}\|_{\text{F}}/\|\mathcal{X}_{\text{last}}\|_{\text{F}}\leq\text{tol}_{\text{als}}, where 𝒳last\mathcal{X}_{\text{last}} is the value at the last iteration, 𝒳new\mathcal{X}_{\text{new}} is constructed from the new iteration, and tolals\text{tol}_{\text{als}} is a predefined threshold.

By denoting Φ(,𝐬)\Phi({\cdot,\mathbf{s}}) as an operator that reshapes a tensor using the size vector 𝐬\mathbf{s}, the overall procedure of TOMD-ALS algorithm can be found in Algorithm 1.

Algorithm 1 TOMD-ALS
  Input: A 4th-order data tensor 𝒳I1×I2×I3×I4\mathcal{X}\in\mathbb{R}^{I_{1}\times I_{2}\times I_{3}\times I_{4}}, predefined TOMD rank, maximum number of iterations (itermax\text{iter}_{\text{max}}), and the threshold for stopping the algorithm tolals=1012\text{tol}_{\text{als}}=10^{-12}, f1=0f_{1}=0, i=0i=0.
  Initialization: Initialize 𝒢(1),,𝒢(4),𝐆(5),𝐔(1),,𝐔(4)\mathcal{G}^{(1)},\cdots,\mathcal{G}^{(4)},\mathbf{G}^{(5)},\mathbf{U}^{(1)},\cdots,\mathbf{U}^{(4)}.
  𝒳new=𝒢×1𝐔(1)×2×4𝐔(4)\mathcal{X}_{\text{new}}=\mathcal{G}\times_{1}\mathbf{U}^{(1)}\times_{2}\cdots\times_{4}\mathbf{U}^{(4)};
  while iitermaxi\leq\text{iter}_{\max} do
     𝒳last=𝒳new\mathcal{X}_{\text{last}}=\mathcal{X}_{\text{new}};
     for  n=1:4n=1:4 do
        𝒜=𝒢×1𝐔(1)×n1𝐔(n1)×n+1𝐔(n+1)×4𝐔(4)\mathcal{A}=\mathcal{G}\times_{1}\mathbf{U}^{(1)}\cdots\times_{n-1}\mathbf{U}^{(n-1)}\times_{n+1}\mathbf{U}^{(n+1)}\cdots\times_{4}\mathbf{U}^{(4)};
        𝐔(n)argmin𝐗(n)𝐔(n)(𝐀(n))F\mathbf{U}^{(n)}\leftarrow\arg\min\|\mathbf{X}_{(n)}-\mathbf{U}^{(n)}(\mathbf{A}_{(n)})\|_{\mathrm{F}};
     end for
     for n=1:4n=1:4 do
        𝒜n=𝒢n×1𝐔(1)×2×n1𝐔(n1)×n+1𝐔(n+1)×n+2×4𝐔(4)\mathcal{A}^{\neq n}=\mathcal{G}^{\neq n}\times_{1}\mathbf{U}^{(1)}\times_{2}\cdots\times_{n-1}\mathbf{U}^{(n-1)}\times_{n+1}\mathbf{U}^{(n+1)}\times_{n+2}\cdots\times_{4}\mathbf{U}^{(4)};
        𝐆(2)(n)argmin𝐗(n)𝐔(n)(𝐆(2)(n))(𝐀n)F\mathbf{G}_{(2)}^{(n)}\leftarrow\arg\min\|\mathbf{X}_{(n)}-\mathbf{U}^{(n)}(\mathbf{G}_{(2)}^{(n)})(\mathbf{A}^{\neq n})\|_{\mathrm{F}};
        𝒢(n)\mathcal{G}^{(n)}=Φ(𝐆(2)(n)\Phi(\mathbf{G}_{(2)}^{(n)}, size(𝒢(n)\operatorname{size}(\mathcal{G}^{(n)}));
     end for
     𝒜~=𝒢5×1𝐔(1)×2×4𝐔(4)\tilde{\mathcal{A}}=\mathcal{G}^{\neq 5}\times_{1}\mathbf{U}^{(1)}\times_{2}\cdots\times_{4}\mathbf{U}^{(4)};
     𝐀~\tilde{\mathbf{A}}=Φ(𝒜~,[,I1I4]\Phi(\tilde{\mathcal{A}},[\sim,I_{1}\cdots I_{4}]), 𝐱\mathbf{x}=Φ(𝒳,[,I1I4])\Phi(\mathcal{X},[\sim,I_{1}\cdots I_{4}]);
     𝐠(5)argmin𝐱(𝐠(5))𝐀~2F\mathbf{g}^{(5)}\leftarrow\operatorname*{argmin}\|\mathbf{x}-(\mathbf{g}^{(5)})\tilde{\mathbf{A}}_{\langle 2\rangle}\|_{\mathrm{F}};
     𝐆(5)\mathbf{G}^{(5)}=Φ(𝐠(5)\Phi(\mathbf{g}^{(5)}, size(𝒢(5)\operatorname{size}(\mathcal{G}^{(5)}));
     𝒳new=𝒢×1𝐔(1)×2×4𝐔(4)\mathcal{X}_{\text{new}}=\mathcal{G}\times_{1}\mathbf{U}^{(1)}\times_{2}\cdots\times_{4}\mathbf{U}^{(4)};
     f1=𝒳last𝒳newF𝒳lastFf_{1}=\frac{\left\|\mathcal{X}_{\text{last}}-\mathcal{X}_{\text{new}}\right\|_{\mathrm{F}}}{\left\|\mathcal{X}_{\text{last}}\right\|_{\mathrm{F}}};
     if f1tolals f_{1}\leq\text{tol}_{\text{als }} then
        break
     end if
     i=i+1i=i+1;
  end while
  Output: 𝒢(1),,𝒢(4),𝐆(5),𝐔(1),,𝐔(4)\mathcal{G}^{(1)},\cdots,\mathcal{G}^{(4)},\mathbf{G}^{(5)},\mathbf{U}^{(1)},\cdots,\mathbf{U}^{(4)}.

4 Tucker-O-Minus decomposition for low rank multi-view clustering model

In this section, we firstly apply TOMD into the low rank tensor optimization model for multi-view clustering. The alternating direction method of multipliers is used to solve it. And its computational complexity is briefly analyzed.

4.1 Optimization Model

To better exploit the low rank data structure in multi-view clustering, we apply the proposed TOMD to the optimization model for clustering. Given a dataset that has VV views {𝐗vCv×N}\left\{\mathbf{X}_{v}\in\mathbb{R}^{C_{v}\times N}\right\} ,v=1,,V,v=1,\cdots,V, where CvC_{v} is the feature dimension of the vv-th feature and NN is the number of data samples in each view. The proposed optimization model based on TOMD for multi-view cluster (TOMD-MVC) can be formulated as follows:

min𝒵,𝐄,𝐌{𝒢(n)}n=15,{𝐔(n)}n=14μv=1Vtr(𝐙vT𝐋𝐙v)+λ𝐌F2+𝐄2,1\displaystyle\min\limits_{\begin{subarray}{c}\mathcal{Z},\mathbf{E},\mathbf{M}\\ \{\mathcal{G}^{(n)}\}_{n=1}^{5},\{\mathbf{U}^{(n)}\}_{n=1}^{4}\end{subarray}}\mu\sum_{v=1}^{V}\operatorname{tr}\left(\mathbf{Z}_{v}^{\mathrm{T}}\mathbf{L}\mathbf{Z}_{v}\right)+\lambda\|\mathbf{M}\|_{\mathrm{F}}^{2}+\|\mathbf{E}\|_{2,1} (10)
s. t.𝐗v=𝐗v𝐙v+𝐄v,v=1,2,,V,\displaystyle~{}\text{s.~{}t.}~{}\mathbf{X}_{v}=\mathbf{X}_{v}\mathbf{Z}_{v}+\mathbf{E}_{v},~{}v=1,2,\cdots,V,
𝐌T𝟏=𝟏,𝟎𝐌𝟏,\displaystyle~{}~{}~{}~{}\mathbf{M}^{\mathrm{T}}\mathbf{1}=\mathbf{1},\mathbf{0}\leq\mathbf{M}\leq\mathbf{1},
𝒵=Φ(TC({𝒢(n)}n=15)×1𝐔(1)×4𝐔(4),[N,N,V]),\displaystyle~{}~{}~{}~{}\mathcal{Z}=\Phi(\operatorname{TC}(\{\mathcal{G}^{(n)}\}_{n=1}^{5})\times_{1}\mathbf{U}^{(1)}\cdots\times_{4}\mathbf{U}^{(4)},[N,N,V]),
𝒵=Ω(𝐙1,𝐙2,,𝐙V),𝐄=[𝐄1;𝐄2;;𝐄V],\displaystyle~{}~{}~{}~{}\mathcal{Z}=\operatorname{\Omega}\left(\mathbf{Z}_{1},\mathbf{Z}_{2},\cdots,\mathbf{Z}_{V}\right),\mathbf{E}=\left[\mathbf{E}_{1};\mathbf{E}_{2};\cdots;\mathbf{E}_{V}\right],

where {𝒢(n)}n=15\{\mathcal{G}^{(n)}\}_{n=1}^{5} are tensor factors, and 𝐋\mathbf{L} is the graph Laplacian matrix of 𝐌\mathbf{M}. 𝐄=[𝐄1;𝐄2;;𝐄V]\mathbf{E}=\left[\mathbf{E}_{1};\mathbf{E}_{2};\cdots;\mathbf{E}_{V}\right] makes all the error matrices {𝐄v}\{\mathbf{E}_{v}\} (v=1,,V)(v=1,\cdots,V) vertically connected along the columns, and function Ω()\operatorname{\Omega}(\cdot) constructs the matrices {𝐙v}(v=1,,V)\{\mathbf{Z}_{v}\}(v=1,\cdots,V) into a 3rd-order tensor. Apart from that, μ\mu and λ\lambda are positive penalty scalars.

After obtaining the affinity matrix 𝐌\mathbf{M}, K-means [11], spectral clustering [34] and other clustering methods can be utilized to yield the final clustering results.

4.2 Solution

The optimization problem (10) can be solved by ADMM. Since variable 𝒵\mathcal{Z} has appeared in the objective function and three constraints simultaneously, we introduce an extra variable 𝒮\mathcal{S} to make 𝒵\mathcal{Z} separable, which result in the equivalence of (10) as follows:

min𝒵,𝐄,𝐌{𝒢(n)}n=15,{𝐔(n)}n=14μv=1Vtr(𝐙vT𝐋𝐙v)+λ𝐌F2+𝐄2,1\displaystyle\min\limits_{\begin{subarray}{c}\mathcal{Z},\mathbf{E},\mathbf{M}\\ \{\mathcal{G}^{(n)}\}_{n=1}^{5},\{\mathbf{U}^{(n)}\}_{n=1}^{4}\end{subarray}}\mu\sum_{v=1}^{V}\operatorname{tr}(\mathbf{Z}_{v}^{\mathrm{T}}\mathbf{L}\mathbf{Z}_{v})+\lambda\|\mathbf{M}\|_{\mathrm{F}}^{2}+\|\mathbf{E}\|_{2,1} (11)
 s. t.𝐗v=𝐗v𝐒v+𝐄v,v=1,2,,V,\displaystyle~{}~{}~{}~{}\text{ s.~{}t.}~{}\mathbf{X}_{v}=\mathbf{X}_{v}\mathbf{S}_{v}+\mathbf{E}_{v},~{}v=1,2,\cdots,V,
𝐌T𝟏=𝟏,𝟎𝐌𝟏,\displaystyle~{}~{}~{}~{}~{}\mathbf{M}^{\mathrm{T}}\mathbf{1}=\mathbf{1},\mathbf{0}\leq\mathbf{M}\leq\mathbf{1},
𝒵=Φ(TC({𝒢(n)}n=15)×1𝐔(1)×4𝐔(4),[N,N,V]),\displaystyle~{}~{}~{}~{}~{}\mathcal{Z}=\Phi(\operatorname{TC}(\{\mathcal{G}^{(n)}\}_{n=1}^{5})\times_{1}\mathbf{U}^{(1)}\cdots\times_{4}\mathbf{U}^{(4)},[N,N,V]),
𝒵=𝒮,\displaystyle~{}~{}~{}~{}~{}\mathcal{Z}=\mathcal{S},
𝒵=Ω(𝐙1,𝐙2,,𝐙V),𝐄=[𝐄1;𝐄2;;𝐄V].\displaystyle~{}~{}~{}~{}~{}\mathcal{Z}=\operatorname{\Omega}\left(\mathbf{Z}_{1},\mathbf{Z}_{2},\cdots,\mathbf{Z}_{V}\right),\mathbf{E}=\left[\mathbf{E}_{1};\mathbf{E}_{2};\cdots;\mathbf{E}_{V}\right].

The augmented Lagrangian function of optimization model (11) is

L(𝒵,𝒮,𝐄,𝐌;𝒲,𝒴)=λ𝐌F2+𝐄2,1+v=1V(μtr(𝐒vT𝐋𝐒v)\displaystyle\operatorname{L}(\mathcal{Z},\mathcal{S},\mathbf{E},\mathbf{M};\mathcal{W},\mathcal{Y})=\lambda\|\mathbf{M}\|_{\mathrm{F}}^{2}+\|\mathbf{E}\|_{2,1}+\sum_{v=1}^{V}(\mu\operatorname{tr}(\mathbf{S}_{v}^{\mathrm{T}}\mathbf{L}\mathbf{S}_{v}) (12)
+𝐖v,𝐗v𝐗v𝐒v𝐄v+τ2𝐗v𝐗v𝐒v𝐄vF2)\displaystyle+\langle\mathbf{W}_{v},\mathbf{X}_{v}-\mathbf{X}_{v}\mathbf{S}_{v}-\mathbf{E}_{v}\rangle+\frac{\tau}{2}\|\mathbf{X}_{v}-\mathbf{X}_{v}\mathbf{S}_{v}-\mathbf{E}_{v}\|_{\mathrm{F}}^{2})
+𝒴,𝒵𝒮+τ2𝒵𝒮F2,\displaystyle+\langle\mathcal{Y},\mathcal{Z}-\mathcal{S}\rangle+\frac{\tau}{2}\|\mathcal{Z}-\mathcal{S}\|_{\mathrm{F}}^{2},

where 𝒲\mathcal{W} and 𝒴\mathcal{Y} are Lagrangian multipliers corresponding to two equations, ,\langle\cdot,\cdot\rangle denotes the inner product, and τ\tau is the penalty parameter. ADMM alternately updates each variable as follows.

Solving 𝒵\mathcal{Z}-subproblem: With the other variables fixed but let 𝒵\mathcal{Z} be the optimization variable, (11) becomes

min𝒵,{𝒢(n)}n=15,{𝐔(n)}n=14𝒵(𝒮t𝒴tτt)F2\displaystyle\min\limits_{\begin{subarray}{c}\mathcal{Z},\{\mathcal{G}^{(n)}\}_{n=1}^{5},\{\mathbf{U}^{(n)}\}_{n=1}^{4}\end{subarray}}\|\mathcal{Z}-(\mathcal{S}^{t}-\frac{\mathcal{Y}^{t}}{\tau^{t}})\|_{\mathrm{F}}^{2} (13)
s. t.𝒵=Φ(TC({𝒢(n)}n=15)×1𝐔(1)×4𝐔(4),[N,N,V]),\displaystyle\text{s. t.}~{}\mathcal{Z}=\Phi(\operatorname{TC}(\{\mathcal{G}^{(n)}\}_{n=1}^{5})\times_{1}\mathbf{U}^{(1)}\cdots\times_{4}\mathbf{U}^{(4)},[N,N,V]),

where tt is the number of iteration. This problem can be solved by TOMD-ALS to obtain the factors 𝒢~(i)(i=1,,5)\tilde{\mathcal{G}}^{(i)}(i=1,\cdots,5) and 𝐔~(i)(i=1,,4)\tilde{\mathbf{U}}^{(i)}(i=1,\cdots,4) of tensor (𝒮t𝒴tτt)(\mathcal{S}^{t}-\frac{\mathcal{Y}^{t}}{\tau^{t}}). Accordingly, the 𝒵\mathcal{Z}-subproblem has following solution:

𝒵t+1=Φ(TC({𝒢(n)~}n=15)×1𝐔~(1)×2×4𝐔~(4),[N,N,V]).\mathcal{Z}^{t+1}=\Phi(\operatorname{TC}(\{\tilde{\mathcal{G}^{(n)}}\}_{n=1}^{5})\times_{1}\tilde{\mathbf{U}}^{(1)}\times_{2}\cdots\times_{4}\tilde{\mathbf{U}}^{(4)},[N,N,V]). (14)

Solving 𝒮\mathcal{S}-subproblem: Similarly, letting 𝒮\mathcal{S} be the only variable, (11) becomes

𝒮t+1\displaystyle\mathcal{S}^{t+1} =argmin𝒮v=1V(τt2𝐗v𝐗v𝐒v𝐄vt+𝐖vtτtF2\displaystyle=\underset{\mathcal{S}}{\operatorname{argmin}}\sum\limits_{v=1}^{V}(\frac{\tau^{t}}{2}\|\mathbf{X}_{v}-\mathbf{X}_{v}\mathbf{S}_{v}-\mathbf{E}_{v}^{t}+\frac{\mathbf{W}_{v}^{t}}{\tau^{t}}\|_{\mathrm{F}}^{2} (15)
+μtr(𝐒vT𝐋𝐒v))+τt2𝒵t+1𝒮+𝒴tτtF2.\displaystyle+\mu\operatorname{tr}(\mathbf{S}_{v}^{\mathrm{T}}\mathbf{L}\mathbf{S}_{v}))+\frac{\tau^{t}}{2}\|\mathcal{Z}^{t+1}-\mathcal{S}+\frac{\mathcal{Y}^{t}}{\tau^{t}}\|_{\mathrm{F}}^{2}.

By separating above problem into VV subproblems, each of which can be solved by:

𝐒vt+1=(τt(𝐭+𝐗vT𝐗v)+2μ𝐋)1(τt𝐙vt+1+𝐘vt+τt𝐗vT(𝐗v𝐄vt+𝐖vtτt)).\begin{split}\mathbf{S}_{v}^{t+1}&=(\tau^{t}(\mathbf{t}+\mathbf{X}_{v}^{\mathrm{T}}\mathbf{X}_{v})+2\mu\mathbf{L})^{-1}(\tau^{t}\mathbf{Z}_{v}^{t+1}\\ &+\mathbf{Y}_{v}^{t}+\tau^{t}\mathbf{X}_{v}^{\mathrm{T}}(\mathbf{X}_{v}-\mathbf{E}_{v}^{t}+\frac{\mathbf{W}_{v}^{t}}{\tau^{t}})).\end{split} (16)

Solving 𝐄\mathbf{E}-subproblem: The optimization subproblem with respect to 𝐄\mathbf{E} is:

𝐄t+1\displaystyle\mathbf{E}^{t+1} =argmin𝐄1τt𝐄2,1\displaystyle=\underset{\mathbf{E}}{\operatorname{argmin}}\frac{1}{\tau^{t}}\|\mathbf{E}\|_{2,1} (17)
+12v=1V𝐄v(𝐗v𝐗v𝐒vt+1+𝐖vtτt)F2.\displaystyle+\frac{1}{2}\sum_{v=1}^{V}\|\mathbf{E}_{v}-(\mathbf{X}_{v}-\mathbf{X}_{v}\mathbf{S}_{v}^{t+1}+\frac{\mathbf{W}_{v}^{t}}{\tau^{t}})\|_{\mathrm{F}}^{2}.

Let 𝐇vt=𝐗v𝐗v𝐒vt+1+𝐖vtτt\mathbf{H}_{v}^{t}=\mathbf{X}_{v}-\mathbf{X}_{v}\mathbf{S}_{v}^{t+1}+\frac{\mathbf{W}_{v}^{t}}{\tau^{t}}, then 𝐇t\mathbf{H}^{t} is constructed by vertically concatenating the matrices {𝐇v},v=1,,V\left\{\mathbf{H}_{v}\right\},v=1,\cdots,V. As it is suggested in [26], (17) has a closed-form solution:

𝐄it+1={𝐇it21τt𝐇it2𝐇it, if 1τt<𝐇it20, otherwise \mathbf{E}_{i}^{t+1}=\left\{\begin{array}[]{ll}\frac{\left\|\mathbf{H}_{i}^{t}\right\|_{2}-\frac{1}{\tau^{t}}}{\left\|\mathbf{H}_{i}^{t}\right\|_{2}}\mathbf{H}_{i}^{t},&\text{ if }\frac{1}{\tau^{t}}<\left\|\mathbf{H}_{i}^{t}\right\|_{2}\\ 0,&\text{ otherwise }\end{array}\right. (18)

where 𝐇it\mathbf{H}_{i}^{t} and 𝐄it+1\mathbf{E}_{i}^{t+1} denote the ii-th column of 𝐇t\mathbf{H}^{t} and 𝐄t+1\mathbf{E}^{t+1}, respectively.

Solving 𝐌\mathbf{M}-subproblem: The optimization subproblem with respect to 𝐌\mathbf{M} is:

min𝐌v=1Vμtr(𝐒vt+1T𝐋𝐒vt+1)+λ𝐌F2\displaystyle\min\limits_{\mathbf{M}}\sum\limits_{v=1}^{V}\mu\operatorname{tr}(\mathbf{S}_{v}^{{t+1}^{\mathrm{T}}}\mathbf{L}\mathbf{S}_{v}^{t+1})+\lambda\|\mathbf{M}\|_{\mathrm{F}}^{2} (19)
s. t.𝐌T𝟏=𝟏,𝟎𝐌𝟏.\displaystyle~{}\text{s. t.}~{}~{}\mathbf{M}^{\mathrm{T}}\mathbf{1}=\mathbf{1},\mathbf{0}\leq\mathbf{M}\leq\mathbf{1}.

According to  (3) and  (4),  (19) can be divided into ii subproblems:

min𝐌ij=1Nv=1Vμ2(𝐒it+1)v(𝐒jt+1)v22mi,j+λj=1Nmi,j2\displaystyle\min\limits_{\mathbf{M}_{i}}\sum\limits_{j=1}^{N}\sum\limits_{v=1}^{V}\frac{\mu}{2}\|(\mathbf{S}_{i}^{t+1})_{v}-(\mathbf{S}_{j}^{t+1})_{v}\|_{2}^{2}m_{i,j}+\lambda\sum\limits_{j=1}^{N}m_{i,j}^{2} (20)
s. t.𝐌iT𝟏=1,𝟎𝐌i𝟏.\displaystyle~{}\text{s. t.}~{}\mathbf{M}_{i}^{\mathrm{T}}\mathbf{1}=1,\mathbf{0}\leq\mathbf{M}_{i}\leq\mathbf{1}.

where 𝐌i\mathbf{M}_{i} is the ii-th column of matrix 𝐌\mathbf{M} and (𝐒i)v(\mathbf{S}_{i})_{v} is the ii-th column of matrix 𝐒v\mathbf{S}_{v}. Following [17, 7] that apply the adaptive neighbor scheme to keep KK largest values constant in 𝐌i\mathbf{M}_{i} and the rest become zero, each subproblem has the solution:

{mijl+1=pi,K+1pijKpi,K+1k=1Kpi,k,jKmijl+1=0,j>K.\left\{\begin{array}[]{l}m_{ij}^{l+1}=\frac{p_{i,K+1}-p_{ij}}{Kp_{i,K+1}-\sum_{k=1}^{K}p_{i,k}},j\leq K\\ m_{ij}^{l+1}=0,j>K.\end{array}\right. (21)

where pi,j=v=1V(𝐬il+1)v(𝐬jl+1)v22p_{i,j}=\sum_{v=1}^{V}\|(\mathbf{s}_{i}^{l+1})_{v}-(\mathbf{s}_{j}^{l+1})_{v}\|_{2}^{2} and 𝐩iN×1\mathbf{p}_{i}\in\mathbb{R}^{N\times 1}.

Updating 𝒲\mathcal{W}, 𝒴\mathcal{Y} and τ\tau: By keeping other variables, i.e., 𝒵\mathcal{Z}, 𝒮\mathcal{S} and 𝐄\mathbf{E} are fixed, the Lagrangian multipliers 𝒲,𝒴\mathcal{W},\mathcal{Y} and the penalty parameter τ\tau are updated as follows:

𝐖vt+1\displaystyle\mathbf{W}_{v}^{t+1} =𝐖vt+τt(𝐗v𝐗v𝐒vt+1𝐄vt+1),\displaystyle=\mathbf{W}_{v}^{t}+\tau^{t}\left(\mathbf{X}_{v}-\mathbf{X}_{v}\mathbf{S}_{v}^{t+1}-\mathbf{E}_{v}^{t+1}\right), (22)
𝒴t+1\displaystyle\mathcal{Y}^{t+1} =𝒴t+τt(𝒵t+1𝒮t+1),\displaystyle=\mathcal{Y}^{t}+\tau^{t}\left(\mathcal{Z}^{t+1}-\mathcal{S}^{t+1}\right),
τt+1\displaystyle\tau^{t+1} =min{βτt,τmax},\displaystyle=\min\left\{\beta\tau^{t},\tau_{max}\right\},

where β>1\beta>1 is utilized to accelerate the convergence, and τmax\tau_{\text{max}} is the predefined maximum value of τ\tau. The affinity matrix 𝐌\mathbf{M}, which will be put into spectral clustering algorithm to yield the final clustering results, can be calculated by 𝐌=1/Vv=1V(|𝐙v|+|𝐙vT|).\mathbf{M}=1/V\sum_{v=1}^{V}(|\mathbf{Z}_{v}|+|\mathbf{Z}_{v}^{\mathrm{T}}|).

Algorithm 2 TOMD for low rank multi-view clustering
  Input: Multi-view data {𝐗v},v=1,,V\{\mathbf{X}_{v}\},v=1,\cdots,V, μ\mu and KK, itermax=150\text{iter}_{\text{max}}=150, and N1,N2,N3,VN_{1},N_{2},N_{3},V.
  Initialization: 𝒮,𝒵,𝐄,𝐌,𝒲,𝒴\mathcal{S},\mathcal{Z},\mathbf{E},\mathbf{M},\mathcal{W},\mathcal{Y} initialized to 𝟎\mathbf{0}, τ=1\tau=1, β=1.5\beta=1.5, tol=107\text{tol}=10^{-7}, iteration t=1t=1.
  while titermaxt\leq\text{iter}_{\text{max}} do
     𝒮~t=Φ((𝒮tΠtτt),[N1,N2,N3,V])\tilde{\mathcal{S}}^{t}=\Phi((\mathcal{S}^{t}-\frac{\Pi^{t}}{\tau^{t}}),[N_{1},N_{2},N_{3},V]);
     Obtain {𝒢~(n)}n=15,{𝐔~(n)}n=14\{\tilde{\mathcal{G}}^{(n)}\}_{n=1}^{5},\{\tilde{\mathbf{U}}^{(n)}\}_{n=1}^{4} by TOMD-ALS and update 𝒵t+1\mathcal{Z}^{t+1} by (14);
     for v=1:Vv=1:V do
        Update 𝐒vt+1\mathbf{S}_{v}^{t+1};
     end for
     Update 𝐄t+1\mathbf{E}^{t+1} and 𝐌t+1\mathbf{M}^{t+1} by (18) and (21), respectively;
     Update 𝒲t+1\mathcal{W}^{t+1}, 𝒴t+1\mathcal{Y}^{t+1} and τt+1\tau^{t+1} by  (22);
     Check convergence conditions:
     max(𝐗v𝐗v𝐒vt+1𝐄vt+1,𝒵t+1𝒮t+1)tol\max(\|\mathbf{X}_{v}-\mathbf{X}_{v}\mathbf{S}_{v}^{t+1}-\mathbf{E}_{v}^{t+1}\|_{\infty},\|\mathcal{Z}^{t+1}-\mathcal{S}^{t+1}\|_{\infty})\leq\text{tol};
     t=t+1;
  end while
  Output: Affinity matrix 𝐌t+1\mathbf{M}^{t+1}.

4.3 Computational Complexity

To analyze the computation complexity of TOMD-MVC, we assume that R=R1==R4=D1==D6R=R_{1}=\cdots=R_{4}=D_{1}=\cdots=D_{6} for convenience of analysis. At each iteration, it takes 𝒪(IN1N2N3VR4)\mathcal{O}\left(IN_{1}N_{2}N_{3}VR^{4}\right) to update the reshaped tensor 𝒵N1×N2×N3×V\mathcal{Z}\in\mathbb{R}^{N_{1}\times N_{2}\times N_{3}\times V}, where II denotes number of iterations of TOMD-ALS. For 𝒮\mathcal{S}-subproblem, we need 𝒪(VN3)\mathcal{O}\left(VN^{3}\right) to compute its closed-form. Updating 𝐄\mathbf{E} and 𝐌\mathbf{M} cost 𝒪(VN2)\mathcal{O}\left(VN^{2}\right) and 𝒪(N2)\mathcal{O}\left(N^{2}\right), respectively. Therefore, the computation in each iteration is 𝒪(V(IN1N2N3R4+N3+N2)+N2)\mathcal{O}\left(V(IN_{1}N_{2}N_{3}R^{4}+N^{3}+N^{2})+N^{2}\right).

In addition, after we get the affinity matrix, we adopt the spectral clustering method to obtain the final memberships of data. Considering the spectral clustering’s computational complexity is 𝒪(N3)\mathcal{O}\left(N^{3}\right) [45], the overall complexity is 𝒪(TV(IN1N2N3R4+N3+N2)+LN2+N3)\mathcal{O}\left(TV(IN_{1}N_{2}N_{3}R^{4}+N^{3}+N^{2})+LN^{2}+N^{3}\right), where TT is the iteration number of TOMD-MVC.

5 Experiments

In this section, image reconstruction experiments have been employed to indicate our proposed tensor network’s advantages in depicting low-rank properties. The clustering performance of our proposed method has also been demonstrated over six multi-view datasets.

5.1 Low-Rank Analysis

To illustrate the superiority of our proposed tensor network, we conduct the image reconstruction experiments on several real word images111http://sipi.usc.edu/database/database.php?volume=misc: ”House”, ”Peppers”, ”Airplane (F-16)”, ”Sailboat on lake”, and ”Airplane”. Since the TOMD has been utilized for 4th-order tensors, the gray forms of these images are further transformed into 16×16×16×1616\times 16\times 16\times 16 tensors.

In this section, the tensors are performed by Tucker-ALS, Tucker and tensor ring decomposition ALS (TuTR-ALS), O-Minus-ALS (Ominus-ALS) and TOMD-ALS algorithms with the same Relative standard error (RSE) [52]. Beyond that, the storage costs of the different decomposition have been calculated and presented in Table 2. In addition to the time needed for pre-processing the images, reconstructing the images, and calculating the storage costs, the running time in our experiments only contains the main body in decomposition, which is also illustrated in Table 2. Considering that the speed of the iterative algorithm primarily affects the running time of those methods, we set the number of iterations for all methods to 500 in our experiment.

As results shown in Table 2, under the same conditions, we can conclude that the running time of TOMD is in the middle, but it has the lowest storage cost and can better exploit the low rank information compared to the other decompositions.

Table 2: Comparison results about Tucker-ALS, TuTR-ALS, Ominus-ALS and TOMD-ALS algorithms based image reconstruction experiments, and the values under the datasets’ name are the predefined RSE.
Dataset Method Time(s) Storage Cost
Tucker-ALS 17.10 2473
House TuTR-ALS 20.21 1872
(0.12) Ominus-ALS 11.34 1545
TOMD-ALS 17.22 1220
Tucker-ALS 80.74 11904
Peppers TuTR-ALS 40.51 4898
(0.12) Ominus-ALS 45.93 4825
TOMD-ALS 47.48 3072
Tucker-ALS 55.66 4608
F-16 TuTR-ALS 29.12 2340
(0.12) Ominus-ALS 12.41 1801
TOMD-ALS 17.22 1484
Tucker-ALS 85.83 10640
Sailboat TuTR-ALS 45.03 5820
(0.12) Ominus-ALS 51.66 6361
TOMD-ALS 47.62 4894
Tucker-ALS 11.30 2849
Airplane TuTR-ALS 30.22 2080
(0.06) Ominus-ALS 12.49 2507
TOMD-ALS 26.81 1516

5.2 Experimental Settings For Multi-view Clustering

5.2.1 Datasets

The performance of TOMD-MVC has been investigated over six multi-view datasets, Yale222http://cvc.cs.yale.edu/cvc/projects/yalefaces/yalefaces.html, MSRCV1333http://research.microsoft.com/en-us/projects/objectclassrecognition/, ExtendYaleB [44], ORL444http://www.uk.research.att.com/facedatabase.html, Reuters555http://lig-membres.imag.fr/grimal/data.html and Handwritten [6], and their statistic information has been summarized in Table 3.

Table 3: Statistics of different datasets.
Datasets Feature(C1,,CvC_{1},\cdots,C_{v}) Samples(N) Views(V) Clusters Objective (N1,N2,N3,VN_{1},N_{2},N_{3},V)
Yale (4096,3304,6750) 165 3 15 Face (165,15,11,3)
MSRCV1 (1302,48,512,100,256,210) 210 6 7 Object (210,15,14,6)
ExtendYaleB (2500,3304,6750) 650 3 10 Face (50,13,650,3)
ORL (4096,3304,6750) 400 3 40 Face (400,20,20,3)
Reuters (4819,4810,4892,4858,4777) 1200 5 6 Text (1200,20,60,5)
Handwritten (240,76,216,47,64,6) 2000 6 10 Digit (200,10,200,60)

5.2.2 Competitors

We compare our TOMD-MVC model with six matrix based methods, GMC [43], GFSC [18], LMSC [55], SFMC [24], EOMSC-CA [29], COMSC [28], and six multi-view clustering methods, LTMSC [53], t-SVD-MSC [48], ETLMSC [45], UGLTL [46], HLR-M2VC [49], TBGL [47], to verify its superiority.

5.2.3 Evaluation Metrics

In our experiments, six commonly used metrics are applied to quantitatively evaluate the clustering performance, namely F-score, precision (P), recall (R), normalized mutual information (NMI), adjusted rand index (AR), and accuracy (ACC). In addition, the higher values of these metrics indicate better clustering performance. Further detailed information please refer to [25, 56, 14, 39, 60].

5.3 Clustering Performance Comparison

5.3.1 Compared with state-of-the-art methods

All detailed clustering results on six data sets are presented in Table 4-9. Since the different initializations may lead to different results, we run 10 trials for each experiment and present their average performance with standard deviations, i.e., mean(standard deviation). The best results are highlighted in boldface and the second-best results are underlined¯\underline{\text{underlined}}.

LTMSC, t-SVD-MSC, ETLMSC, UGLTL, HLR-M2VC, and TBGL are all tensor-based methods, whose clustering performance is almost higher than matrix based methods. Therefore, employing the tensor based low rank constraints on self-representation tensor is a better way to capture the higher-order correlations. Nevertheless, our proposed TOMD-MVC can still achieve the best results among these tensor-based methods under all six evaluation metrics. Specifically, the results of our proposed model on Yale, ORL and Reuters data sets, are 1.5%1.5\%, 1.13%1.13\%, 6.57%6.57\% higher than the second-best method, respectively. Especially on the ExtendYaleB dataset, the improvement of TOMD-MVC is at least 30%30\% higher than the second-best performance of t-SVD-MSC. Based on this, we can demonstrate the superiority of our proposed tensor network in low rank multi-view clustering.

Table 4: Clustering results on Yale dataset. For TOMD-MVC, we set K=10K=10 and μ=1\mu=1.
Method F-score Precision Recall NMI AR ACC
GMC 0.446(0.000) 0.378(0.000) 0.544(0.000) 0.668(0.000) 0.403(0.000) 0.618(0.000)
GFSC 0.433(0.028) 0.384(0.040) 0.497(0.016) 0.647(0.015) 0.391(0.032) 0.621(0.029)
LMSC 0.519(0.003) 0.475(0.006) 0.572(0.000) 0.717(0.002) 0.484(0.004) 0.679(0.007)
SFMC 0.480(0.000) 0.438(0.000) 0.532(0.000) 0.663(0.000) 0.443(0.000) 0.618(0.000)
EOMSC-CA 0.469(0.000) 0.418(0.000) 0.533(0.000) 0.654(0.000) 0.430(0.000) 0.648(0.000)
LTMSC 0.620(0.009) 0.599(0.011) 0.643(0.006) 0.764(0.006) 0.594(0.009) 0.736(0.004)
t-SVD-MSC 0.902(0.066) 0.891(0.075) 0.915(0.058) 0.946(0.036) 0.896(0.071) 0.934(0.053)
ETLMSC 0.542(0.055) 0.509(0.053) 0.580(0.061) 0.706(0.043) 0.510(0.059) 0.654(0.044)
UGLTL 0.867(0.047) 0.852(0.057) 0.883(0.039) 0.931(0.023) 0.858(0.050) 0.912(0.041)
HLR-M2VC 0.695(0.010) 0.673(0.011) 0.718(0.010) 0.817(0.006) 0.674(0.011) 0.772(0.012)
COMSC 0.627(0.000) 0.607(0.000) 0.953(0.000) 0.762(0.000) 0.602(0.000) 0.739(0.000)
TBGL 0.587(0.000) 0.556(0.000) 0.623(0.000) 0.739(0.000) 0.559(0.000) 0.703(0.000)
TOMD-MVC 0.916(0.013) 0.914(0.014) 0.918(0.013) 0.949(0.008) 0.910(0.014) 0.959(0.006)
Table 5: Clustering results on MSRCV1 dataset. For TOMD-MVC, we set K=5K=5 and μ=50\mu=50.
Method F-score Precision Recall NMI AR ACC
GMC 0.824(0.000) 0.812(0.000) 0.837(0.000) 0.846(0.000) 0.795(0.000) 0.910(0.000)
GFSC 0.609(0.050) 0.581(0.065) 0.642(0.031) 0.661(0.042) 0.542(0.062) 0.730(0.061)
LMSC 0.656(0.081) 0.646(0.081) 0.667(0.080) 0.677(0.065) 0.600(0.094) 0.783(0.086)
SFMC 0.763(0.000) 0.701(0.000) 0.836(0.000) 0.790(0.000) 0.720(0.000) 0.838(0.000)
EOMSC-CA 0.813(0.000) 0.787(0.000) 0.840(0.000) 0.837(0.000) 0.781(0.000) 0.876(0.000)
LTMSC 0.727(0.001) 0.714(0.001) 0.742(0.000) 0.750(0.000) 0.682(0.001) 0.829(0.000)
t-SVD-MSC 0.962(0.000) 0.961(0.000) 0.963(0.000) 0.960(0.000) 0.955(0.000) 0.981(0.000)
ETLMSC 0.934(0.079) 0.924(0.099) 0.946(0.056) 0.946(0.055) 0.923(0.092) 0.950(0.077)
UGLTL 1.000(0.000) 1.000(0.000) 1.000(0.000) 1.000(0.000) 1.000(0.000) 1.000(0.000)
HLR-M2VC 0.990(0.000) 0.990(0.000) 0.990(0.000) 0.989(0.000) 0.989(0.000) 0.995(0.000)
COMSC 0.861(0.000) 0.856(0.000) 0.961(0.000) 0.863(0.000) 0.838(0.000) 0.929(0.000)
TBGL 1.000(0.000) 1.000(0.000) 1.000(0.000) 1.000(0.000) 1.000(0.000) 1.000(0.000)
TOMD-MVC 1.000(0.000) 1.000(0.000) 1.000(0.000) 1.000(0.000) 1.000(0.000) 1.000(0.000)
Table 6: Clustering results on ExtendYaleB dataset. For TOMD-MVC, we set K=15K=15 and μ=50\mu=50.
Method F-score Precision Recall NMI AR ACC
GMC 0.256(0.000) 0.201(0.000) 0.351(0.000) 0.405(0.000) 0.149(0.000) 0.386(0.000)
GFSC 0.204(0.006) 0.161(0.004) 0.282(0.024) 0.327(0.023) 0.090(0.005) 0.348(0.025)
LMSC 0.352(0.001) 0.309(0.001) 0.410(0.001) 0.513(0.001) 0.270(0.001) 0.530(0.001)
SFMC 0.290(0.000) 0.197(0.000) 0.551(0.000) 0.494(0.000) 0.170(0.000) 0.509(0.000)
EOMSC-CA 0.202(0.000) 0.169(0.000) 0.252(0.000) 0.250(0.000) 0.095(0.000) 0.289(0.000)
LTMSC 0.323(0.005) 0.296(0.006) 0.355(0.004) 0.456(0.004) 0.242(0.006) 0.455(0.002)
t-SVD-MSC 0.483(0.005) 0.456(0.007) 0.513(0.004) 0.618(0.003) 0.422(0.006) 0.577(0.001)
ETLMSC 0.262(0.017) 0.257(0.017) 0.590(0.008) 0.307(0.021) 0.179(0.019) 0.325(0.011)
UGLTL 0.406(0.038) 0.395(0.036) 0.417(0.041) 0.503(0.044) 0.339(0.042) 0.463(0.037)
HLR-M2VC 0.575(0.000) 0.553(0.000) 0.599(0.000) 0.706(0.000) 0.527(0.000) 0.673(0.000)
COMSC 0.453(0.000) 0.414(0.000) 0.881(0.000) 0.596(0.000) 0.387(0.000) 0.616(0.000)
TBGL 0.230(0.000) 0.156(0.000) 0.437(0.000) 0.382(0.000) 0.100(0.000) 0.403(0.000)
TOMD-MVC 0.981(0.001) 0.981(0.001) 0.981(0.001) 0.982(0.001) 0.979(0.001) 0.990(0.001)
Table 7: Clustering results on ORL dataset. For TOMD-MVC, we set K=10K=10 and μ=30\mu=30.
Method F-score Precision Recall NMI AR ACC
GMC 0.368(0.000) 0.239(0.000) 0.805(0.000) 0.861(0.000) 0.345(0.000) 0.660(0.000)
GFSC 0.501(0.053) 0.413(0.063) 0.644(0.035) 0.828(0.019) 0.487(0.055) 0.636(0.046)
LMSC 0.776(0.016) 0.717(0.024) 0.847(0.013) 0.932(0.005) 0.771(0.017) 0.823(0.016)
SFMC 0.702(0.000) 0.586(0.000) 0.876(0.000) 0.908(0.000) 0.694(0.000) 0.775(0.000)
EOMSC-CA 0.461(0.000) 0.365(0.000) 0.627(0.000) 0.791(0.000) 0.446(0.000) 0.613(0.000)
LTMSC 0.766(0.025) 0.725(0.028) 0.813(0.023) 0.920(0.010) 0.760(0.026) 0.817(0.021)
t-SVD-MSC 0.987(0.017) 0.979(0.027) 0.996(0.006) 0.997(0.003) 0.987(0.017) 0.987(0.017)
ETLMSC 0.928(0.027) 0.894(0.036) 0.966(0.019) 0.982(0.009) 0.927(0.028) 0.929(0.025)
UGLTL 0.948(0.017) 0.932(0.022) 0.965(0.013) 0.985(0.005) 0.947(0.017) 0.954(0.016)
HLR-M2VC 0.991(0.015) 0.984(0.026) 0.997(0.005) 0.998(0.003) 0.990(0.016) 0.991(0.015)
COMSC 0.770(0.000) 0.730(0.000) 0.989(0.000) 0.913(0.000) 0.764(0.000) 0.835(0.000)
TBGL 0.611(0.000) 0.470(0.000) 0.871(0.000) 0.895(0.000) 0.599(0.000) 0.848(0.000)
TOMD-MVC 1.000(0.000) 1.000(0.000) 1.000(0.000) 1.000(0.000) 1.000(0.000) 1.000(0.000)
Table 8: Clustering results on Reuters dataset. For TOMD-MVC, we set K=20K=20 and μ=50\mu=50.
Method F-score Precision Recall NMI AR ACC
GMC 0.283(0.000) 0.167(0.000) 0.927(0.000) 0.075(0.000) 0.003(0.000) 0.187(0.000)
GFSC 0.328(0.010) 0.254(0.020) 0.475(0.055) 0.226(0.024) 0.143(0.026) 0.375(0.045)
LMSC 0.407(0.000) 0.374(0.000) 0.447(0.000) 0.342(0.000) 0.276(0.000) 0.562(0.000)
SFMC 0.284(0.000) 0.166(0.000) 0.966(0.000) 0.019(0.000) 0.001(0.000) 0.182(0.000)
EOMSC-CA 0.312(0.000) 0.291(0.000) 0.335(0.000) 0.223(0.000) 0.163(0.000) 0.427(0.000)
LTMSC 0.304(0.000) 0.253(0.000) 0.380(0.000) 0.196(0.001) 0.130(0.000) 0.382(0.001)
t-SVD-MSC 0.904(0.000) 0.902(0.000) 0.907(0.000) 0.885(0.000) 0.885(0.000) 0.950(0.000)
ETLMSC 0.898(0.114) 0.891(0.128) 0.907(0.098) 0.896(0.087) 0.877(0.138) 0.920(0.118)
UGLTL 0.934(0.000) 0.933(0.000) 0.934(0.000) 0.914(0.000) 0.920(0.000) 0.966(0.000)
HLR-M2VC 0.714(0.001) 0.704(0.001) 0.724(0.001) 0.708(0.001) 0.656(0.001) 0.831(0.001)
COMSC 0.407(0.000) 0.349(0.000) 0.764(0.000) 0.330(0.000) 0.264(0.000) 0.547(0.000)
TBGL 0.284(0.000) 0.167(0.000) 0.951(0.000) 0.029(0.000) 0.002(0.000) 0.179(0.000)
TOMD-MVC 0.964(0.013) 0.963(0.014) 0.966(0.012) 0.957(0.016) 0.957(0.016) 0.981(0.007)
Table 9: Clustering results on Handwritten dataset. For TOMD-MVC, we set K=20K=20 and μ=40\mu=40.
Method F-score Precision Recall NMI AR ACC
GMC 0.854(0.000) 0.793(0.000) 0.925(0.000) 0.910(0.000) 0.836(0.000) 0.857(0.000)
GFSC 0.630(0.049) 0.573(0.052) 0.700(0.054) 0.715(0.026) 0.584(0.056) 0.714(0.053)
LMSC 0.818(0.001) 0.815(0.001) 0.820(0.001) 0.816(0.001) 0.797(0.001) 0.901(0.000)
SFMC 0.868(0.000) 0.820(0.000) 0.920(0.000) 0.892(0.000) 0.852(0.000) 0.881(0.000)
EOMSC-CA 0.710(0.000) 0.613(0.000) 0.842(0.000) 0.775(0.000) 0.672(0.000) 0.725(0.000)
LTMSC 0.818(0.016) 0.815(0.016) 0.821(0.015) 0.833(0.010) 0.797(0.017) 0.896(0.012)
t-SVD-MSC 0.999(0.000) 0.999(0.000) 0.999(0.000) 0.999(0.000) 0.999(0.000) 1.000(0.000)
ETLMSC 0.919(0.071) 0.880(0.107) 0.965(0.027) 0.964(0.030) 0.909(0.080) 0.902(0.089)
UGLTL 1.000(0.000) 1.000(0.000) 1.000(0.000) 1.000(0.000) 1.000(0.000) 1.000(0.000)
HLR-M2VC 1.000(0.000) 1.000(0.000) 1.000(0.000) 1.000(0.000) 1.000(0.000) 1.000(0.000)
COMSC 0.885(0.000) 0.883(0.000) 0.977(0.000) 0.881(0.000) 0.872(0.000) 0.940(0.000)
TBGL 0.849(0.000) 0.792(0.000) 0.914(0.000) 0.894(0.000) 0.831(0.000) 0.843(0.000)
TOMD-MVC 1.000(0.000) 1.000(0.000) 1.000(0.000) 1.000(0.000) 1.000(0.000) 1.000(0.000)

5.3.2 Compared with Tucker decomposition based methods

In this section, we implement a number of experiments to evaluate the influence of applying higher-order form and O-minus structure on self-representation tensor 𝒵\mathcal{Z} in multi-view clustering. The primary procedure of clustering algorithm are same except for updating tensor 𝒵\mathcal{Z}. Specifically, with 4-Dimensional Tucker decomposition based multi-view clustering model (4DT-MVC), we reshape 𝒵\mathcal{Z} into a 4th-order tensor and employs Tucker decomposition to capture its low-rank information while solving the problem with respect to 𝒵\mathcal{Z}. As for the 3-Dimensional Tucker decomposition based MVC method (3DT-MVC), the representation tensor maintained 3rd-order. The average results after 10 runs of each experiment with six metrics are all presented in Table 10.

As results show, the improvements of 4DT-MVC are around 29.50%29.50\%, 10.57%10.57\%, 6.87%6.87\%, 8.09%8.09\%, 2.67%2.67\% about the average of six metrics over 3DT-MVC on Yale, MSRCV1, ExtendYaleB, ORL and Handwritten datasets, respectively. Especially for the Reuters data set, 4DT-MVC is more than twice as much as 3DT-MVC. The main reason may be that more low-rank information can be discovered in high-dimensional space. In addition, TOMD has a further 7.16%7.16\%, 2.76%2.76\%, 0.56%0.56\%, 0.42%0.42\%, 5.21%5.21\%, 2.92%2.92\% development over 4DT-MVC on the same datasets, demonstrating the superiority of our proposed tensor network TOMD.

Table 10: Clustering results on six datasets obtained by different tensor decomposition constraints.
Datasets Method F-score Precision Recall NMI AR ACC
3DT-MVC 0.635 0.611 0.660 0.781 0.610 0.738
Yale 4DT-MVC 0.842 0.825 0.861 0.913 0.832 0.927
TOMD-MVC 0.916 0.914 0.918 0.949 0.910 0.959
3DT-MVC 0.879 0.866 0.892 0.869 0.840 0.943
MSRCV1 4DT-MVC 0.972 0.971 0.973 0.971 0.967 0.986
TOMD-MVC 1.000 1.000 1.000 1.000 1.000 1.000
3DT-MVC 0.909 0.908 0.911 0.905 0.899 0.954
ExtendYaleB 4DT-MVC 0.975 0.974 0.975 0.976 0.973 0.988
TOMD-MVC 0.981 0.981 0.981 0.982 0.979 0.990
3DT-MVC 0.908 0.888 0.928 0.970 0.905 0.933
ORL 4DT-MVC 0.995 0.994 0.995 0.998 0.995 0.998
TOMD-MVC 1.000 1.000 1.000 1.000 1.000 1.000
3DT-MVC 0.445 0.377 0.543 0.441 0.310 0.584
Reuters 4DT-MVC 0.922 0.921 0.924 0.917 0.907 0.957
TOMD-MVC 0.964 0.963 0.966 0.957 0.957 0.981
3DT-MVC 0.945 0.944 0.945 0.934 0.939 0.972
Handwritten 4DT-MVC 0.971 0.971 0.971 0.963 0.968 0.986
TOMD-MVC 1.000 1.000 1.000 1.000 1.000 1.000

5.4 Model Discussion

5.4.1 Representation Visualization

Fig. 2 visualizes the affinity matrices learned by different clustering methods. Due to the limitations of space, we only present the results of LMSC, t-SVD-MSC and TOMD-MVC on ORL dataset. Compared with the other three methods, our proposed TOMD-MVC is much better as its visualization matrix gives a more clear block-diagonal structure on ORL dataset.

Refer to caption
(a) LMSC
Refer to caption
(b) t-SVD-MSC
Refer to caption
(c) TOMD-MVC
Figure 2: Comparison of affinity matrix on ORL dataset by LMSC, t-SVD-MSC and TOMD-MVC.

5.4.2 Convergence

In Fig. 3, we respectively plot the convergence curves of 3DT-MVC, 4DT-MVC and TOMD-MVC methods, where the X-axis is the number of iterations, and Y-axis is the value of Reconstruction Error=(1/V)v=1V𝐗v𝐗v𝐒v𝐄v\text{Reconstruction Error}=(1/V)\sum_{v=1}^{V}\left\|\mathbf{X}_{v}-\mathbf{X}_{v}\mathbf{S}_{v}-\mathbf{E}_{v}\right\|_{\infty} and Match Error=(1/V)v=1V𝐙v𝐒v\text{Match Error}=(1/V)\sum_{v=1}^{V}\left\|\mathbf{Z}_{v}-\mathbf{S}_{v}\right\|_{\infty}. As the results show, the 3DT-MVC convergences faster on the MSRCV1 dataset. However, its clustering performance does not perform well compared to the 4DT-MVC and TOMD-MVC. The convergence speed of the TOMD algorithm is similar to that of 4DT-MVC as presented in Fig. 3. Nonetheless, our proposed method has a further 2.76%2.76\% development over 4DT-MVC on MSRCV1 as illustrated in Table 10. Empirically, the number of optimization iteration is usually located within the range of (100,150)(100,150).

Refer to caption
(a) 3DT-MVC
Refer to caption
(b) 4DT-MVC
Refer to caption
(c) TOMD-MVC
Figure 3: The convergence curves on MSRCV1 dataset.

5.4.3 Parameter Sensitivity Analysis

As suggested in [7], the parameter λ\lambda can be determined by the number of adaptive neighbors KK. Therefore, there are three free parameters in our model: balanced parameter μ\mu, the number of adaptive neighbors KK and tensor ranks (R1,,R4,D1,,D6)(R_{1},\cdots,R_{4},D_{1},\cdots,D_{6}). Due to the limitations of the paper, we only show the results of ACC and NMI on Yale dataset by the combination of different μ\mu and KK in Fig. 4. We can see that the performance of TOMD-MVC is relatively stable when μ=[5,50]\mu=[5,50] and K=[5,15]K=[5,15] on Yale dataset. Besides, the evaluation results on Yale dataset have been presented in Table 11 with different rank tuning. The results show that TOMD-MVC can achieve promising performance while the pre-defined ranks are set as (30,15,11,V,4,4,4,4,4,4)(30,15,11,V,4,4,4,4,4,4), where VV is the number of views.

Refer to caption
Figure 4: Parameters tuning (μ\mu and KK) in terms of ACC and NMI on MSRCV1.
Table 11: Clustering results of TOMD-MVC with different ranks on Yale dataset.
(R1,,R4,D1,,D6)(R_{1},\cdots,R_{4},D_{1},\cdots,D_{6}) F-score Precision Recall NMI AR ACC
(10,5,5,V,2,2,2,2,2,2)(10,5,5,V,2,2,2,2,2,2) 0.626(0.018) 0.595(0.016) 0.660(0.021) 0.812(0.010) 0.600(0.020) 0.723(0.029)
(10,10,5,V,2,2,2,2,2,2)(10,10,5,V,2,2,2,2,2,2) 0.624(0.019) 0.594(0.016) 0.657(0.023) 0.808(0.012) 0.598(0.020) 0.730(0.020)
(10,5,10,V,2,2,2,2,2,2)(10,5,10,V,2,2,2,2,2,2) 0.623(0.008) 0.583(0.007) 0.670(0.011) 0.813(0.005) 0.597(0.009) 0.707(0.020)
(20,5,5,V,2,2,2,2,2,2)(20,5,5,V,2,2,2,2,2,2) 0.630(0.017) 0.598(0.015) 0.666(0.020) 0.813(0.010) 0.605(0.019) 0.724(0.034)
(10,5,5,V,4,4,4,4,4,4)(10,5,5,V,4,4,4,4,4,4) 0.490(0.032) 0.464(0.035) 0.520(0.030) 0.711(0.021) 0.455(0.035) 0.652(0.034)
(20,10,10,V,4,4,4,4,4,4)(20,10,10,V,4,4,4,4,4,4) 0.844(0.022) 0.835(0.022) 0.852(0.022) 0.910(0.014) 0.833(0.023) 0.921(0.012)
(20,15,10,V,4,4,4,4,4,4)(20,15,10,V,4,4,4,4,4,4) 0.843(0.012) 0.833(0.015) 0.854(0.010) 0.909(0.006) 0.833(0.013) 0.922(0.006)
(20,10,11,V,4,4,4,4,4,4)(20,10,11,V,4,4,4,4,4,4) 0.887(0.015) 0.884(0.015) 0.891(0.015) 0.932(0.010) 0.880(0.016) 0.944(0.008)
(30,10,10,V,4,4,4,4,4,4)(30,10,10,V,4,4,4,4,4,4) 0.858(0.012) 0.849(0.014) 0.866(0.010) 0.918(0.006) 0.848(0.013) 0.929(0.006)
(30,15,11,V,6,6,6,6,6,6)(30,15,11,V,6,6,6,6,6,6) 0.637(0.007) 0.613(0.010) 0.663(0.005) 0.782(0.004) 0.613(0.008) 0.738(0.006)
(30,15,11,V,4,4,4,4,4,4)(30,15,11,V,4,4,4,4,4,4) 0.916(0.013) 0.914(0.014) 0.918(0.013) 0.949(0.008) 0.910(0.014) 0.959(0.006)
(40,15,11,V,4,4,4,4,4,4)(40,15,11,V,4,4,4,4,4,4) 0.896(0.008) 0.892(0.009) 0.899(0.008) 0.937(0.005) 0.889(0.009) 0.948(0.004)

6 Conclusion

In this paper, we propose a novel tensor network called TOMD to considerably capture the low-rank property of representation tensor and fully explore the latent correlations across different views in clustering. In addition, we drive a detailed algorithm to solve the optimization problem. Accordingly, TOMD-MVC has been developed and its superiority performance has been demonstrated by extensive experimental results on six real-world datasets.

Acknowledgment

This work was supported by the National Natural Science Foundation of China (NSFC) under Grant 62171088, Grant 6220106011 and Grant U19A2052.

References

  • [1] Tülay Adali, Furkan Kantar, Mohammad Abu Baker Siddique Akhonda, Stephen Strother, Vince D Calhoun, and Evrim Acar. Reproducibility in matrix and tensor decompositions: Focus on model match, interpretability, and uniqueness. IEEE Signal Processing Magazine, 39(4):8–24, 2022.
  • [2] Johann A Bengua, Ho N Phien, Hoang Duong Tuan, and Minh N Do. Efficient tensor completion for color image and video recovery: Low-rank tensor train. IEEE Transactions on Image Processing, 26(5):2466–2479, 2017.
  • [3] M Billó, F Fucito, GP Korchemsky, A Lerda, and JF Morales. Two-point correlators in non-conformal 𝒩\mathcal{N}= 2 gauge theories. Journal of High Energy Physics, 2019(5):1–56, 2019.
  • [4] J Douglas Carroll and Jih-Jie Chang. Analysis of individual differences in multidimensional scaling via an n-way generalization of “eckart-young” decomposition. Psychometrika, 35(3):283–319, 1970.
  • [5] Hongyang Chen, Fauzia Ahmad, Sergiy Vorobyov, and Fatih Porikli. Tensor decompositions in wireless communications and mimo radar. IEEE Journal of Selected Topics in Signal Processing, 15(3):438–453, 2021.
  • [6] Peng Chen, Liang Liu, Zhengrui Ma, and Zhao Kang. Smoothed multi-view subspace clustering. In International Conference on Neural Computing for Advanced Applications, pages 128–140. Springer, 2021.
  • [7] Yongyong Chen, Xiaolin Xiao, Chong Peng, Guangming Lu, and Yicong Zhou. Low-rank tensor graph learning for multi-view subspace clustering. IEEE Transactions on Circuits and Systems for Video Technology, 2021.
  • [8] Zefeng Chen, Guoxu Zhou, and Qibin Zhao. Hierarchical factorization strategy for high-order tensor and application for data completion. IEEE Signal Processing Letters, 2021.
  • [9] Andrzej Cichocki, Namgil Lee, Ivan V Oseledets, A-H Phan, Qibin Zhao, and D Mandic. Low-rank tensor networks for dimensionality reduction and large-scale optimization problems: Perspectives and challenges part 1. Foundations and Trends in Machine Learning, 9(4-5):249–429, 2016.
  • [10] Lanlan Feng, Ce Zhu, and Yipeng Liu. Multi-mode tensor singular value decomposition for low-rank image recovery. In International Conference on Image and Graphics, pages 238–249. Springer, 2021.
  • [11] John A Hartigan and Manchek A Wong. Algorithm as 136: A k-means clustering algorithm. Journal of The Royal Statistical Society. Series C (Applied Statistics), 28(1):100–108, 1979.
  • [12] Lifang He, Chun-Ta Lu, Yong Chen, Jiawei Zhang, Linlin Shen, S Yu Philip, and Fei Wang. A self-organizing tensor architecture for multi-view clustering. In 2018 IEEE International Conference on Data Mining (ICDM), pages 1007–1012. IEEE, 2018.
  • [13] Chenping Hou, Feiping Nie, Hong Tao, and Dongyun Yi. Multi-view unsupervised feature selection with adaptive similarity and view weight. IEEE Transactions on Knowledge and Data Engineering, 29(9):1998–2011, 2017.
  • [14] Dong Huang, Chang-Dong Wang, Jian-Sheng Wu, Jian-Huang Lai, and Chee-Keong Kwoh. Ultra-scalable spectral clustering and ensemble clustering. IEEE Transactions on Knowledge and Data Engineering, 32(6):1212–1226, 2019.
  • [15] Yuwang Ji, Qiang Wang, Xuan Li, and Jie Liu. A survey on tensor techniques and applications in machine learning. IEEE Access, 7:162950–162990, 2019.
  • [16] Rohit Kumar Kaliyar, Anurag Goswami, and Pratik Narang. Deepfake: improving fake news detection using tensor decomposition-based deep neural network. The Journal of Supercomputing, 77(2):1015–1037, 2021.
  • [17] Zhao Kang, Haiqi Pan, Steven CH Hoi, and Zenglin Xu. Robust graph learning from noisy data. IEEE Transactions on Cybernetics, 50(5):1833–1843, 2019.
  • [18] Zhao Kang, Guoxin Shi, Shudong Huang, Wenyu Chen, Xiaorong Pu, Joey Tianyi Zhou, and Zenglin Xu. Multi-graph fusion for multi-view spectral clustering. Knowledge-Based Systems, 189:105102, 2020.
  • [19] Henk AL Kiers. Towards a standardized notation and terminology in multiway analysis. Journal of Chemometrics: A Journal of the Chemometrics Society, 14(3):105–122, 2000.
  • [20] Misha E Kilmer and Carla D Martin. Factorization strategies for third-order tensors. Linear Algebra and its Applications, 435(3):641–658, 2011.
  • [21] Misha E Kilmer, Carla D Martin, and Lisa Perrone. A third-order generalization of the matrix svd as a product of third-order tensors. Tufts University, Department of Computer Science, Tech. Rep. TR-2008-4, 2008.
  • [22] Tamara G Kolda and Brett W Bader. Tensor decompositions and applications. SIAM Review, 51(3):455–500, 2009.
  • [23] Pieter M Kroonenberg and Jan De Leeuw. Principal component analysis of three-mode data by means of alternating least squares algorithms. Psychometrika, 45(1):69–97, 1980.
  • [24] Xuelong Li, Han Zhang, Rong Wang, and Feiping Nie. Multiview clustering: A scalable and parameter-free bipartite graph fusion method. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(1):330–344, 2020.
  • [25] Kun-Yu Lin, Ling Huang, Chang-Dong Wang, and Hong-Yang Chao. Multi-view proximity learning for clustering. In International Conference on Database Systems for Advanced Applications, pages 407–423. Springer, 2018.
  • [26] Guangcan Liu, Zhouchen Lin, Shuicheng Yan, Ju Sun, Yong Yu, and Yi Ma. Robust recovery of subspace structures by low-rank representation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(1):171–184, 2012.
  • [27] Jiani Liu, Ce Zhu, and Yipeng Liu. Smooth compact tensor ring regression. IEEE Transactions on Knowledge and Data Engineering, 2020.
  • [28] Jiyuan Liu, Xinwang Liu, Yuexiang Yang, Xifeng Guo, Marius Kloft, and Liangzhong He. Multiview subspace clustering via co-training robust data representation. IEEE Transactions on Neural Networks and Learning Systems, 2021.
  • [29] Suyuan Liu, Siwei Wang, Pei Zhang, Kai Xu, Xinwang Liu, Changwang Zhang, and Feng Gao. Efficient one-pass multi-view subspace clustering with consensus anchors. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pages 7576–7584, 2022.
  • [30] Yipeng Liu, Jiani Liu, Zhen Long, and Zhu Ce. Tensor Computation for Data Analysis. Springer International Publishing, 2021.
  • [31] Yu Liu, Quanming Yao, and Yong Li. Generalizing tensor decomposition for n-ary relational knowledge bases. In Proceedings of The Web Conference 2020, pages 1104–1114, 2020.
  • [32] Zhen Long, Yipeng Liu, Longxi Chen, and Ce Zhu. Low rank tensor completion for multiway visual data. Signal processing, 155:301–316, 2019.
  • [33] Zhen Long, Ce Zhu, Jiani Liu, and Yipeng Liu. Bayesian low rank tensor ring for image recovery. IEEE Transactions on Image Processing, 30:3568–3580, 2021.
  • [34] Andrew Y Ng, Michael I Jordan, and Yair Weiss. On spectral clustering: Analysis and an algorithm. In Advances in Neural Information Processing Systems, pages 849–856, 2002.
  • [35] Feiping Nie, Guohao Cai, and Xuelong Li. Multi-view clustering and semi-supervised classification with adaptive neighbours. In Thirty-first AAAI Conference on Artificial Intelligence, 2017.
  • [36] Feiping Nie, Jing Li, and Xuelong Li. Parameter-free auto-weighted multiple graph learning: a framework for multiview clustering and semi-supervised classification. In International Joint Conference on Artificial Intelligence, pages 1881–1887, 2016.
  • [37] Feiping Nie, Jing Li, and Xuelong Li. Self-weighted multiview clustering with multiple graphs. In International Joint Conference on Artificial Intelligence, pages 2564–2570, 2017.
  • [38] Shi-Ju Ran, Emanuele Tirrito, Cheng Peng, Xi Chen, Luca Tagliacozzo, Gang Su, and Maciej Lewenstein. Tensor Network Contractions: Methods and Applications to Quantum Many-Body Systems. Springer Nature, 2020.
  • [39] Hinrich Schütze, Christopher D Manning, and Prabhakar Raghavan. Introduction to information retrieval, volume 39. Cambridge University Press Cambridge, 2008.
  • [40] Yao Sui, Guanghui Wang, and Li Zhang. Sparse subspace clustering via low-rank structure propagation. Pattern Recognition, 95:261–271, 2019.
  • [41] Ledyard R Tucker. Some mathematical notes on three-mode factor analysis. Psychometrika, 31(3):279–311, 1966.
  • [42] Ledyard R Tucker et al. The extension of factor analysis to three-dimensional matrices. Contributions to mathematical psychology, 110119, 1964.
  • [43] Hao Wang, Yan Yang, and Bing Liu. Gmc: Graph-based multi-view clustering. IEEE Transactions on Knowledge and Data Engineering, 32(6):1116–1129, 2019.
  • [44] Xiaobo Wang, Xiaojie Guo, Zhen Lei, Changqing Zhang, and Stan Z Li. Exclusivity-consistency regularized multi-view subspace clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 923–931, 2017.
  • [45] Jianlong Wu, Zhouchen Lin, and Hongbin Zha. Essential tensor learning for multi-view spectral clustering. IEEE Transactions on Image Processing, 28(12):5910–5922, 2019.
  • [46] Jianlong Wu, Xingyu Xie, Liqiang Nie, Zhouchen Lin, and Hongbin Zha. Unified graph and low-rank tensor learning for multi-view clustering. In Proceedings of the AAAI conference on artificial intelligence, volume 34, pages 6388–6395, 2020.
  • [47] Wei Xia, Quanxue Gao, Qianqian Wang, Xinbo Gao, Chris Ding, and Dacheng Tao. Tensorized bipartite graph learning for multi-view clustering. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022.
  • [48] Yuan Xie, Dacheng Tao, Wensheng Zhang, Yan Liu, Lei Zhang, and Yanyun Qu. On unifying multi-view self-representations for clustering by tensor multi-rank minimization. International Journal of Computer Vision, 126(11):1157–1179, 2018.
  • [49] Yuan Xie, Wensheng Zhang, Yanyun Qu, Longquan Dai, and Dacheng Tao. Hyper-laplacian regularized multilinear multiview self-representations for clustering and semisupervised learning. IEEE transactions on cybernetics, 50(2):572–586, 2020.
  • [50] Yan Yang and Hao Wang. Multi-view clustering: A survey. Big Data Mining and Analytics, 1(2):83–107, 2018.
  • [51] Yuyuan Yu, Guoxu Zhou, Ning Zheng, Shengli Xie, and Qibin Zhao. Graph regularized nonnegative tensor ring decomposition for multiway representation learning. arXiv preprint arXiv:2010.05657, 2020.
  • [52] Longhao Yuan, Jianting Cao, Xuyang Zhao, Qiang Wu, and Qibin Zhao. Higher-dimension tensor completion via low-rank tensor ring decomposition. In 2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), pages 1071–1076. IEEE, 2018.
  • [53] Changqing Zhang, Huazhu Fu, Si Liu, Guangcan Liu, and Xiaochun Cao. Low-rank tensor constrained multiview subspace clustering. In Proceedings of the IEEE International Conference on Computer Vision, pages 1582–1590, 2015.
  • [54] Changqing Zhang, Huazhu Fu, Jing Wang, Wen Li, Xiaochun Cao, and Qinghua Hu. Tensorized multi-view subspace representation learning. International Journal of Computer Vision, 128(8):2344–2361, 2020.
  • [55] Changqing Zhang, Qinghua Hu, Huazhu Fu, Pengfei Zhu, and Xiaochun Cao. Latent multi-view subspace clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 4279–4287, 2017.
  • [56] Guang-Yu Zhang, Yu-Ren Zhou, Chang-Dong Wang, Dong Huang, and Xiao-Yu He. Joint representation learning for multi-view subspace clustering. Expert Systems with Applications, 166:113913, 2021.
  • [57] Qibin Zhao, Guoxu Zhou, Shengli Xie, Liqing Zhang, and Andrzej Cichocki. Tensor ring decomposition. arXiv preprint arXiv:1606.05535, 2016.
  • [58] Qinghai Zheng and Jihua Zhu. Multi-view subspace clustering with view correlations via low-rank tensor learning. Computers and Electrical Engineering, 100:107939, 2022.
  • [59] Qinghai Zheng, Jihua Zhu, Zhiqiang Tian, Zhongyu Li, Shanmin Pang, and Xiuyi Jia. Constrained bilinear factorization multi-view subspace clustering. Knowledge-Based Systems, 194:105–514, 2020.
  • [60] Shi Zhong and Joydeep Ghosh. A unified framework for model-based clustering. The Journal of Machine Learning Research, 4:1001–1037, 2003.
  • [61] Wenzhang Zhuge, Feiping Nie, Chenping Hou, and Dongyun Yi. Unsupervised single and multiple views feature extraction with structured graph. IEEE Transactions on Knowledge and Data Engineering, 29(10):2347–2359, 2017.