Multi-Tensor Network Representation for High-Order Tensor Completion
Abstract
This work studies the problem of high-dimensional data (referred to as tensors) completion from partially observed samplings. We consider that a tensor is a superposition of multiple low-rank components. In particular, each component can be represented as multilinear connections over several latent factors and naturally mapped to a specific tensor network (TN) topology. In this paper, we propose a fundamental tensor decomposition (TD) framework: Multi-Tensor Network Representation (MTNR), which can be regarded as a linear combination of a range of TD models, e.g., CANDECOMP/PARAFAC (CP) decomposition, Tensor Train (TT), and Tensor Ring (TR). Specifically, MTNR represents a high-order tensor as the addition of multiple TN models, and the topology of each TN is automatically generated instead of manually pre-designed. For the optimization phase, an adaptive topology learning (ATL) algorithm is presented to obtain latent factors of each TN based on a rank incremental strategy and a projection error measurement strategy. In addition, we theoretically establish the fundamental multilinear operations for the tensors with TN representation, and reveal the structural transformation of MTNR to a single TN. Finally, MTNR is applied to a typical task, tensor completion, and two effective algorithms are proposed for the exact recovery of incomplete data based on the Alternating Least Squares (ALS) scheme and Alternating Direction Method of Multiplier (ADMM) framework. Extensive numerical experiments on synthetic data and real-world datasets demonstrate the effectiveness of MTNR compared with the start-of-the-art methods.
Index Terms:
Multi-tensor network representation, tensor completion, tensor networks, tensor decomposition, adaptive topology learning, low-rank.I Introduction
With the rapid development of sensing and information technology in recent years, multimodal and large-scale data displayed in the form of a multi-dimensional array (referred to as tensors) are ubiquitous in various types, e.g., color images [1, 2], videos [3], and seismic data [4], etc. Tensors provide a favorable data structure and widely utilized in data processing and analysis, spanning from the fields of computer vision [5, 6], machine learning [7, 8], dimension compression [9], to signal processing [10, 11]. Unfortunately, missing entry and noise contamination problems are prevalent during data acquisition and transformation, leading to local detail incompleteness and global structure sparseness of the obtained results. That severely limits the practical value and causes the performance degradation of related applications [12]. The estimation of missing elements based on the inherent structure information of the observed entries is termed as tensor completion (TC) [13, 14, 15, 16].
The theoretical foundation of TC is the low-rank representation framework [12], which captures the global information by exploiting the low-rank property of incomplete data. Mathematically, the corresponding low-rank TC (LRTC) problem can be derived into the following optimization model as
(1) |
where is the observed incomplete tensor and is a linear operator that projects the elements in the observing entries set to itself and others to zeros.
For the 2D cases, the matrix rank is widely adopted to describe the degree of low-rankness of the data [17]. However, the rank function is nonconvex and hard to optimize. Since the nuclear norm is the tightest convex envelope of the rank of a matrix, which is frequently applied in the low-rank model [13, 18] and addressed by the singular value threshold (SVT) [19] algorithm. For the matrix of rank r with observed entries satisfying certain incoherence conditions [20], the missing elements will be recovered exactly with high probability. Nevertheless, the recovered matrix based on nuclear norm minimization strategy is usually suboptimal. Recently, several nonconvex continuous surrogate functions are employed to approximate the rank function, e.g., norm [21], Capped norm [22], and have achieved better estimation in the assist of nonconvex optimization techniques [23, 24].
For the N-D cases, the high-order LRTC problem is challenging due to the flexibility and diversity of tensor algebra operations and inconsistent definition of tensor rank [1]. In the multilinear products algebraic framework [25], multilinear tensor rank [2, 6] is defined as a vector consisting of the ranks of all mode unfolding matrices. In this context, the Sum of Nuclear Norm (SNN) [6, 26] is extensively utilized as a relaxed convex surrogate of tensor rank. The relaxed convex formulation of (1) can be written as
(2) |
where represents the mode- unfolding matrix of and the is nuclear norm defined as the sum of singular values, e.g., . However, the unfolding operation results in the loss of structural information, and SNN is not the tightest convex relaxation of multilinear tensor rank [27]. In the algebra framework of tensor-product (t-product) [28], third-order tensors can be regarded as extended linear operators of matrices. Kilmer et al. [29] introduce the definition of tensor tubal rank based on the tensor singular value decomposition(t-SVD) [28], which accurately estimated the low-rank structure of 3th-order tensors with a complete theoretical guarantee. Tensor nuclear norm (TNN) [30], a convex surrogate of tensor tubal rank, has a tight recovery bound and is applied for LRTC. Meantime, t-products with invertible transform can be extended to higher dimensions directly [31, 32] and implemented in other domains [33].
The huge advantage of the aforementioned rank-minimization-based TC methods is to estimate the tensor rank implicitly without specifying it in advance. Nevertheless, all existing approaches require a lot of singular value decomposition (SVD)[34] calculations in the optimization stage, which is computationally prohibitive and not applicable for large-scale scenariosa [35]. Another promising direction to address (1) effectively is tensor decomposition (TD) [25, 36, 37], which captures the intrinsic global structure of the data on latent space with multilinear operations over a set of latent factors. The TD-based LRTC model can formulated as
(3) |
here denotes the recovered tensor generated by a set of latent factors . Two of the most classic and universally applied TD models are CANDECOMP/PARAFAC (CP) decomposition [38] and Tucker decomposition [39]. CP decomposition [38] represents an N-order tensor as a sum of rank-1 tensor factors and defines the CP rank as the smallest number of factors to eliminate structural loss. Tucker decomposition [39] decomposes a tensor as multilinear operations among a core tensor and multiple factor matrix. Tensor network (TN) [40, 41], one of the powerful tools in physics, is broadly applied to solving large-scale optimization problems and provides graphical computation interpretation. TD models based on TN representation has recently attracted much attention, including tensor train (TT) [42] representation and tensor ring (TR) [43] representation, which can overcome the Curse of Dimensionality [40] with linear storage cost . Lately, TN is exploited to explore more complex topology structures [42, 43, 44] for TD, e.g., Projected Entangled Pair States (PEPS) [45] and Fully-Connected TN (FCTN) [46].
However, most of the available high-dimensional data suffer the difficulties of mode-dimension unbalance [47] (e.g., filters in convolution neural networks), and mode-correlation discrepancy (e.g., spatial modes in color image are strongly intercorrelated but their weakly correlated with channel mode). Such problems lead to the limited representation and flexibility of TD models with fixed topology. Thus, the following two limitations must be considered: (i) model selection and (ii) rank determination. Notice that the critical difference among various TD models lies in the latent factors connections (also called multilinear operations), as shown in Fig.2. Thus, a current encouraging direction is to explore the data-independent decomposition topology by virtue of TN[44, 48]. In [44], Hashemizadeh et al. gradually update the TD topology structure based on greedy search and rank incremental strategy. However, the selection of promising edges is tiring since it requires traversing and iterating all possible options. The genetic algorithm is used in [48] by Li and Sun to determine the connection relationship of factors. Since the rank determination for a fixed topology is NP-hard [49]. Cheng et al. [50] adopt reinforcement learning (RL) to determine the rank of TR representation and applied it to neural network compression tasks. In [51], the TR rank is gradually increased based on the measured sensitivity of the approximation error of factor. Also, the sparsity-inducing prior is employed by authors in [52, 53] to determining the optimal rank for TD models. Although many TD-based approaches exhibit favorable characteristics, they are either computationally prohibitive [46, 44, 48] or with limited generality [51, 53, 54].
According to the subspace clustering theory [55], usually the high-dimensional data, e.g., images and videos, which are located close to low-order structures related to several classes to which the data belong, and this property are widely used in unsupervised learning tasks include subspace clustering [56], infrared small target detection [57], latent convex tensor decomposition [58]. In this paper, we present a Multi-Tensor Network Representation (MTNR) framework for high-order tensor decomposition and apply it to the LRTC problem. Our goal is to decompose high-dimensional data into multiple low-rank components, and each component with TN representation can be represented as multilinear connections over several latent factors and naturally mapped to a specific topology. In particular, the structure of MTNR is explored adaptively from the instance data and can selectively establish connections with appropriate intensity for any two factors to capture correlations between data modes. To the best of our knowledge, this work is the first to utilize a combination of multiple TN models for high-order tensor decomposition and completion. In summary, the contribution of this article is threefold.

-
We propose a Multi-Tensor Network Representation (MTNR) framework for high-order tensor decomposition, which can be regarded as a combination of multiple TD models. In theory, we define the fundamental algebraic operations among the tensors with unified TN representation. Note that these properties allow us to perform multilinear operations among tensors represented by different TN models in a latent factor space with linear complexity instead of multilinear space with exponential complexity.
-
We present an adaptive topology learning (ATL) algorithm for MTNR based on a rank incremental strategy and a projection error measurement strategy. The structure of MTNR is data-dependent and can approximate complex low-rank information from different subspaces and obtains a better low-rank representation.
-
We apply MTNR to the typical TC task and develop two efficient algorithms: MTNR-ALS and MTNR-ADMM. The former is basic and based on the Alternating Least Squares (ALS) scheme [59] while the latter is built on the Alternating Direction Method of Multiplier (ADMM) framework [60], involving latent low-rank regularizers. The effectiveness of MTNR is demonstrated on both synthetic data and real-world datasets via extensive experiments.
The remainder of this paper is organized as follows. In Section II, the preliminaries of tensor notations and operations are introduced. Section III presents the MTNR framework and TN-based tensor algebraic operations in detail. In Section IV, two algorithms are proposed for solving the LRTC problem. The numerical analysis and experimental comparisons are given in Section V. Finally, we summarize this work in Section VI.
II Notations and Preliminaries
In this paper, we use boldface calligraphic letters to denote tensors of order , e.g.,; boldface capital letters to denote matrices, e.g.,. The boldface lowercase letters and lowercase letters respectively denote vectors and scalars, e.g., and . For a N-order tensor , its -th entry is denote as or , where . The denotes a set including integers from 1 to . The inner product of is defined as . The Frobenius norm and norm of are defined as and , which also applies to matrices and vectors. For a matrix , the rank of the is represented as , and the nuclear norm of is denoted as , where represents the -th eigenvalue of . The conjugate transpose, inverse, and pseudo-inverse of matrix are denoted as , and , respectively. indicates the identity matrix of size . In addition, some definitions of tensor operations [2, 25, 43] involved in the paper are given as follows.
Definition II.1.
(Vectorization and Matricization). For an -order tensor . The vectorization of is denote as , with entries . The mode- matricization (also known as unfolding or flattening) of arranges the mode-k fibers into the columns of the resulting matrix, denote as and its elements are given by
Furthermore, the -matricization of is denote as , which can be regarded as mode-1 matricization of after merging the first modes and its elements are given by
Note that the multi-indices is defined by little-endian convention [40] as with .
Definition II.2.
(Multilinear Tensor Rank). For an -order tensor . The multilinear tensor rank (also known as Tucker rank) of is a vector and denoted as , where represents mode- Matricization of .
Then the SNN can be defined as . We further present several basis tensor operations used throughout this paper.
Definition II.3.
(Hadamard Product). Let two -order tensors . The Hadamard product between these two tensors is element-wise product and defined as , with entries .
Definition II.4.
(Kronecker Product). Let two -order tensors and . The Kronecker product between these two tensors yields an -order tensor of size , with entries
Definition II.5.
(Mode-n Khatri-Rao product). Let an -order tensor and have . The mode-n Khatri-Rao product between these two tensors yields an -order tensor of size , with entries and subtensors .
Definition II.6.
(Outer Product). Let an -order tensor and th-order tensor . The outer product between these two tensors yields an th-order tensor of size , with entries .

Definition II.7.
(Mode- product). Let an -order tensor and matrix . The mode- product between these two tensors yields an -order tensor of size , with entries .
Moreover, the complicated relations and operations among tensors can represent graphically and intuitively by TN [40]. As shown in Fig.1, the vertices with edges denote an th-order tensor, and each edge denotes one mode. The standard edge, connecting two tensors, represents the modes involved in multilinear operations. To summation along the standard edge between two connecting tensors is called tensor contraction.
Definition II.8.
(Tensor Contraction). For an -order tensor and th-order tensor with common modes, . Let and . The tensor contraction between these two tensors alone the common modes yields an th-order tensor of size , with entries .
In essence, tensor contraction is a high-dimensional generalization of matrix multiplication [40] since . Thus, the mode- product and multilinear product can be classified as tensor contraction.
III Multi-Tensor Network Representation
Equipped with the above preliminaries, we present the MTNR framework in detail in this section. We first introduce the standard TN representation and extend it to multiple cases. Then, an adaptive topology learning algorithm is proposed for MTNR. Finally, the fundamental algebraic operations are established among the tensors with TN representations.
III-A Tensor Network Representation
The TN model represents a high-order tensor as the multilinear operation over a set of latent factors, and the number of factors is equivalent to the order of the tensor without considering the internal vertices. As illustrated in Fig.2, the difference among various TD models lies in the topological structure constructed by the connections among latent factors, e.g., the TR model [43] establishes a connection between the head and tail factors based on TT [42]. For the sake of clarity, we employ to denote the TN model uniformly. Suppose that an -order tensor is decomposed by and we obtain factors with . Hence, the rank of can be naturally expressed as a matrix , which is
(4) |
where represents the dimension of common mode between and . We call it edge rank and it satisfies (). The element-wise form of decomposition can be expressed as
(5) |
where denotes tensor contraction operation and indicates the recovered tensor generated by factors. Then we introduce the new definition as follows:
Definition III.1.
(Subnetwork). Let an -order tensor be decomposed by and has TN representations , where for . Given a vector , which is reordering of , then the contraction of subnetwork that consisting of the factors will obtain an -order tensor of size .
The calculation of in (5) can be completed with times matrix multiplication because the mode-n matricization of can be written as
(6) |
where is an -order tensor of size , obtained by contracting the subnetwork of composed of N-1 factors except . Similarly, the , and are defined as , and , respectively.
Note that the TN representation (5) does not demand a concrete topology strictly, and conceptually generalizes a family of TD models, including TT and TR representation. This unified representation gives us the chance to learn adaptive decomposition topologies for various data. Another tensor form of decomposition can be formulated as
(7) |
where denote th mode- fiber of factor . Then the reconstructed tensors can be expressed as the superposition of the rank-1 tensors. In general, TN representation with arbitrary topology is essentially an undirected graph composed of several vertices and edges, and exist a bijective with the simple graph [48]. If condition is satisfied, then the TN representation of can be regard as a complete graph, as illustrated graphically in Fig.2(f).
Remark III.1.
All the rank-1 edge is removed before the tensor network contraction since those edges are meanless and do not affect the computation consistency. Hence, two connected factors indicate the corresponding edge rank is larger than one, and the isolated subnetwork is calculated by the outer product [48]. In addition, the self-loops and multiple edges structure is prohibited to avoid confusion.
Since the topology of the is not strictly limited. For the TN representation with factors, there exist kind of topology when taking no account of the edge rank magnitude, from the (e.g., TN representation with isolated vertices) to a fully connected . Hence, the selection of TD models can also be viewed as a topology searching process, which is a discrete optimization problem analogous to the layer-wise architecture search in convolutional neural networks [50]. From (5), we can observe that the edge rank determines the topology and complexity of . Supposing that all the elements in are equal to and , then the number of parameters requires of decomposition is and the computational complexity of operation is . However, such storage complexity increases exponentially with the tensor order and the contraction is computationally too expensive for large-scale data. In order to solve this problem, we set the maximum connections of each latent factor as , where is a constant and satisfies . The existence of decreases the degree of freedom of feasible topology and contraction complexity. Thus, the decomposition required parameters are , which is linear with the tensor order, and this property reduces the complexity of calculation and storage significantly.
The relationship of the rank of and multilinear rank is established in the following theorem.
Theorem III.1.
Let an -order tensor be represented by Equation (5), then the following two inequalities hold for :
(8) |
As a side note, the generalized TN representation defined in this section prohibits the existence of internal factors (e.g., the cores tensor in Tucker decomposition) since they can be eliminated by tensor renormalization group (TRG) [61] algorithm or tensor contraction strategies. For instance, an efficient TR decomposition algorithm is proposed in [62] base on Tucker compression.

III-B Multi-Tensor Network Representation
Based on subspace clustering [55] theory, high-dimensional data can be regarded as the combination of low-rank components from different subspaces. As such, an -order tensors can be represented by an addition of multiple components as
(9) |
where represents the th low-rank component for . We can observe that if each component is a rank-1 tensor, then (9) is equivalent to CP decomposition and is just CP-rank.
Most of the high-dimensional data are complicated and mode-interrelated. The TD model with simple topology (e.g., ring-shaped TR) to capture the intrinsic low-rank structure inside the data and only require linear storage complexity [43]. However, the connections involved in topology only exist in the adjacent factors, leading to exponential correlation decay and expecting a larger rank to capture more global features. On the contrary, the TD model with complex topology can achieve better results, e.g., FCTN [46], but the storage and computational complexity are exponential with the order of the data. In order to better weigh the representation capability and complexity of TN models. As shown in Fig.3, MTNR aims to discover TN topologies from data and manipulate multiple TD models with different topologies to decompose components separately, which is critical compared with other TD models. Then, (9) can be rewritten as
(10) |
where represents the th TD models with a simple topological structure for , and the corresponding latent factor is obtained via . From (10), we see that MTNR can be considered as a combination of multiple TD models, which provides MTNR with more compelling representation and generalization capabilities. The amount of parameters for MTNR is , which is linear to the tensor order. Note that the structure of each TN model in MTNR is data-adaptive to better capture the low-rank structure and reduce approximation deviation. Besides, the combination of multiple identical TD models retains the properties of a single one, such as inverse permutation invariance in TT, and circular dimensional permutation invariance in TR.
III-C Adaptive Topology Learning Algorithm
A key problem of low-rank approximation based on TN representation is to search adaptive topology for various complicated and mode-imbalanced data. Existing data-driven approaches are difficult for data with different orders and sizes (such as neural architecture search and reinforcement learning). In this part, we present a novel efficient adaptive topology learning (ATL) algorithm for MTNR.
The main objective of ATL is to obtain a low-rank approximation of data with less storage resources. Given an -order tensor and relative error , the optimization problems of MTNR can be formulated as
(11) |
Here denotes storage cost function and is the predefined upper bound of the th low-rank component representation. In fact, instead of specifying the number of components , we choose to sequentially optimize each low-rank component to fit the current residual estimation, which can reduce the complexity of the topological search and obtain more stable outcomes. To solve the optimization problem of each low-rank component is based on the following two strategies.
1) Rank incremental strategy. The rank incremental strategy [44, 51] suggests that tensors with TN representation starts from the rank-1 tensor, successively selects a factor pair and increases its corresponding edge rank to greatly reduce approximate error. The topology and ranks obtained in this manner have better data adaptability. Specifically, for the th low-rank component, , the corresponding optimization problem can be derived from (11) as
(12) |
where represents residual estimation tensor. The (12) includes inter-coupled variables, so the well-known ALS [43] scheme or gradient descent algorithm [63] can be exploited to obtain the numerical solution. The principle of the ALS scheme is to update each variable sequentially while fixing the others until all the variables satisfy the convergence criterion. Then the subproblem w.r.t.the th factor variable can be expressed in a mode- matrix form as
(13) |
which is a least-square optimization problem and has a closed-form solution
(14) |
All factors are circulantly updated until satisfy , where is current iteration number and is a specific error level. In this context, we will identify a promising factor pair and increase its corresponding edge rank according to the following projection error measurement strategy if the complexity restraints are satisfied.
2) Projection error measurement strategy. The residual tensor of size is approximated by the th low-rank component , with TN representation , where . Since the correlation within different modes of the reconstructed tensor is established through the connection within the factors, we can increase the edge rank to improve the correlation between the and modes of . Each selection for an -order tensor produces cases. A simple technique is greedy search [44], which uses brute-force search to test all options and for each trial it requires several additional iterations to return the most promising edges to improve the model performance. However, that is too computationally expensive. Inspired by [51], we propose an improved projection error measurement strategy for identifying the optimal factor pair. For satisfies , consider the following problem
(15) |
where of size denotes the mode- matricization of the approximation error tensor , and is viewed as the error projection matrix on the factors and , computed by . We use the , which we called projection error measurement, to evaluate the importance or sensitivity of the factors pair . In each iteration, the order of factors is updated randomly to evade the influence of mode arrangement. In section V, we demonstrate can effectively measure the importance of the local topology of the TN representation. Moreover, we set three conditions to guarantee the effectiveness of the selection process.
(i) Refuse to select the identical factor pair twice consecutively.
(ii) The number of selections for each factor pair is less than the corresponding mode dimension.
(iii) The maximal connections of each factor are set as .
The above conditions are used to ensure the diversity and feasibility of the TD topologies generated by ATL. In addition, these topologies are robust to the arrangement of data modes. Overall, we will continuously collect the low-rank components until the stop criterion holds. Algorithm 1 summarizes the whole learning process of MTNR in detail. The significance of the low-rank components obtained by MTNR is decreasing gradually, and the corresponding TN topology is expected to seek the optimal one under the current structure error. In this way, MTNR is prone to form multiple TN models with different topologies with linear parameters.
III-D Multilinear operations with TN representation
An interesting conjecture is whether it is possible to perform multilinear operations in a latent factor space instead of a multilinear space for tensors with TN representation. This section gives the affirmative answer and theoretically establishes basic algebra operations, including multilinear, Hadamard, outer, inner, Kronecker, and mode-n Khatri-Rao products and addition operation. These properties allow us to process large-scale data with linear storage cost and significantly reduce the computational complexity. In [42, 43], the authors establish basic multilinear operations for the tensors with TT and TR format, and here we further extend it to a generalized scenario and thus we have the following propositions.
Proposition III.1.
Suppose that an -order tensor is decomposed by , having a TN representation of , where . Then the mode- product (see Def. II.7) of and a matrix can be written as:
(16) |
Moreover, the multilinear product of with a set of vectors, for , can be computed by
(17) |
The multilinear product is performed on each factor and its result is also given in a TN format. For instance, if the number of connections of each factor is no more than , then the computational complexity of (16) and (17) is equivalent to and , respectively.
Proposition III.2.
Let two -order tensors be respectively decomposed by and , which have TN representations and , where and for .
1) The Hadamard product (see Def. II.3) of and yields a tensor , having a TN representation of , where each factor of size , for , can be computed by
(18) |
where is th subtensor of in the th mode.
Proof. The n-matricization of and can be formulated as and , then the following equation holds
(19) |
we can futher obtain
(20) |
Therefore, Eq.(18) holds by mathematical induction and the Hadamard product of tensors with TN representation can perform in their factors via mode-n Khatri-Rao product.
2) The inner product of and can be computed by
(21) |
where is a vector with all values of one. Naturally, the Frobenius norm of can be obtained by .
Proposition III.3.
Let an -order tensor and th-order tensor be respectively decomposed by and , which have the TN representations of and . The outer product (see Def. II.6) of and yields a tensor , which has a TN representation of .
Moreover, the Kronecker and mode-n Khatri-Rao product of tensors with TN representation can perform in their factors and the result will also follow the TN format.
Proposition III.4.
Let two -order tensors and be decomposed by and , which has a TN representations of and , where and for . The Kronecker product (see Def. II.4) of and yields a tensor , having the TN representation of , where each factor of size , for , can be computed by
(22) |
Proof. The Kronecker product of and can be written as
(23) |
Proposition III.5.
Let two -order tensors and be decomposed by and , which have the TN representations of and , where and for . Suppose that , then the mode-n Khatri-Rao product (see Def. II.5) between these two tensors yields a tensor , having a TN representation of , where each factor , for , can be computed by
(24) |
The Propositions III.1III.5 establishes the essential multilinear operation of tensors with TN representation, which is implemented straightly on the latent space and obtain the TN format results. Note that these operations may cause changes in the size of the factors and aggravate the computational complexity of the tensor contraction. We further define the addition of the tensors with TN representation in Theorem III.2, indicating that the combination of multiple TD models essentially can be transformed into a single one.
Theorem III.2.
Let two -order tensors be decomposed by and , which have the TN representations of and , where and for . Then the addition of these two tensors, , having a TN representation of , where each factor of size , for , can be computed by
(25) |
The establishment of Theorem III.2 requires that the topology of the resulting tensor TN representation cannot contain isolated factors; otherwise we will randomly construct an additional edge for the factors with zero connections. It can be observed that the addition of tensors with TN representations is the combination of the corresponding factors into a higher-order diagonalized format. Therefore, we have the following remark:
Remark III.2.
-
(1)
The addition of tensors represented by the same TN models without isolated vertices can obtain a tensor in a TN format with identical topology.
-
(2)
At least addition (resp. ) tensors with TR (resp. TT) representation to guarantee that the result tensor corresponds topology is a complete graph.
In summary, multilinear operations among higher-order tensors can be carried out implicitly in the latent space, and the storage cost is linear with the tensor order.
IV MTNR FOR LOW-RANK TENSOR COMPLETION
This section applies MTNR to the LRTC task, aiming to obtain a low-rank approximation of the data from limited samplings. We propose two practical algorithms, named MTNR-ALS and MTNR-ADMM, and their computational complexity is analyzed.
IV-A MTNR-ALS Algorithm
The MTNR-ALS is an extension of Algorithm 1 on the LRTC task. Given an -order incomplete observation tensor and entries index set . Our objective is to obtain the approximated tensor, , in the MTNR format (10). Hence, the MTNR-based optimizing model for LRTC can be formulated as
(26) |
where represents the th low-rank component . For the sake of clarity, we omit the constraint condition in (11). The (26) cannot be optimized directly due to is assumed to be an unknown coefficient. Consistent with (12), we consider the th low-rank component’s optimization model
(27) |
We use the to fit the present residual observation tensor . The variables and are interrelated. We employ the ALS scheme to address (27) by alternately updating
(28) |
where denotes indices set of missing entries, and the closed-form solution of -subproblem is given in (14). We employ the same strategy as the ATL algorithm to perform the update of the factor structure. The optimizing model (26) can be regarded as MTNR if the observed data is no missing elements. The whole implementation process of the MTNR-ALS algorithm is listed in Algorithm 2.
Compared with other TD-based approaches for LRTC, e.g., TT-ALS and TR-ALS [54], the main advantage of MTNR-ALS is the TN representation with adaptive topology and the number of low-rank components is adjusted dynamically, which is helpful for capturing the low-rankness inside the high-dimension data. Nevertheless, the low-rankness of the recovered tensor is weakened during the increase of the number of low-rank components, which may be unfavorable for the results. We consider developing a new algorithm to guarantee the low-rank property of the reconstructed tensor.
IV-B MTNR-ADMM Algorithm
The TD-based LRTC methods incorporating low-rank regularizations can achieve better performance [58] as the explicit tensor rank constraint and the implicit convex surrogate function (e.g., SNN [6]) are applied to the recovered tensor to better capture the global low-rank structure information. However, the computational complexity of SNN that directly acts on the recovery tensor is , which grows exponentially with the data order and obtains suboptimal results. Thereby, we consider applying implicit low-rank regularizations to the factors of TN representation to reduce the calculation. In a nutshell, we impose SNN regularization to the factors instead of the original format with the computational complexity , which is linear to tensor order . Specifically, the optimization model of th low-rank component is reformulated as
(29) |
where is a trade-off coefficient. Note that the low-rank regularization is dynamically adjusted as the factor structure update during the optimization process. If a mode- matricization of is present in vector format, then the corresponding nuclear norm will be invalid. Since variables in (29) are inter-coupled and cannot optimized independently, we introduce auxiliary variables to reduce the optimization difficulty and the model (29) is deduced to
(30) |
where and represent and , respectively. The above multivariable optimization problem can be solved via the ADMM framework. The augmented Lagrangian function of (30) can be formulated as follows
(31) |
where is Lagrangian multiplier tensor set and is a positive penalty parameter. The solution of (31) is given via alternately updating variables and . The update mechanism is to optimize one variable each time and fix others, which is as follows:
-
(1)
Update . Extract all items that are associated with from (31). Then the -subproblem, for , can be formulated as:
(32) The closed-form solution of is given in mode- matricization format as follows:
(33) where is a identity matrix of size .
-
(2)
Update . The optimization subprobelem for , , is formulated to
(34) The closed-form solution of above model is given in the mode- matricization format as follows:
(35) where is the singular value thresholding (SVT) [34] operator. Given a matrix , whoseich singular value decomposition is expressed as , then , where is the identity matrix of the same size as .
-
(3)
Update and . The update rules w.r.t , for , is
(36) The update strategy of is consistent with (28), which can be computed by
(37)
Algorithm 3 summarizes the overall optimization steps of MTNR-ADMM.
IV-C Computational Complexity and Convergence Analysis
In this part, we first analyze the computational complexity of Algorithms 13. Since the TN topology updates dynamically in the optimization process, we only consider the general case, that is, the number of connections of each factor and the edge rank are and , respectively. For an -order incomplete tensor with , the computational cost of Algorithm 1 (ATL) on each iteration mainly includes two parts: 1) TN contraction in step 5; 2) updating in step 6. The TN contraction aims to obtain th low-rank component with tensor format, which conducts the matrix multiplication -1 times and the computational complexity is . In step 6, the computational complexity of calculating and updating are and . Moreover, Algorithms 2 and 1 have the same computational complexity due to the fact that the update cost of is negligible to the whole algorithm. Hence, the whole computational complexity of MTNR-ALS at each iteration is . In practice, such a heuristic algorithm requires more iterations due to the updating strategy. Compared with other TD-based methods including TR-ALS [43] and FCTN-PAM [46] with computational complexities costs being and , respectively, the overall computational complexity of MTNR-ALS is higher than TT-ALS, but it is lower than FCTN when facing higher-order tensors.
For Algorithm 3 (MTNR-ADMM) with implicit low-rank regularization, the computational complexities of updating (step 7) is . Then the computational complexities of Algorithm 3 is . In comparison, for rank-minimization-based methods, SiLRTC [6] and TRNNM [64] require performing the full SVD in matricization format matrix and the total cost are and at each iteration. Therefore, imposing low-rank restraint in the latent factor can reduce considerable calculations.
We establish the convergence guarantee of the proposed Algorithm 2. Naturally, the convergence of Algorithms 1 can also be obtained in a similar manner [65]. Since each low-rank component is designed to fit the current observation residuals, we just need to prove the convergence of the factors under a fixed structure.
Theorem IV.1.
The Lagrangian function of (27) is , where is introduced as a Lagrangian multiplier. The solving strategy of model (27) is a special scenario in [66], so satisfying Theorem IV.1 is based on the following four conditions.
(a) The sequence of is non-increasing and is a bounded sequence;
(b) has Lipschitz constant on any bounded set;
(c) has a Lipschitz constant with respect to , and there exists a constraint , such that for all and ;
(d) satisfies the Kurdyka–Łojasiewicz property [67] at .




V EXPERIMENTAL RESULTS
In this section, we evaluate the proposed algorithms, MTNR-ALS and MTNR-ADMM, and compare them with other state-of-the-art TC methods. We first give the basic experimental settings. Then, extensive experimental results on synthetic data, benchmark color images, face datasets and color videos are presented to demonstrate the superiority and effectiveness of our approach.
V-A Experimental Setting
1) Parameter setting and environment. In our method, the two essential hyperparameters are the maximum number of connections and parameters upper bound , which control the computational and storage complexity of the algorithm. Unless otherwise stated, we set and , where and are the order and mode dimension of the instance tensor, respectively. Hence, the storage complexity of TN representation of low-rank components can be guaranteed to be linear with the data order. For Algorithms 13, the relative error , threshold , and are set to -2, -3, and , respectively. For algorithm 3, the trade-off coefficient , positive penalty parameter , and are set to 1, -1 and 30, respectively. All our experiments are performed on the same platform under Windows 10 on a PC with a 3.0 GHz CPU and 64G RAM.
2) Compared methods. For comparison, we choose several start-of-the-art tensor completion algorithms, including TD-based TT-ALS [42], TR-ALS [54], FCTN-PAM [46], and TR-WOPT [68], where the FCTN-PAM corresponding TN topology can be regarded as a complete graph since it establishes connections among any two factors. The rank-minimization-based methods include SiLRTC [6], HaLRTC [6], and TRNNM [64]. The compare method based on both TD and rank-minimization is TRLRF [35]. We carefully adjust the parameters as suggested in the original paper for the above methods to obtain the best results. The computational complexity of the compared methods is shown in Table I.
3) Evaluation indices. We adopt relative square error (RSE) [5], peak signal-to-noise ratio (PSNR) [1], and structure similarity (SSIM) [1] as quality indices to evaluate the performance of the reconstruction results. For synthetic data, RSE is applied as the quantitative evaluation metric and defined as
(38) |
where and are the recovered tensor and ground truth, respectively. A smaller RSE value implies a better performance. For images data, PSNR and SSIM are applied as the quantitative evaluation metric and defined as
(39) |
where and are the recovered image and ground truth image; and are the mean value and standard variance of ; , , , and are the maximum value, number of entries, mean value, and standard variance of , respectively. The covariance of and is , and are constant. Specifically, PSNR and SSIM measure the visual quality and structure similarity, and a large value of PSNR or SSIM indicates a better recovery accuracy.
Image | Metric | TT-ALS | TR-ALS | FCTN-PAM | TR-WOPT | MTNR-ALS | SiLRTC | HaLRTC | TRNNM | TRLRF | MTNR-ADMM |
---|---|---|---|---|---|---|---|---|---|---|---|
Barbara | RSE | 0.2272 | 0.1686 | 0.1738 | 0.2884 | 0.1477 | 0.3518 | 0.3025 | 0.2356 | 0.1644 | 0.1489 |
PSNR | 19.31 | 21.89 | 21.63 | 17.23 | 23.02 | 15.50 | 16.81 | 18.98 | 22.11 | 22.96 | |
SSIM | 0.5498 | 0.6667 | 0.6455 | 0.4286 | 0.7359 | 0.4553 | 0.4991 | 0.6095 | 0.6786 | 0.7351 | |
House | RSE | 0.1664 | 0.1061 | 0.1148 | 0.1714 | 0.1028 | 0.2591 | 0.2086 | 0.1518 | 0.0980 | 0.09710 |
PSNR | 20.32 | 24.11 | 23.42 | 19.98 | 24.37 | 16.37 | 18.31 | 21.06 | 24.79 | 24.88 | |
SSIM | 0.5999 | 0.7348 | 0.6852 | 0.5879 | 0.7506 | 0.5931 | 0.6558 | 0.7556 | 0.7605 | 0.7830 | |
Peppers | RSE | 0.2603 | 0.1789 | 0.1896 | 0.2734 | 0.1811 | 0.3838 | 0.3576 | 0.2622 | 0.1723 | 0.1691 |
PSNR | 18.17 | 21.00 | 20.53 | 17.36 | 20.95 | 14.67 | 15.28 | 17.95 | 21.32 | 21.45 | |
SSIM | 0.5064 | 0.6303 | 0.5741 | 0.4430 | 0.6301 | 0.4594 | 0.4830 | 0.5945 | 0.6498 | 0.6778 | |
Sailboat | RSE | 0.2422 | 0.1832 | 0.1821 | 0.2383 | 0.1824 | 0.3284 | 0.3011 | 0.2254 | 0.1848 | 0.1666 |
PSNR | 17.81 | 20.11 | 20.09 | 17.80 | 20.39 | 14.96 | 15.81 | 18.36 | 20.03 | 20.83 | |
SSIM | 0.4893 | 0.6013 | 0.5996 | 0.4885 | 0.6120 | 0.4531 | 0.4873 | 0.6065 | 0.5944 | 0.6555 | |
Lena | RSE | 0.2030 | 0.1610 | 0.1755 | 0.2086 | 0.1628 | 0.2996 | 0.2627 | 0.2065 | 0.1544 | 0.1499 |
PSNR | 19.15 | 21.07 | 20.33 | 18.84 | 20.96 | 15.80 | 16.94 | 18.96 | 21.41 | 21.67 | |
SSIM | 0.5483 | 0.6244 | 0.5530 | 0.5049 | 0.6312 | 0.5045 | 0.5458 | 0.6343 | 0.6458 | 0.6861 | |
Baboon | RSE | 0.2745 | 0.2582 | 0.2856 | 0.2592 | 0.2499 | 0.3473 | 0.3240 | 0.2656 | 0.2537 | 0.2357 |
PSNR | 16.58 | 17.11 | 16.22 | 17.07 v | 17.38 | 14.53 | 15.15 | 16.86 | 17.25 | 17.89 | |
SSIM | 0.3630 | 0.3793 | 0.3371 | 0.3682 | 0.3902 | 0.3487 | 0.3591 | 0.4155 | 0.3834 | 0.4338 | |
Airplane | RSE | 0.1552 | 0.1150 | 0.1203 | 0.1865 | 0.1124 | 0.2078 | 0.1832 | 0.1400 | 0.1197 | 0.1058 |
PSNR | 19.11 | 21.61 | 21.18 | 17.35 | 21.79 | 16.43 | 17.59 | 19.96 | 21.33 | 22.25 | |
SSIM | 0.5721 | 0.6592 | 0.6440 | 0.4801 | 0.6943 | 0.5318 | 0.5960 | 0.7026 | 0.6747 | 0.7275 | |
Facade | RSE | 0.1507 | 0.1224 | 0.1482 | 0.1942 | 0.1114 | 0.2639 | 0.2015 | 0.1598 | 0.1188 | 0.1174 |
PSNR | 22.19 | 24.01 | 22.35 | 19.99 | 24.82 | 17.33 | 19.68 | 21.69 | 24.26 | 24.36 | |
SSIM | 0.7044 | 0.7843 | 0.7193 | 0.5767 | 0.817 | 0.4813 | 0.5732 | 0.7115 | 0.7913 | 0.7948 | |
Starfish | RSE | 0.1216 | 0.0993 | 0.0925 | 0.1167 | 0.0833 | 0.1641 | 0.1252 | 0.0923 | 0.0907 | 0.07422 |
PSNR | 20.35 | 22.02 | 22.64 | 20.65 | 23.53 | 17.65 | 20.00 | 22.66 | 22.82 | 24.54 | |
SSIM | 0.6610 | 0.7374 | 0.7010 | 0.6369 | 0.7552 | 0.6127 | 0.6994 | 0.7777 | 0.7457 | 0.7927 | |
Sea | RSE | 0.1372 | 0.1049 | 0.1128 | 0.1509 | 0.1024 | 0.2224 | 0.1713 | 0.1258 | 0.1021 | 0.1029 |
PSNR | 21.11 | 23.45 | 22.81 | 20.08 | 23.64 | 16.95 | 19.18 | 21.86 | 23.66 | 24.37 | |
SSIM | 0.6769 | 0.7319 | 0.7220 | 0.6025 | 0.7455 | 0.5848 | 0.6707 | 0.7329 | 0.7412 | 0.7615 |
Data | Permutation | TT-ALS | TR-ALS | FCTN-PAM | TR-WOPT | MTNR-ALS |
---|---|---|---|---|---|---|
Syn-1 | Original | 0.192 | 0.137 | 0.125 | 0.202 | 0.108 |
Transpositional | 0.224 | 0.149 | 0.125 | 0.216 | 0.110 | |
Syn-2 | Original | 0.1524 | 0.0946 | 0.0853 | 0.162 | 0.0735 |
Transpositional | 0.147 | 0.0901 | 0.0853 | 0.155 | 0.0733 | |
Syn-3 | Original | 0.0969 | 0.082 | 0.0621 | 0.102 | 0.0640 |
Transpositional | 0.0918 | 0.0784 | 0.0620 | 0.096 | 0.0636 | |
Syn-4 | Original | 0.0750 | 0.0498 | 0.0234 | 0.108 | 0.024 |
Transpositional | 0.0628 | 0.0446 | 0.023 | 0.0854 | 0.022 |










V-B Synthetic Data
To verify the robustness and effectiveness of the algorithms, we generated four different sorts of data. The Syn-1 is an ordinary th-order tensor of size , Syn-2 is a mode-unbalanced th-order tensor of size , Syn-3 is a th-order tensor of size with a large discrepancy in mode correlation, and Syn-4 is a th-order tensor of size with both mode imbalance and discrepancy. The Syn-1 and Syn-2 are obtained via the addition of 32 rank-1 tensors of the same size as the target tensor. The Syn-3 and Syn-4 are generated with TT representation of rank because TT only establishes connections within adjacent factors and has the problem of exponential decay of mode correlation [40]. Moreover, the elements of the generating factors are randomly sampled from the standard normal distribution . The compared methods include TT-ALS, TR-ALS, FCTN-PAW, TR-WOPT, and MTNR-ALS, where TR-WOPT and FCTN-PAW algorithms are based on gradient descent strategy and PAM framework [46], respectively.

1) Different missing rates. We set the missing rate (MR) varying from 10% to 99%, and all entries are missing at random (MAR). In Fig. 4, we report the RSE values (best results of 10 independent experiments) of five TD-based methods. As can be seen, with the increase of MR, the completion performance of all algorithms decrease. Compared with other methods, MTNR-ALS achieves the best performance, and the advantage of MTNR-ALS is more apparent when the MR is larger than 80%. Conversely, TR-WOPT yields the worst result and cannot obtain reliable convergence. Fig. 4 shows that TR-WOPT cannot get a reasonable lower RSE value even under a lower MR (e.g., 10%). In terms of time performance, MTNR-ALS requires more time than the counterparts since it requires updating structure additionally. It is worth noting that the ranks of the comparison methods, except MTNR-ALS, require being manually tuned to obtain the best results. In general, a smaller rank is recommended for an incomplete tensor with a large MR. However, the rank setting is challenging for high-dimensional data with different MR conditions, and it will significantly affect the accuracy of completion. Hence, a unique advantage of MTNR-ALS is not required to specify the rank in advance, and the rank of MTNR-ALS is generated automatically based on the rank incremental strategy. We noticed that MTNR-ALS delivered satisfactory results under different MR conditions, which attributes to its adaptive topology and rank. Compared with other TD models with fixed rank, our approach has better robustness and higher completion performance on synthetic data when MR is 99%. This indicates the effectiveness of adaptive structural decomposition in TN completion based approaches.
2) Different mode permutations. As we know, TD models tend to sensitive to mode permutation, which is related to their corresponding topological structure. For instance, the ring-shaped TR model possesses circular dimensional permutation invariance [43], but transposing adjacent modes of the original tensor will change the decomposition outcome. According to Section III, we know that the topology structure of MTNR is automatically generated during the optimization process and independent of the mode permutation. Therefore, MTNR has the property of transpositional invariance [46] as FCTN, which means the represented factors do not depend on the strict arrangement of the initial tensor mode. Notice that the factor contraction order is irrelevant to the results, and such property is produced during the tensor representation.
We analyze the influence of the initial mode’s arrangement of the incomplete tensor on the performance of the inpainting result by different algorithms. We transpose the 1st and 4th modes of Syn-1, the 2nd and 5th modes of Syn-2, the 2nd and 4th modes of Syn-3, and the 1st and 5th modes of Syn-4, respectively. The quantitative RSE values of inpainting results by different methods on four synthetic data (original and transpositional) are listed in Table III, where the MR is 90%. From Table III, the recovery results of TT and TR based approaches vary significantly under different mode arrangements. Fortunately, the MTNR and FCTN based approaches are robust to tensor mode transposition and obtain better results. This further provides empirical evidence for the mode arrangement invariance of MTNR.
Image | Missing pattern | Metric | TT-ALS | TR-ALS | FCTN-PAM | TR-WOPT | MTNR-ALS | SiLRTC | HaLRTC | TRNNM | TRLRF | MTNR-ADMM |
---|---|---|---|---|---|---|---|---|---|---|---|---|
House | MAR | RSE | 0.0826 | 0.0517 | 0.0474 | 0.1513 | 0.03108 | 0.0702 | 0.0441 | 0.0442 | 0.0509 | 0.0488 |
PSNR | 26.33 | 30.34 | 31.10 | 21,06 | 35.64 | 27.72 | 31.74 | 31.52 | 30.45 | 30.84 | ||
SSIM | 0.8502 | 0.9228 | 0.9313 | 0.5989 | 96.56 | 0.8984 | 0.9566 | 0.9534 | 0.9250 | 0.9202 | ||
RMAR | RSE | 0.5046 | 0.4851 | 0.0974 | 0.2194 | 0.09984 | 0.2372 | 0.2872 | 0.1940 | 0.1743 | 0.0928 | |
PSNR | 10.76 | 11.01 | 24.93 | 17.82 | 24.76 | 17.12 | 15.26 | 18.89 | 19.86 | 25.39 | ||
SSIM | 0.3018 | 0.2708 | 0.8379 | 0.4788 | 0.8506 | 0.6585 | 0.5036 | 0.6253 | 0.6438 | 0.8617 | ||
CMAR | RSE | 0.5387 | 0.4435 | 0.0649 | 0.1812 | 0.1121 | 0.1163 | 0.1099 | 0.1476 | 0.1076 | 0.0693 | |
PSNR | 10.10 | 11.74 | 28.95 | 19.48 | 23.62 | 23.38 | 23.90 | 21.29 | 24.05 | 27.81 | ||
SSIM | 0.2456 | 0.2930 | 0.8980 | 0.5429 | 0.8087 | 0.7900 | 0.8070 | 0.7131 | 0.8258 | 0.8862 | ||
RCMAR | RSE | 0.4552 | 0.3008 | 0.0729 | 0.1718 | 0.1146 | 0.1016 | 0.0857 | 0.1276 | 0.0958 | 0.0686 | |
PSNR | 11.47 | 15.21 | 27.40 | 19.98 | 23.52 | 24.52 | 26.01 | 22.60 | 25.18 | 27.92 | ||
SSIM | 0.3006 | 0.5226 | 0.8503 | 0.5627 | 0.8101 | 0.8127 | 0.8519 | 0.7527 | 0.8260 | 0.8825 | ||
Lena | MAR | RSE | 0.1040 | 0.0814 | 0.0628 | 0.1901 | 0.0537 | 0.0904 | 0.0741 | 0.06224 | 0.0802 | 0.0679 |
PSNR | 24.87 | 26.93 | 29.18 | 19.73 | 30.53 | 26.08 | 27.79 | 29.27 | 27.06 | 28.50 | ||
SSIM | 0.8382 | 0.8901 | 0.9232 | 0.6286 | 0.9328 | 0.8781 | 0.9181 | 0.9360 | 0.8923 | 0.9176 | ||
RMAR | RSE | 0.3474 | 0.4807 | 0.09132 | 0.1264 | 0.1683 | 0.2024 | 0.1863 | 0.1611 | 0.1246 | 0.0867 | |
PSNR | 14.50 | 11.86 | 25.88 | 23.15 | 20.86 | 19.20 | 20.06 | 21.16 | 23.31 | 26.40 | ||
SSIM | 0.4905 | 0.2812 | 0.8496 | 0.7458 | 0.8011 | 0.6094 | 0.7477 | 0.7099 | 0.8036 | 0.8871 | ||
CMAR | RSE | 0.5872 | 0.5282 | 0.1253 | 0.1159 | 0.1464 | 0.1797 | 0.1785 | 0.2366 | 0.1169 | 0.1253 | |
PSNR | 9.87 | 10.89 | 23.26 | 23.90 | 21.94 | 20.13 | 20.17 | 17.83 | 23.83 | 23.83 | ||
SSIM | 0.1952 | 0.2521 | 0.7414 | 0.7913 | 0.7257 | 0.6629 | 0.6638 | 0.5758 | 0.8138 | 0.7785 | ||
RCMAR | RSE | 0.3446 | 0.3134 | 0.0927 | 0.1904 | 0.1489 | 0.1367 | 0.1289 | 0.1617 | 0.1711 | 0.0892 | |
PSNR | 14.53 | 15.52 | 25.83 | 19.65 | 21.72 | 22.51 | 22.97 | 21.09 | 20.64 | 26.16 | ||
SSIM | 0.5040 | 0.5526 | 0.84907 | 0.6304 | 0.7213 | 0.7629 | 0.7791 | 0.7034 | 0.6875 | 0.8687 |
Data | Metric | TT-ALS | TR-ALS | FCTN-PAM | TR-WOPT | MTNR-ALS | SiLRTC | HaLRTC | TRNNM | TRLRF | MTNR-ADMM |
---|---|---|---|---|---|---|---|---|---|---|---|
Data-1 | RSE | 0.2347 | 0.2503 | 0.2582 | 0.2804 | 0.2462 | 0.4010 | 0.2540 | 0.2909 | 0.2306 | 0.2257 |
PSNR | 29.09 | 28.45 | 28.28 | 27.53 | 28.66 | 24.26 | 28.62 | 28.16 | 29.48 | 29.61 | |
SSIM | 0.9015 | 0.8691 | 0.8556 | 0.8387 | 0.8768 | 0.8425 | 0.8969 | 0.8994 | 0.8924 | 0.8919 | |
Data-2 | RSE | 0.1940 | 0.1766 | 0.1998 | 0.2831 | 0.1779 | 0.4199 | 0.2144 | 0.2504 | 0.1718 | 0.1728 |
PSNR | 27.10 | 27.89 | 26.93 | 24.13 | 27.89 | 22.13 | 26.61 | 25.75 | 28.28 | 28.33 | |
SSIM | 0.8932 | 0.9052 | 0.8864 | 0.8098 | 0.8991 | 0.8500 | 0.9106 | 0.9074 | 0.9140 | 0.9150 | |
Data-3 | RSE | 0.2054 | 0.1839 | 0.2184 | 0.3015 | 0.1809 | 0.4049 | 0.2149 | 0.2569 | 0.1884 | 0.1755 |
PSNR | 26.34 | 27.31 | 25.90 | 23.14 | 26.90 | 21.99 | 26.38 | 25.29 | 27.26 | 27.69 | |
SSIM | 0.8798 | 0.8870 | 0.8606 | 0.7876 | 0.8854 | 0.8394 | 0.9102 | 0.8996 | 0.8971 | 0.8936 | |
Data-4 | RSE | 0.2503 | 0.2653 | 0.2638 | 0.3187 | 0.2408 | 0.4635 | 0.2805 | 0.3196 | 0.2600 | 0.2396 |
PSNR | 24.84 | 24.33 | 24.44 | 22.84 | 25.09 | 21.22 | 24.48 | 23.77 | 25.03 | 25.36 | |
SSIM | 0.8540 | 0.8361 | 0.8274 | 0.7773 | 0.8516 | 0.8200 | 0.8862 | 0.8835 | 0.8642 | 0.8644 |




V-C Image Completion
We further evaluate the performance of MTNR-ALS and MTNR-ADMM in the color image completion task, and compare them with eight other start-of-the-art algorithms. Ten benchmark color images of size are shown in Fig. 5, which are reshaped to 5th-order tensor of size in the experimental process.
1) Completion with uniformly random missing. We use 10% randomly sampled entries (with a MR of 0.9) to perform the image completion task. For comparison methods, the parameter settings follow their original paper to achieve the best result. The methods TR-ALS, TR-WOPT, and TRLRF, represent the target tensor with TR format and the rank is set as 8. The rank of TT-ALS is a vector . In addition, the rank of FCTN-PAM is set to 4 due to limited computing resources. The quantitative indices RSE, PSNR, and SSIM are used and the best inpainting results among twenty independent experiments are reported in Table II. Note that the overall recovery performance is unsatisfactory due to the local structural information loss during the reshaping process and a higher MR.
For the TD-based methods in the left hand side of Table II, MTNR-ALS achieved the best completion performance on most of the test examples. Many results show FCTN-PAM is a little inferior to TR-ALS due to its limited robustness to target tensors with unbalanced mode dimensions. Besides, for higher-order tensors (), the optimization of FCTN-PAM is expensive since its computational and storage complexity increases exponentially with the tensor order. We find the gradient descent-based method TR-WOPT unstable during the optimization process and obtain poor reconstruction accuracy. For the TC methods with rank-minimization regularization in the right half of Table II, the SiLRTC, HaLRTC, and TRNNM directly impose conditions on the target tensor and obtain large RSE and small PSNR and SSIM values. The TRLRF and MTNR-ADMM with latent restraints achieve better performance than their counterparts TR-ALS and MTNR-ALS, demonstrating the performance gains due to TD and rank-minimization. Finally, our proposed MTNR-ADMM achieves the best performance in natural image completion tasks with uniformly random missing.
2) Completion with nonrandom missing. We also consider the image completion in the case of Row-wise missing at random (RMAR), Colum-wise missing at random (CMAR), and Row-Column-wise missing at random (RCMAR). The images ”House” and ”Lena” were selected for non-randomly missing completion experiments. We report the impainting image in Fig. 6 and report the quantitative indices in Table. IV by ten different methods under different miss patterns with a miss rate of 40%. Each image in Fig. 6 shows a magnified map of a local patch (in the red box) for clarity.
We observe that the TD-based algorithms with ALS solver hardly deal with nonrandom missing cases, e.g., TT-ALS and TR-ALS. Surprisingly, the gradually generated TD model can greatly alleviate this problem and obtain acceptable results. Although the inpainting performance of MTNR-ALS is inferior to FCTN, which is optimized under the PAM framework, it is far superior to TR-ALS. The SiLRTC and HaLRTC have similar poor recovery performance in nonrandom missing cases, which implies the limitations of SNN. Most of the other methods have a relatively apparent residual stripe in the visualization results, while MTNR-ADMM has better visual effects. The quantitative indices of MTNR-ADMM in Table IV also achieve the most satisfactory outcomes, especially in the RMAR pattern of the second image, and the SSIM value is 0.0375 higher than the second one. These experiments suggest that MTNR-ADMM can still work excellently under other unrandom missing patterns.
V-D YaleFace Dataset Completion
In this section, we verify the performance of the proposed algorithm on Extended Yale Face Database B. The test data include the 5th, 10th, 15th, and 20th people in the first pose under 64 illumination conditions, and each image was resized to . Hence, we get four sets of data (Data-14)with a size of , and in the experiment, we reshape them to and randomly select 100% pixel values as the observation data. We set the observation level , and and of MTNR are set to 2 and , respectively. Note that when , MTNR is also easy to form a ring topology, but the ranks of different edges and mode arrangements are dynamically adjusted. The completion task for human faces is difficult since the features under different illuminations are more complicated, and the downsampled images have a lower local similarity.
We report the average completion performance of each data by different algorithms in Table V. Unlike the previous experimental results, we found that the inpainting results of FCTN-PAM are inferior to TT-ALS and TR-ALS. Two reasons are considered: 1) Excessive connections within factors will lead to over-parameterization of TD models, which is unfavourable for complicated data representation; 2) The identical edge rank will cause the inflexible representation since a small rank change will lead to a more significant storage cost adjustment. That implies that we don’t need to establish a connection between any factors, and the informative features of data’s different modes can be propagated along the edge. From the partial visualization results in Fig. 7, the global structure of the inpainting results of TRLRF and MTNR-ADMM are more intact.
V-E Video Completion
We also evaluate the proposed two algorithms on color video completion tasks and the test video sequence ”Bus” comes from the YUV dataset [12]. We selected the first 64 frames of the ”Bus” video and each frame with size . Then, the size of the test tensor is . Compared with the face data in the previous subsection, the adjacent frames of the video sequence are highly similar, and there is more redundant information. During the experiment, the ranks of TT, TR and FCTN based approaches are set as 20, 18, and 8, respectively. We focus on comparing MTNR and FCTN to highlight the advantage of the ATL algorithm on the rank determination of TD models. The and of MTNR are equivalent to 3 and . We randomly selected 40% of pixels as observation entries for the video completion task. We visualize the inpainting results of the 16th, 32th, 48th, and 64th frames in Fig. 8 and compare each frame’s PSNR and SSIM values in Fig. 9. From the visualization of video completion, the edges of the recovery images of SiLRTC and TR-WOPT are discontinuous and blurry. At the same time, MTNR-ADMM delivers the best completion results, and the inpainting images are clearer than the others. Notably, we only generated one low-rank component for MTNR in the experimental process, which corresponds to an FCTN model with ranks (16,3,12;16,3,13;3,3,3). Empirical evidence shows that the rank determination based on the ATL algorithm is more effective than the simple manual setting.
VI Conclusion
This paper proposed a generalized MTNR framework, which expresses a high-order tensor as a superposition of multiple low-rank components in a TN format. The topology of MTNR models is data-adaptive and robust to the mode arrangement of the data. We defined the basic multilinear operations of the tensor with TN format in theory and developed a novel ATL algorithm for latent factors learning of MTNR in practice. Then, two algorithms, called MTNR-ALS and MTNR-ADMM, are presented for the LTRC task. Extensive experimental results on synthetic data, benchmark color image, face dataset, and video demonstrate that the performance of MTNR is outperforming the state-of-the-art TD and rank-minimization based methods.
Acknowledgment
Postgraduate Research & Practice Innovation Program (KYCX21_0304) of Jiangsu Province, China. The authors would like to thank Andong Wang for his suggestions to improve this paper.
References
- [1] J.-L. Wang, T.-Z. Huang, X.-L. Zhao, T.-X. Jiang, and M. K. Ng, “Multi-dimensional visual data completion via low-rank tensor representation under coupled transform,” IEEE Transactions on Image Processing, vol. 30, pp. 3581–3596, 2021.
- [2] A. Cichocki, N. Lee, I. V. Oseledets, A. H. Phan, Q. Zhao, and D. P. Mandic, “Low-rank tensor networks for dimensionality reduction and large-scale optimization problems: Perspectives and challenges part 1.” arXiv preprint arXiv:1609.00893, 2016.
- [3] Y. Luo, D. Tao, K. Ramamohanarao, C. Xu, and Y. Wen, “Tensor canonical correlation analysis for multi-view dimension reduction,” in 2016 IEEE 32nd International Conference on Data Engineering (ICDE), 2016, pp. 1460–1461.
- [4] L. Zhu and X. Zhou, “Seismic moment tensor inversion using 3d velocity model and its application to the 2013 lushan earthquake sequence,” Physics and Chemistry of The Earth, vol. 95, pp. 10–18, 2016.
- [5] L. Zhang, L. Song, B. Du, and Y. Zhang, “Nonlocal low-rank tensor completion for visual data,” IEEE Transactions on Systems, Man, and Cybernetics, vol. 51, no. 2, pp. 673–685, 2021.
- [6] J. Liu, P. Musialski, P. Wonka, and J. Ye, “Tensor completion for estimating missing values in visual data,” in IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 35, no. 1, 2009, pp. 208–220.
- [7] Y. Ji, Q. Wang, X. Li, and J. Liu, “A survey on tensor techniques and applications in machine learning,” IEEE Access, vol. 7, pp. 162 950–162 990, 2019.
- [8] D. Tao, X. Li, W. Hu, S. Maybank, and X. Wu, “Supervised tensor learning,” in Fifth IEEE International Conference on Data Mining (ICDM’05), vol. 13, no. 1, 2005, pp. 450–457.
- [9] H. Lu, K. Plataniotis, and A. Venetsanopoulos, “Mpca: Multilinear principal component analysis of tensor objects,” IEEE Transactions on Neural Networks, vol. 19, no. 1, pp. 18–39, 2008.
- [10] N. D. Sidiropoulos, L. D. Lathauwer, X. Fu, K. Huang, E. E. Papalexakis, and C. Faloutsos, “Tensor decomposition for signal processing and machine learning,” IEEE Transactions on Signal Processing, vol. 65, no. 13, pp. 3551–3582, 2017.
- [11] L.-H. Lim and P. Comon, “Multiarray signal processing: Tensor decomposition meets compressed sensing,” Comptes Rendus Mecanique, vol. 338, no. 6, pp. 311–320, 2010.
- [12] L. Zhang, W. Wei, Q. Shi, C. Shen, A. van den Hengel, and Y. Zhang, “Accurate tensor completion via adaptive low-rank representation,” IEEE Transactions on Neural Networks, vol. 31, no. 10, pp. 4170–4184, 2020.
- [13] Z. Zhang and S. Aeron, “Exact tensor completion using t-svd,” IEEE Transactions on Signal Processing, vol. 65, no. 6, pp. 1511–1526, 2017.
- [14] H. Tan, G. Feng, J. Feng, W. Wang, Y.-J. Zhang, and F. Li, “A tensor based method for missing traffic data completion,” Transportation Research Part C-emerging Technologies, vol. 28, pp. 15–27, 2013.
- [15] P. Zhou, C. Lu, Z. Lin, and C. Zhang, “Tensor factorization for low-rank tensor completion,” IEEE Transactions on Image Processing, vol. 27, no. 3, pp. 1152–1163, 2018.
- [16] Y. Xu and W. Yin, “A block coordinate descent method for regularized multiconvex optimization with applications to nonnegative tensor factorization and completion,” Siam Journal on Imaging Sciences, vol. 6, no. 3, pp. 1758–1789, 2012.
- [17] E. Candès and B. Recht, “Exact matrix completion via convex optimization,” Communications of The ACM, vol. 55, no. 6, pp. 111–119, 2012.
- [18] B. Recht, M. Fazel, and P. A. Parrilo, “Guaranteed minimum-rank solutions of linear matrix equations via nuclear norm minimization,” Siam Journal on Control and Optimization, 2010.
- [19] J.-F. Cai, E. J. Candès, and Z. Shen, “A singular value thresholding algorithm for matrix completion,” Siam Journal on Optimization, vol. 20, no. 4, pp. 1956–1982, 2010.
- [20] Y. Chen, “Incoherence-optimal matrix completion,” IEEE Transactions on Information Theory, vol. 61, no. 5, pp. 2909–2923, 2015.
- [21] lldiko E. Frank and J. H. Friedman, “A statistical view of some chemometrics regression tools,” Technometrics, vol. 35, no. 2, pp. 109–135, 1993.
- [22] T. Zhang, “Analysis of multi-stage convex relaxation for sparse regularization,” Journal of Machine Learning Research, vol. 11, no. 35, pp. 1081–1107, 2010.
- [23] J. Bolte, S. Sabach, and M. Teboulle, “Proximal alternating linearized minimization for nonconvex and nonsmooth problems,” Mathematical Programming, vol. 146, no. 1, pp. 459–494, 2014.
- [24] C. Lu, J. Tang, S. Yan, and Z. Lin, “Nonconvex nonsmooth low rank minimization via iteratively reweighted nuclear norm,” IEEE Transactions on Image Processing, vol. 25, no. 2, pp. 829–839, 2016.
- [25] T. G. Kolda and B. W. Bader, “Tensor decompositions and applications,” Siam Review, vol. 51, no. 3, pp. 455–500, 2009.
- [26] S. Gandy, B. Recht, and I. Yamada, “Tensor completion and low-n-rank tensor recovery via convex optimization,” Inverse Problems, vol. 27, no. 2, p. 25010, 2011.
- [27] B. Romera-Paredes and M. Pontil, “A new convex relaxation for tensor completion,” in Advances in Neural Information Processing Systems 26, vol. 26, 2013, pp. 2967–2975.
- [28] M. E. Kilmer and C. D. Martin, “Factorization strategies for third-order tensors,” Linear Algebra and its Applications, vol. 435, no. 3, pp. 641–658, 2011.
- [29] M. E. Kilmer, K. S. Braman, N. Hao, and R. C. Hoover, “Third-order tensors as operators on matrices: A theoretical and computational framework with applications in imaging,” SIAM Journal on Matrix Analysis and Applications, vol. 34, no. 1, pp. 148–172, 2013.
- [30] C. Lu, J. Feng, Y. Chen, W. Liu, Z. Lin, and S. Yan, “Tensor robust principal component analysis with a new tensor nuclear norm,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 42, no. 4, pp. 925–938, 2020.
- [31] A. Bibi and B. Ghanem, “High order tensor formulation for convolutional sparse coding,” in 2017 IEEE International Conference on Computer Vision (ICCV), 2017, pp. 1790–1798.
- [32] A. Wang, C. Li, Z. Jin, and Q. Zhao, “Robust tensor decomposition via orientation invariant tubal nuclear norms,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. 4, 2020, pp. 6102–6109.
- [33] E. Kernfeld, M. Kilmer, and S. Aeron, “Tensor-tensor products with invertible linear transforms,” Linear Algebra and its Applications, vol. 485, pp. 545–570, 2015.
- [34] L. De Lathauwer, B. De Moor, and J. Vandewalle, “A multilinear singular value decomposition,” SIAM journal on Matrix Analysis and Applications, vol. 21, no. 4, pp. 1253–1278, 2000.
- [35] L. Yuan, C. Li, D. P. Mandic, J. Cao, and Q. Zhao, “Tensor ring decomposition with rank minimization on latent space: An efficient approach for tensor completion,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, no. 1, 2019, pp. 9151–9158.
- [36] A. Cichocki, D. Mandic, A.-H. Phan, C. Caiafa, G. Zhou, Q. Zhao, and L. D. Lathauwer, “Tensor decompositions for signal processing applications from two-way to multiway component analysis,” arXiv preprint arXiv:1403.4462, 2014.
- [37] L. D. Lathauwer, B. D. Moor, and J. Vandewalle, “A multilinear singular value decomposition,” SIAM Journal on Matrix Analysis and Applications, vol. 21, no. 4, pp. 1253–1278, 2000.
- [38] R. A. Harshman, “Foundations of the parafac procedure: Models and conditions for an “explanatory“ multi-model factor analysis,” vol. 16, pp. 1–84, 1970.
- [39] L. R. Tucker, “Some mathematical notes on three-mode factor analysis,” Psychometrika, vol. 31, no. 3, pp. 279–311, 1966.
- [40] S.-J. Ran, E. Tirrito, C. Peng, X. Chen, L. Tagliacozzo, G. Su, and M. Lewenstein, Tensor Network Contractions: Methods and Applications to Quantum Many-Body Systems, 2020.
- [41] J. C. Bridgeman and C. T. Chubb, “Hand-waving and interpretive dance: An introductory course on tensor networks,” Journal of Physics A, vol. 50, no. 22, p. 223001, 2017.
- [42] I. V. Oseledets, “Tensor-train decomposition,” SIAM Journal on Scientific Computing, vol. 33, no. 5, pp. 2295–2317, 2011.
- [43] Q. Zhao, G. Zhou, S. Xie, L. Zhang, and A. Cichocki, “Tensor ring decomposition,” arXiv preprint arXiv:1606.05535, 2016.
- [44] M. Hashemizadeh, M. Liu, J. Miller, and G. Rabusseau, “Adaptive tensor learning with tensor networks,” arXiv preprint arXiv:2008.05437, 2020.
- [45] R. Orús, “A practical introduction to tensor networks: Matrix product states and projected entangled pair states,” Annals of Physics, vol. 349, pp. 117–158, 2014.
- [46] Y.-B. Zheng, T.-Z. Huang, X.-L. Zhao, Q. Zhao, and T.-X. Jiang, “Fully-connected tensor network decomposition and its application to higher-order tensor completion,” in the AAAI Conference, 2021.
- [47] Y.-D. Kim, E. Park, S. Yoo, T. Choi, L. Yang, and D. Shin, “Compression of deep convolutional neural networks for fast and low power mobile applications,” arXiv preprint arXiv:1511.06530, 2015.
- [48] C. Li and Z. Sun, “Evolutionary topology search for tensor network decomposition,” in ICML 2020: 37th International Conference on Machine Learning, vol. 1, 2020, pp. 5947–5957.
- [49] C. J. Hillar and L.-H. Lim, “Most tensor problems are np-hard,” Journal of the ACM, vol. 60, no. 6, p. 45, 2013.
- [50] Z. Cheng, B. Li, Y. Fan, and Y. Bao, “A novel rank selection scheme in tensor ring decomposition based on reinforcement learning for deep neural networks,” in ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2020, pp. 3292–3296.
- [51] F. Sedighin, A. Cichocki, and A.-H. Phan, “Adaptive rank selection for tensor ring decomposition,” IEEE Journal of Selected Topics in Signal Processing, vol. 15, no. 3, pp. 454–463, 2021.
- [52] Q. Zhao, L. Zhang, and A. Cichocki, “Bayesian cp factorization of incomplete tensors with automatic rank determination,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 37, no. 9, pp. 1751–1763, 2015.
- [53] Z. Long, C. Zhu, J. Liu, and Y. Liu, “Bayesian low rank tensor ring for image recovery,” IEEE Transactions on Image Processing, vol. 30, pp. 3568–3580, 2021.
- [54] W. Wang, V. Aggarwal, and S. Aeron, “Efficient low rank tensor ring completion,” in 2017 IEEE International Conference on Computer Vision (ICCV), 2017, pp. 5698–5706.
- [55] E. Elhamifar and R. Vidal, “Sparse subspace clustering: Algorithm, theory, and applications,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 35, no. 11, pp. 2765–2781, 2013.
- [56] C. Lu, J. Feng, Z. Lin, T. Mei, and S. Yan, “Subspace clustering by block diagonal representation,” IEEE transactions on pattern analysis and machine intelligence, vol. 41, no. 2, pp. 487–501, 2018.
- [57] Y. Dai and Y. Wu, “Reweighted infrared patch-tensor model with both nonlocal and local priors for single-frame small target detection,” IEEE journal of selected topics in applied earth observations and remote sensing, vol. 10, no. 8, pp. 3752–3767, 2017.
- [58] C. Li, M. E. Khan, Z. Sun, G. Niu, B. Han, S. Xie, and Q. Zhao, “Beyond unfolding: Exact recovery of latent convex tensor decomposition under reshuffling,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. 04, 2020, pp. 4602–4609.
- [59] P. Comon, X. Luciani, and A. L. F. de Almeida, “Tensor decompositions, alternating least squares and other tales,” Journal of Chemometrics, vol. 23, pp. 393–405, 2009.
- [60] S. Boyd, N. Parikh, E. Chu, B. Peleato, and J. Eckstein, Distributed Optimization and Statistical Learning Via the Alternating Direction Method of Multipliers, 2011.
- [61] M. Levin and C. P. Nave, “Tensor renormalization group approach to two-dimensional classical lattice models.” Physical Review Letters, vol. 99, no. 12, pp. 120 601–120 601, 2007.
- [62] S. Ahmadi-Asl, A. Cichocki, A. H. Phan, M. G. Asante-Mensah, M. M. Ghazani, T. Tanaka, and I. Oseledets, “Randomized algorithms for fast computation of low rank tensor ring model,” Machine Learning: Science and Technology, vol. 2, no. 1, p. 011001, 2020.
- [63] S. Ruder, “An overview of gradient descent optimization algorithms,” arXiv preprint arXiv:1609.04747, 2016.
- [64] J. Yu, C. Li, Q. Zhao, and G. Zhao, “Tensor-ring nuclear norm minimization and application for visual : Data completion,” in ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2019, pp. 3142–3146.
- [65] Y. Wang, W. Yin, and J. Zeng, “Global convergence of admm in nonconvex nonsmooth optimization,” Journal of Scientific Computing, vol. 78, no. 1, pp. 29–63, 2019.
- [66] Y. Xu and W. Yin, “A globally convergent algorithm for nonconvex optimization based on block coordinate update,” Journal of Scientific Computing, vol. 72, no. 2, pp. 700–734, 2017.
- [67] J. Bolte, A. Daniilidis, and A. Lewis, “The łojasiewicz inequality for nonsmooth subanalytic functions with applications to subgradient dynamical systems,” SIAM Journal on Optimization, vol. 17, no. 4, pp. 1205–1223, 2007.
- [68] L. Yuan, J. Cao, X. Zhao, Q. Wu, and Q. Zhao, “Higher-dimension tensor completion via low-rank tensor ring decomposition,” in 2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), 2018, pp. 1071–1076.