Condition numbers for the truncated total least squares problem and their estimations
Qing-Le Meng1, Huai-An Diao2, Zheng-Jian Bai3
1
School of Mathematical Sciences, Xiamen University, Xiamen 361005, P.R. China. Email: [email protected]
2 Corresponding author.
School of Mathematics and Statistics, Northeast Normal University, No. 5268 Renmin
Street, Changchun 130024, P.R. China. Email: [email protected]
3 School of Mathematical Sciences and Fujian Provincial Key Laboratory on Mathematical Modeling & High Performance Scientific Computing, Xiamen University, Xiamen 361005, P.R. China. Email: [email protected]
The research of Z.-J. Bai is partially supported by the National Natural Science Foundation of China (No. 11671337) and the Fundamental Research Funds for the Central Universities (No. 20720180008).
Abstract.
In this paper, we present explicit expressions for the mixed and componentwise condition numbers of the truncated total least squares (TTLS) solution of under the genericity condition, where is a real data matrix and is a real -vector. Moreover, we reveal that normwise, componentwise and mixed condition numbers for the TTLS problem can recover the previous corresponding counterparts for the total least squares (TLS) problem when the truncated level of for the TTLS problem is . When is a structured matrix, the structured perturbations for the structured truncated TLS (STTLS) problem are investigated and the corresponding explicit expressions for the structured normwise, componentwise and mixed condition numbers for the STTLS problem are obtained. Furthermore, the relationships between the structured and unstructured normwise, componentwise and mixed condition numbers for the STTLS problem are studied. Based on small sample statistical condition estimation, reliable condition estimation algorithms are proposed for both unstructured and structured normwise, mixed and componentwise cases, which utilize the SVD of the augmented matrix . The proposed condition estimation algorithms can be integrated into the SVD-based direct solver for the small and medium size TTLS problem to give the error estimation for the numerical TTLS solution. Numerical experiments are reported to illustrate the reliability of the proposed condition estimation algorithms.
Keywords: Truncated total least squares, normwise perturbation, componentwise perturbation, structured perturbation, singular value decomposition, small sample statistical condition estimation.
AMS subject classifications: 15A09, 65F20, 65F35
1. Introduction
In this paper, we consider the following linear model
(1.1)
where the data matrix and the observation vector are both perturbed. When , the linear model (1.1) is overdetermined. To find a solution to (1.1), one may solve the following minimization problem:
(1.2)
where means the Frobenius matrix norm. This is the classical total least square (TLS) problem, which was originally proposed by Golub and Van Loan [11].
The TLS problem is often used for the linear model (1.1) when the augmented matrix is rank deficient, i.e., the small singular values of are assumed to be separated from the others. More interestingly, the truncated total least square (TTLS) method aims to solve the linear model (1.1) in the sense that the small singular values of are set to be zeros. For the discussion of the TTLS, one may refer to [27, §3.6.1] and [6, 8]. The TTLS problem arises in various applications such as linear system theory, computer vision, image reconstruction, system identification, speech and audio processing, modal and spectral analysis, and astronomy, etc. The overview of the TTLS can be found in [25].
Let be the predefined truncated level, where . The TTLS problem aims to solve the following problem:
(1.3)
where denotes the Euclidean vector norm or its induced matrix norm and is the best rank- approximation of in the Frobenius norm. The TTLS problem can be viewed as the regularized TLS (1.2) by truncating the small singular values of to be zero (cf.[6, 8]).
In order to solve (1.3), we first recall the singular value decomposition (SVD) of which is given by
(1.4)
where and are orthogonal matrices and is a real matrix with a vector on its diagonal and , where . Here, is a diagonal matrix with a vector on its diagonal and the superscript
“” takes the transpose of a matrix or vector. Then we have , where . Suppose the truncation level satisfies the condition
(1.5)
Define
(1.6)
If
(1.7)
then the TTLS problem is generic and the TTLS solution is given by [14]
(1.8)
Condition numbers measure the worst-case sensitivity of an input data to small perturbations. Normwise condition numbers for the TLS problem (1.2) under the genericity condition were studied in [1, 16], where SVD-based explicit formulas for normwise condition numbers were derived. The normwise condition number of the truncated SVD solution to a linear model as (1.1) was introduced in [2]. When the data is sparse or badly scaled, it is more suitable to consider the componentwise perturbations since normwise perturbations only measure the perturbation for the data by means of norm and may ignore the relative size of the perturbation on its small (or zero) entries (cf. [15]). There are two types of condition numbers in componentwise perturbations. The mixed condition number measures the errors in the output using norms and the input perturbations componentwise, while the componentwise condition number measures both the error in the output and the perturbation in the input componentwise (cf. [9]). The Kronecker product based formulas for the mixed and componentwise condition numbers to the TLS problem (1.2) were derived in [31, 4].
The corresponding componentwise perturbation analysis for the multidimensional TLS problem and mixed least squares-TLS problem can be found in [29, 30].
Gratton et al. in [14] investigated the normwise condition number for the TTLS problem (1.3). The normwise condition number formula and its computable upper bounds for the TTLS solution (1.3) were derived (cf. [14, Theorems 2.4-2.6]), which rely on the SVD of the augmented matrix . Since the normwise condition number formula for the TTLS problem (1.3) involves Kronecker product, which is not easy to compute or evaluate even for the medium size TTLS problem, the condition estimation method based on the power method [15] or the Golub-Kahan-Lanczos (GKL) bidiagonalization algorithm [10] was proposed to estimate the spectral norm of Fréchet derivative matrix related to (1.3). Furthermore, as point in [14], first-order perturbation bounds based on the normwise condition number can significantly improve the pervious normwise perturbation results in [7, 28] for (1.3).
As mentioned before, when the TTLS problem (1.3) is sparse or badly scaled, which often occurs in scientific computing, the conditioning based on normwise perturbation analysis may severely overestimate the true error of the numerical solution to (1.3). Indeed, from the numerical results for Example 5.1 in Section 5, the TTLS problem (1.3) with respect to the specific data and is well-conditioned under componentwise perturbation analysis while it is very ill-conditioned under normwise perturbation, which implies that the normwise relative errors for the numerical solution to (1.3) are pessimistic. In this paper, we propose the mixed and componentwise condition number for the TTLS problem (1.3) and the corresponding explicit expressions are derived, which can capture the true conditioning of (1.3) with respect to the sparsity and scaling for the input data. As shown in Example 5.1, the introduced mixed and componentwise condition numbers for (1.3) can be much smaller than the normwise condition number appeared in [14], which can improve the first-order perturbation bounds for (1.3) significantly. Furthermore, when the truncated level in (1.3) is selected to be , (1.3) reduces to (1.2). The normwised, mixed and componentwise condition numbers for the TTLS problem (1.3) are shown to be mathematically equivalent to the corresponding ones [1, 16, 31] for the untruncated case from their explicit expressions.
Structured TLS problems [17, 22, 25] had been studied extensively in the past decades. For structured TLS problems, it is suitable to investigate structured perturbations on the input data, because structure-preserving algorithms that preserve the underlying matrix structure can enhance the accuracy and efficiency of the TLS solution computation. Structured condition numbers for structured TLS problems can be found in [23, 4, 5] and references therein. In this paper, we introduce structured perturbation analysis for the structured TTLS (STTLS) problem. The explicit structured normwise, mixed and componentwise condition numbers for the STTLS problem are obtained, and their relationships corresponding to the unstructured ones are investigated.
The Kronecker product based expressions, for both unstructured and structured normwise, mixed and componentwise condition numbers of the TTLS solution in Theorems 2.1 and 2.2, involve higher dimensions and thus prevent the efficient calculations of these condition numbers. In practice, it is important to estimate condition numbers efficiently since the forward error for the numerical solution can be obtained via combining condition numbers with backward errors. In this paper, based on the small sample statistical condition estimation (SCE) [18], we propose reliable condition estimation algorithms for both unstructured and structured normwise, mixed and componentwise condition numbers of the TTLS solution, which utilize the SVD of to reduce the computational cost. Furthermore, the proposed condition estimation algorithms can be integrated into the SVD-based direct solver for the small or medium size TTLS problem (1.3). Therefore, one can obtain the reliable forward error estimations for the numerical TTLS solution after implementing the proposed condition estimation algorithms. The main computational cost in condition number estimations for (1.3) is to evaluate the directional derivatives with respect to the generated direction during the loops in condition number estimations algorithms. We point out that the power method [15] for estimating the normwise condition number in [14] needs to evaluate the directional derivatives twice in one loop. However, only evaluating direction derivative once is needed in the loop of Algorithms 1 to 3. Therefore, compared with the normwise condition number estimation algorithm proposed in [14], our proposed condition number estimations algorithms in this paper are more efficient in terms of the computational complexity, which are also applicable for estimating the componentwise and structured perturbations for (1.3). For recent SCE’s developments for (structured) linear systems, linear least squares and TLS problem, we refer to [20, 19, 21, 5] and references therein.
The rest of this paper is organized as follows. In Section 2 we review pervious perturbation results on the TTLS problem and derive explicit expressions of the mixed and componentwise condition numbers. The structured normwise, mixed and componentwise condition numbers are also investigated in Section 2, where the relationships between the unstructured normwise, mixed and componentwise condition numbers for (1.3) with the corresponding structured counterparts are investigated. In Section 3 we establish the relationship between normwise, componentwise and mixed condition numbers for the TTLS problem and the corresponding counterparts for the untruncated TLS. In Section 4 we are devoted to propose several condition estimation algorithms for the normwise, mixed and componentwise condition numbers of the TTLS problem. Moreover, the structured condition estimation is considered. In Section 5, numerical examples are shown to illustrate the efficiency and reliability of the proposed algorithms and report the perturbation bounds based on the proposed condition number. Finally, some concluding remarks are drawn in the last section.
2. Condition numbers for the TTLS problem
In this section we review previous perturbation results on the TTLS problem. The explicit expressions of the mixed and componentwise condition numbers for the TTLS problem are derived. Furthermore, for the structured TTLS problem, we propose the normwise, mixed and componentwise condition numbers, where explicit formulas for the corresponding counterparts are derived. The relationships between the unstructured normwise, mixed and componentwise condition numbers for (1.3) with the corresponding structured counterparts are investigated. We first introduce some conventional notations.
Throughout this paper, we use the following notation. Let be the vector -norm or its induced matrix norm. Let be the identity matrix of order . Let be the -th column vector of an identity matrix of an appropriate dimension. The superscripts “” and“” mean the inverse and the Moore-Penrose inverse of a matrix respectively. The symbol “” means componentwise multiplication of two conformal dimensional matrices. For any matrix , let , where denote the absolute value of . For any two matrices , represents for all and . For any , we define by
Let be a column vector obtained by stacking the columns of on top of one another. For a vector , let , where for and . The symbol “” means the Kronecker product and is a permutation matrix defined by
(2.1)
Given the matrices , , and , and with appropriate dimensions, we have the following propertes of the Kronecker product and vec operator [13]:
(2.2)
2.1. Preliminaries
In this subsection, we recall the definition of absolute normwise condition number of the TTLS solution defined by (1.3) (cf. [14]). The absolute normwise condition number of in (1.3) is defined by
(2.3)
where the function is given by
(2.4)
Let the SVD of be given by (1.4). If the truncation level satisfies the conditions (1.5) and (1.7), then the explicit expression of is given by [14, Theorem 2.4]
(2.5)
where
(2.6)
with
Please be noted the the dimension of may be large even for medium size TTLS problems. The explicit formula given by (2.5) involves the computation of the spectral norm of . Hence, upper bounds for is obtain in [14, §2.4], which only rely on the singular values of and . When the data is sparse or badly scaled, the normwise condition number may not reveal the conditioning of (1.3), since normwise perturbations ignore the relative size of the perturbation on its small (or zero) entries. Therefore, it is more suitable to consider the componentwise perturbation analysis for (1.3) when the data is sparse or badly scaled. In the next subsection, we shall introduce the mixed and componentwise condition number for (1.3).
In [14, §2.3], if both and the full SVD of are available, then one may compute by using the power method [15, Chap. 15] to or the Golub-Kahan-Lanczos (GKL) bidiagonalization algorithm [10] to , where only the matrix-vector product is needed. However, as pointed in the introduction part, the normwise condition number estimation algorithm in [14] are devised based on the power method [15], which needs to evaluate the matrix-vector products and in one loop for some suitable dimensional vectors and . In Section 4, SCE-based condition estimation algorithms for (1.3) shall be proposed, where in one loop we only need to compute the directional derivative but the matrix-vector product is not involved. Therefore, compared with normwise condition number estimation algorithm in [14], SCE-based condition estimation algorithms in Section 4 are more efficient.
2.2. Mixed and componentwise condition numbers
In Lemma 2.1 below, the first order perturbation expansion of with respect to the perturbations of the data and is reviewed, which involves the Kronecker product. In order to avoid forming Kronecker product explicitly in the explicit expression for the directional derivative of , we derive the corresponding equivalent formula (2.7) in Lemma 2.2. Furthermore, the directional derivative (2.7) can be used to save computation memory of SCE-based condition estimation algorithms in Section 4.
Lemma 2.1.
[14, Theorem 2.4]
Let the SVD of the augmented matrix be given by (1.4).
Suppose is a truncation level such that and . If with sufficiently small, then, for the TTLS solution of and the TTLS solution of , we have
Lemma 2.2.
Under the same assumptions as in Lemma 2.1, if with sufficiently small, then the directional derivative of at in the direction is given by
When the data is sparse or badly-scaled, it is more suitable to adopt the componentwise perturbation analysis to investigate the conditioning of the TTLS problem. In the following definition, we introduce the relative mixed and componentwise condition numbers for the TTLS problem.
Definition 2.1.
Suppose the truncation level is chosen such that and . The mixed and componentwise condition numbers for the TTLS problem (1.3) are defined as follows:
In the following theorem, we give the Kronecker product based explicit expressions of and .
Theorem 2.1.
Suppose the truncation level is chosen such that and . Then the mixed and componentwise condition numbers and defined in Definition 2.1 for the TTLS problem (1.3) can be characterized by
(2.13a)
(2.13b)
Proof. Let , where and . For any
, it follows from that
where is a diagonal matrix such that =sign for . Using (2.2) we have
Therefore, we derive (2.13a). One can use the similar argument to obtain (2.13b).
∎
Remark 2.1.
Based on (2.3) and (2.5), the relative normwise condition number for the TTLS problem (1.3) can be defined and has the following expression
(2.15)
Using the fact that
it is easy to see that
(2.16)
From Example 5.1, we can see and can be much smaller than when the data is sparse and badly scaled. Therefore, one should adopt the mixed and componentwise condition number to measure the conditioning of (1.3) instead of the normwise condition number when is spare or badly scaled. However, since the explicit expressions of and are based on Kronecker product, which involves large dimensional computer memory to form them explicitly even for medium size TLS problems, it is necessary to propose efficient and reliable condition estimations for and , which will be investigated in Section 4.
In [3, 17, 22], the structured TLS (STTLS) problem has been studied extensively. Hence, it is interesting to study the structured perturbation analysis for the STTLS problem. In the following, we propose the structured normwise, mixed and componentwise condition numbers for the STTLS problem, where is a linear structured data matrix. Assume that is a linear subspace which consists of a class of basis matrices. Suppose there are () linearly independent matrices in , where are matrices of constants, typically 0’s and 1’s. For any , there is a uniques vector such that
(2.17)
In the following, we study the sensitivity of the STTLS solution to perturbations on the data and , which is defined by
(2.18)
where is the unique solution to the STTLS problem (1.3) and (2.17).
Definition 2.2.
Suppose the truncation level is chosen such that and . The absolute structured normwise, mixed and componentwise condition number for the STTLS problem (1.3) and (2.17) are defined as follows:
In the following lemma, we provide the first order expansion of the STTLS solution with respect to the structured perturbations on and on , which help us to derive the structured condition number expressions for the STTLS problem (1.3) and (2.17). In view of the fact that , we can prove the following lemma from Lemma 2.1. The detailed proof is omitted here.
Lemma 2.3.
Under the same assumptions of Lemma 2.1, if with sufficiently small, then, for the STTLS solution of and the STTLS solution of , we have
where .
The following theorem concerns with the explicit expressions for the structured normwise, mixed and componentwise condition numbers , , and defined in Definition 2.2 when can be expressed by (2.17). Since the proof is similar to Theorem 2.1, we omit it here.
Theorem 2.2.
Suppose the truncation level is chosen such that and . The absolute structured normwise, mixed and componentwise condition numbers , , and defined in Definition 2.2 for the STTLS problem (1.3) and (2.17) can be characterized by
Remark 2.2.
Based on Definition 2.2 and Theorem 2.2, the relative normwise condition number for the STTLS problem (1.3) and (2.17) can be defined and has the following expression
In Example 5.2, we can see can be much smaller than . Hence, structured condition number can explain that structure-preserving algorithms can enhance the accuracy of the numerical solution, since structure-preserving algorithms preserve the underlying matrix structure.
In the following proposition, we show that, when is a linear structured matrix defined by (2.17), the structured normwise, mixed and componentwise condition numbers , , and are smaller than the corresponding unstructured condition numbers , and respectively.
Proposition 2.1.
Using the notations above, we have . Moreover, if , then we have
Proof.
From [23, Theorem 4.1], the matrix is column orthogonal. Hence, and it is not difficult to see that
by comparing their expressions. Using the monotonicity of infinity norm, it can be obtained that
therefore we prove that and can be proved similarly. ∎
3. Revisiting condition numbers of the untruncated TLS problem
In this section, we investigate the relationship between normwise, componentwise and mixed condition numbers for the TTLS problem and the previous corresponding counterparts for the untruncated TLS. In the following, let be the untruncated TLS solution to (1.2). First let us review previous results on condition numbers for the untruncated TLS problem.
Let be the smallest singular value of . As noted in [11], if
(3.1)
then the TLS problem (1.2) has a unique TLS solution
Let be a linear function of the TLS solution , where is a fixed matrix with .
We define the mapping
(3.2)
As in [1], the absolute normwise condition number of can be characterized by
(3.3)
where
(3.4)
Later, in [16], an equivalent expression of was given by
Recall that is given by (2.5). The relationship between the upper bound for and the corresponding counterpart for was studied in [14, §2.5]. The following theorem shows the equivalence of and .
Theorem 3.1.
For the untruncated TLS problem (1.2), the explicit expression of given by (2.5) with is equivalent to that of given by (3.5)
Proof.
When , it is easy to see that . Also we have
(3.6)
where , , , and
(3.7)
When , under the genericity condition (3.1), the following identities hold for the TLS solution (cf. [11])
(3.8)
where has the partition in (1.6) and given by (1.8).
Thus it is not difficult to see that
(3.9)
Since
we know that
thus, it can be verified that
(3.10)
Combining (3.9) and (3.10) with the expression of given by (3.6), we have
(3.11)
This, together with , yields
where is defined in (3.5). Therefore, when the expression of given by (2.5) is reduced to
The proof is complete.
∎
In [31], Zhou et al. defined and derived the relative mixed and componentwise condition numbers for the untruncated TLS problem (1.2) as follows: Let , where and are the perturbations of and respectively. When the norm is small enough, for the TLS solution of and the TTLS solution of , we have
(3.12)
(3.13)
where
and are the -th column of and respectively.
Recently, in [4], Diao and Sun defined and gave mixed and componentwise condition numbers for the linear function as follows.
where . Moreover, when , the expressions of and were equivalent to the explicit expressions of and given by (3.12)–(3.13) (cf.[4, Theorem 3.2]).
In the following theorem, we prove that, when , and given by Theorem 2.1 are reduced to those of and , respectively.
Theorem 3.2.
Using the notations above, when , we have and .
Proof. From the proof of [5, Lemma 2], for given by (3.4) we have
where and are defined in (3.6) and (1.6), respectively. Thus, when , using (3.11) and (3.7) we have
(3.14)
where
(3.15)
Partition and as follows:
(3.16)
By comparing the expressions of and with , we only need to show that
(3.17)
where
Using the SVD of in (1.4), the partitions of , in (1.6) and in (2.5), it follows that
(3.18)
where is the -th column of . We note that
(3.19)
where is the first column of . From the SVD of and (3.8), it is easy to check that
(3.20)
Substituting (3.18), (3.8) and (3.20) into the expression of and yields
The structured condition numbers for the untruncated TLS problem with linear structures were studied by
Li and Jia in [23]. For the structured matrix defined by (2.17), denote
(3.26)
where is defined by (3.4). The structured mixed condition number is characterized
as [23]
(3.27)
In [23, Theorem 4.3], Li and Jia proved that and , where is given by (3.12). Hence we have the following proposition. Indeed, we prove that, when , the expression of given by (3.27) is reduced to that of in Theorem 2.2.
Proposition 3.1.
Using the notations above, when , we have .
4. Small sample statistical condition estimation
Based on small sample statistical condition estimation (SCE), reliable condition estimation algorithms for both unstructured and structured normwise, mixed and componentwise are devised, which utilize the SVD of the augmented matrix . In the following, we first review the basic idea of SCE. Let be a differentiable function. For the input vector , we want to estimate the sensitivity of the output with respect to small perturbation on , where is a unit vector and is a small positive number. The Taylor expansion of at is given by
where is the gradient of at
. Neglecting the second and higher order terms of we have
from which we conclude that the local sensitivity can be measured by
. Let the Wallis factor be given by [18]
As in [18], if is selected uniformly and randomly from the unit -sphere
(denoted as ), then the expected value is . In practice, the Wallis factor can be
approximated accurately by [18]
(4.1)
Therefore, the following quantity
can be used as a condition estimator for
with high probability for the function at . For example, let , which indicates the accuracy of the estimator, it is shown that
In general, we are interested in
finding an estimate that is accurate to a factor of 10 (). The accuracy of the condition estimator can enhanced by using multiple samples of , denoted . The -sample condition estimation is given by
where the matrix is orthonormalized after are selected uniformly and
randomly from . Usually, at most two or three samples are sufficient for high accuracy. For example, the accuracies of and are given by [18]
As an illustration, for and , the estimator has probability , which is within a relative factor of the true condition number .
The above results can be easily extended to vector-valued or matrix-valued functions.
4.1. Normwise perturbation analysis
In this subsection, we propose an algorithm for the normwise condition
estimation of the TTLS problem (1.3) based on SCE. The input data of Algorithm 1 includes the matrix , the vector , the SVD of , and the computed solution . The output includes the condition vector and the estimated relative condition number .
Algorithm 1 Small sample condition estimation for the TTLS problem under normwise perturbation analysis
1.
Generate matrices with entries being in , the standard Gaussian distribution. Orthonormalize the following matrix
to obtain via the modified Gram-Schmidt orthogonalization process.
Set
where and are defined in (2.7) with . Estimate the absolute condition vector
Here, for any vector , and .
4.
Compute the normwise condition number as follows,
where
Next, we give some remarks on the computational cost of Algorithm 1.
In Step 1, the modified Gram-Schmidt orthogonalization process [12] is adopted to form an orthonormal matrix
and the total flop count is about .
The cost associated with step 3 is about flops that is mainly from computing the directional derivative in Lemma 2.2. The last step needs flops.
We note that generates a good condition estimation. In this case, the total cost of Algorithm 1 is , which does not exceed the cost of computing the SVD of and . Furthermore, the directional derivative (4.2) only be computed once in one loop of Algorithm 1. On the contrary, Gratton et. al [14] proposed the normwise condition number estimation algorithm through using the power method [15], which needs to evaluate the
matrix-vector products and in one loop for some suitable dimensional vectors and . Therefore, Algorithm 1 is more efficient compared with the normwise condition number estimation method in [14].
4.2. Componentwise perturbation analysis
If the perturbation in the input data is measured componentwise rather
than by norm, it may help us to measure the sensitivity of
a function more accurately [26]. The SCE method can also be used to measure the sensitivity of componentwise perturbations [18], which may give a more realistic indication
of the accuracy of a computed solution than that from the normwise condition number. In componentwise perturbation analysis, for a perturbation of , we assume that . Therefore, the perturbation can be rewritten as with
and each entry of being in the interval . Based on the above observations, we can obtain
a componentwise sensitivity estimate of the solution of the TTLS problem (1.3) as follows. The detailed descriptions are given in Algorithm 2, which is a modification of Algorithm 1 directly.
Algorithm 2 Small sample condition estimation for
the TTLS problem under componentwise perturbation analysis
1.
Generate matrices with entries being in . Orthonormalize the following matrix
to obtain via the modified Gram-Schmidt orthogonalization process.
Set
For calculate by (4.2).
Estimate the absolute condition vector
4.
Set the relative condition vector
.
Compute the mixed and componentwise condition estimations
and as follows,
To estimate the mixed and componentwise condition numbers via Algorithm 2, we only need additional computational cost flops comparing with Algorithm 1.
4.3. Structured perturbation analysis
For the structured TLS problem, it is reasonable to consider the case that the perturbation has the same structure as .
Suppose the matrix takes the form of (2.17), i.e., . Here, are linearly independent matrices of a linear subspace .
It is easy to see that
where and .
For
we have ,
where the Matlab-routine notation denotes a Toeplitz matrix having as its first column and as its first row, and , where . This means that a Toeplitz matrix can be obtained by taking and letting be the map
The SCE method maintains the desired matrix structure by working with the perturbations of and in the linear space of . This produces only slight changes in the SCE algorithm. By simply generating and randomly instead of and as in Algorithm 1, we obtain an algorithm to estimate the condition of a composite map . We summarize the structured normwise condition estimation in Algorithm 3, which also includes the structured componentwise condition estimation. The computational cost of Algorithm 3 is reported in Table 1.
Algorithm 3 Small sample condition estimation for the STTLS problem under structured perturbation analysis
1.
Generate matrices with
entries in where and . Orthonormalize the following matrix
to obtain an orthonormal matrix by using the modified Gram-Schmidt orthogonalization process.
Set
In this section, we present some numerical examples to illustrate the reliability of
the SCE for the TTLS problem (1.3). For a given TTLS problem, the TTLS solution with truncation level can be computed by utilizing the SVD of and (3.8). The corresponding exact condition numbers are computed by their explicit expressions associated with the given data . All the sample number in Algorithms 1 to 3 are set to be .
All the numerical experiments are carried out on Matlab R2019b with the machine epsilon under Microsoft Windows 10.
Example 5.1.
Let
where .
In Table 2, we compare our SCE-based estimations , and from Algorithms 1 and 2 with the corresponding exact condition numbers for Example 5.1. The symbol in Table 2 means the condition numbers , and are not defined for the truncation level . From the numerical results listed in Table 2, it is easy to find that the normwise condition number defined by (2.15) are much greater than results of mixed and componentwise condition numbers and given by Theorem 2.1. The (untruncated) TTL problem (1.3) is well conditioned under componentwise perturbations regardless of the choice of and . Compared with the normwise condition number, the mixed and componentwise condition numbers may capture the true conditioning of this TTLS problem. We also observe that the SCE-based condition estimations can provide reliable estimations. Moreover,
numerical results of the untruncated mixed and componentwise condition numbers and given in Theorem 2.1 are equal to corresponding values of the untruncated ones and given by (3.12)–(3.13), respectively. The similar conclusion can be drawn for comparing and for .
Table 2. Comparisons of exact normwise, mixed and componentwise condition numbers with SCE-based condition estimations for Example 5.1 via Algorithms 1 and 2 with .
3
Example 5.2.
[27]Let the data matrix and the observation vector be given by
Since the first - singular values of the augmented matrix are equal and larger than the -th singular value , the truncated level can only be . It is clear that is a Toeplitz matrix.
A condition estimation is said to be reliable if the estimations fall within one tenth to ten times of the corresponding exact condition numbers (cf. [15, Chapter 15]). Table 3 displays the numerical results for Example 5.2 by choosing from to . From Table 3, we can conclude that Algorithms 3 can provide reliable mixed and componentwise condition estimations for this specific Toeplitz matrix and the observation vector , while the normwise condition estimation may seriously overestimate the true relative normwise condition number. We also see that the unstructured mixed and componentwise condition numbers , given by Theorem 2.1 are not smaller than the corresponding structured ones , shown in Theorem 2.2, which is consistent with Proposition 2.1. Numerical values of the structured normwise, mixed and componentwise condition number are smaller than the corresponding counterparts.
Table 3. Comparisons of true structured normwise, mixed and componentwise condition numbers with SCE-based condition estimations for Example 5.2 via Algorithms 3 with the truncated level and .
Example 5.3.
This test problem comes from [14, §3.2]. The augmented matrix are builded by using the SVD . Here, is an arbitrary orthogonal matrix with the size of , and be a diagonal matrix with equally spaced singular values in . The matrix is generated as following: compute the QR decomposition of the matrix
with the Q-factor , where and are random matrices, and are normalized random vectors. Here, is truncation level. Then we set to be an orthogonal matrix that commutes the first and last rows of . It is easy to verify that and . In this test, we take , , , and . The perturbation matrices and of and are generated as follows:
(5.1)
where and are random matrices
whose entries are uniformly distributed in the open interval , represents the magnitude of the perturbation.
We note that both and are componentwise perturbations on and respectively. In order to illustrate the validity of these estimators and
via Algorithms 1 and
2, we define
the normwise, mixed and componentwise over-estimation ratios as follows
Typically the ratios
in are acceptable [15, Chap. 15].
(a) Normwise over-estimation ratios
(b) Mixed over-estimation ratios
(c) Componentwise over-estimation ratios
Figure 1. SCE for the TLS problem under unstructured componentwise perturbations with 1000 samples for Example 5.3.
Figure 1 displays the numerical results for Example 5.3, where we generate 1000 random samples .
From Figure 1, we see that may seriously overestimate the true relative normwise error. All the normwise over-estimation ratios of 1000 samples is more than , and the mean value of these estimations is . All elements of the mixed over-estimation ratios of 1000 samples are within and the mean value of these estimations is . There are only 6 entries of the componentwise over-estimation ratios are greater than 10 and the mean value of these estimations is . The maximal values of the normwise, mixed and componentwise over-estimation vector are and , while their corresponding minimum are accordingly. Therefore the mixed and componentwise condition estimations and are reliable.
Example 5.4.
Let the Toeplitz matrix and the vector are defined in Example 5.2, where . We generate 1000 structured componentwise perturbations and 1000 unstructured componentwise perturbations , where is a random Toeplitz matrix and is a random matrix whose entries are uniformly distributed in the open internal . And , where is a random matrix with components uniformly distributing in the open interval .
For Example 5.4, we use , and to denote the structured normwise, mixed and componentwise over-estimation ratios corresponding to structured componentwise perturbations of and , and , and are unstructured normwise, mixed and componentwise over-estimation ratios corresponding to unstructured componentwise perturbations of and , where
Figure 2 displays the numerical results for Example 5.4 with and . Here, in Figure 2(A)–Figure 2(C), the symbol “+” in the blue color denote the numerical values of , and corresponding to 1000 structured perturbations
while the symbol “*” in the red color denote the numerical values of , and corresponding to 1000 unstructured perturbations.
From Figure 2, we observe that the mixed and componentwise condition estimations , , and are reliable, while the structured normwise condition estimation may seriously over-estimate the true relative normwise error. Furthermore, we can also conclude that the over-estimation ratios associated with the structured mixed and component condition numbers are smaller than the unstructured counterparts in most cases, which are consistent with the conclusion in Proposition 2.1. The mean values of , , , and of samples are , , , and respectively. Moreover, all these unstructured and structured mixed and componentwise condition over-estimation ratios are with , which indicate mixed and componentwise condition estimations are reliable.
(a) Unstructured and structured normwise over-estimation ratios
(b) Unstructured and structured mixed over-estimation ratios
(c) Unstructured and structured componentwise over-estimation ratios
Figure 2. SCE for the TLS problem under unstructured and structured componentwise perturbations with 1000 unstructured and structured perturbations for Example 5.4.
Example 5.5.
This example comes from the model that restructures the image named Shepp-Logan “head phantom” (Shepp and Logan 1974) by the TTLS technique, which is widely used in inverse scattering studies. In fact, the TTLS method has been used to study the ultrasound inverse scattering imaging [24]. Here, we utilize the MATLAB file “paralleltomo.m” from the testprobs suite111Netlib: http://www.netlib.org/numeralgo/ or GitHub: https://github.com/jakobsj/AIRToolsII. to create parallel-beam CT test problem and obtain the exact phantom. The input parameters of “paralleltomo.m” are set to be , , and , hence we can obtain a -by- matrix and -by- right-hand vector . The 500 perturbations and of and are generated as in (5.1).
Table 4 lists the condition estimations , and with truncation level and the corresponding relative errors with respect to different magnitude of the perturbation for Example 5.5. From Table 4, we can see that the relative errors are bounded by the product of and corresponding condition estimations, which means that the proposed condition estimations can give reliable error bounds. The componentwise condition estimation is much larger than
and since the minimum absolute component of the TTLS solution is too small, which is order of . Hence the component of the TTLS solution in the sense of the tiny magnitude is very sensitive to small perturbations on the underlying component of .
Table 4. Comparison of the relative errors and SCE-based estimations by Algorithms 1 and 2 under 500 perturbations with different perturbation magnitudes for Example 5.5.
6. Concluding remarks
In this paper, we study the mixed and componentwise condition numbers of the TTLS problem under the genericity condition. We also consider the structured condition estimation for the STTLS problem and investigate the relationship between the unstructured condition numbers and the corresponding structured counterparts. When the TTLS problem degenerates the untrunctated TLS problem, we prove that condition numbers for the TTLS problem can recover the previous condition numbers for the TLS problem from their explicit expressions. Based on SCE, normwise, mixed and componentwise condition estimations algorithms are proposed for the TTLS problem, which can be integrated into the SVD-based direct solver for the TTLS problem. Numerical examples indicate that, in practice, it is better to adopt the componentwise perturbation analysis for the TTLS problem and the proposed algorithms are reliable, which provide posterior error estimations of high accuracy. The results in this paper can be extended to the truncated singular value solution of a linear ill-posed problem [2]. We will report our progresses on the above topic elsewhere in the future.
References
[1]M. Baboulin and S. Gratton, A contribution to the conditioning of the total least-squares
problem, SIAM J. Matrix Anal. Appl., 32(3) (2011), pp. 685–699.
[2]E. H. Bergou, S. Gratton, and J. Tshimanga, The exact condition number of the truncated singular value solution of a linear ill-posed problem, SIAM J. Matrix Anal. Appl., 35(3) (2014), pp. 1073–1085.
[3]A. Beck and A. Ben-Tal, A global solution for the structured total least squares problem with
block circulant matrices, SIAM J. Matrix Anal. Appl., 27(1) (2005), pp 238–255.
[4]H. Diao and Y. Sun, Mixed and componentwise condition numbers for a linear function of the solution of the total least squares problem, Linear Algebra Appl., 544 (2018), pp. 1–29.
[5]H. Diao, Y. Wei, and P. Xie, Small sample statistical condition estimation for the total least
squares problem, Numer. Algorithms, 75(2) (2017), pp. 435–455.
[6]R. D. Fierro and J. R. Bunch, Collinearity and total least squares, SIAM J. Matrix Anal. Appl., 15(4) (1994), pp. 1167–1181.
[7]R. D. Fierro and J. R. Bunch, Perturbation theory for orthogonal projection methods with
applications to least squares and total least squares, Linear Algebra Appl., 234(2) (1996), pp. 71–96.
[8]R. D. Fierro, G. H. Golub, P. C. Hansen, and D. P. O’Leary, Regularization by truncated total least squares, SIAM J. Sci. Comput., 18(4) (1997), pp. 1223–1241.
[9]I. Gohberg and I. Koltracht, Mixed, componentwise, and structured condition numbers, SIAM J. Matrix Anal. Appl., 14(3) (1993), pp. 688–704.
[10]G. Golub and W. Kahan, Calculating the singular values and pseudo-inverse of a matrix,
J. SIAM Ser. B Numer. Anal., 2(2) (1965), pp. 205–224.
[11]G. H. Golub and C. F. Van Loan, An analysis of the total least squares problem, SIAM J. Numer. Anal., 17(6) (1980), pp. 883–893.
[12]G. H. Golub and C. F. Van Loan, Matrix Computations, 4th ed., Johns Hopkins University Press, Baltimore, MD, 2012.
[13]A. Graham, Kronecker Products and Matrix Calculus: with Applications,
Ellis Horwood, London, 1981.
[14]S. Gratton, D. Titley-Peloquin, and J. T. Ilunga, Sensitivity and conditioning of the truncated total least squares
solution, SIAM J. Matrix Anal. Appl.,
34(3) (2013), pp. 1257–1276.
[15]N. J. Higham, Accuracy and Stability of Numerical Algorithms, 2nd ed., SIAM, Philadelphia, PA, 2002.
[16]Z. Jia and B. Li, On the condition number of the total least squares problem, Numer. Math., 125(1) (2013), pp. 61–87.
[17]J. Kamm and J. G. Nagy, A total least squares method for Toeplitz systems of equations, BIT, 38(3) (1998), pp. 560–582.
[18]C. S. Kenney and A. J. Laub, Small-sample statistical condition estimates for general matrix
functions, SIAM J. Sci. Comput., 15(1) (1994), pp. 36–61.
[19]C. S. Kenney, A. J. Laub, and M. S. Reese, Statistical condition estimation for linear least squares, SIAM J. Matrix Anal. Appl., 19(4) (1998), pp. 906–923.
[20]C. S. Kenney, A. J. Laub, and M. S. Reese, Statistical condition estimation for linear systems, SIAM J. Sci. Comput., 19(2) (1998), pp. 566–583.
[21]A. J. Laub and J. Xia, Applications of statistical condition estimation to the solution of linear systems, Numer. Linear Algebra Appl., 15 (2008), pp. 489–513.
[22]P. Lemmerling and S. Van Huffel, Analysis of the structured total least squares problem for
Hankel/Toeplitz matrices, Numer. Algorithms, 27(1) (2001), pp. 89–114.
[23]B. Li and Z. Jia, Some results on condition numbers of the scaled total least squares
problem, Linear Algebra Appl, 435(3) (2011), pp. 674–686.
[24]C. Liu, Y. Wang and P. Heng, A comparison of truncated total least squares with Tikhonov regularization in imaging by ultrasound inverse scattering, Phy. in Med. Biol., 48(15) (2003), pp. 2437–2451.
[25]I. Markovsky and S. Van Huffel, Overview of total least-squares methods,
Signal Processing, 87(10) (2007), pp. 2283–2302.
[26]R. D. Skeel, Scaling for numerical stability in Gaussian elimination, J. ACM, 26(3) (1979), pp. 494–526.
[27]S. Van Huffel and J. Vandewalle, The Total Least Squares Problem: Computational Aspects and
Analysis, SIAM, Philadelphia, PA, 1991.
[28]M. Wei, The analysis for the total least squares problem with more than one
solution, SIAM J. Matrix Anal. Appl., 13(3) (1992), pp. 746–763.
[29]B. Zheng, L. Meng, and Y. Wei, Condition numbers of the multidimensional total least squares problem, SIAM J. Matrix Anal. Appl., 38 (2017), pp. 924–948.
[30]B. Zheng and Z. Yang, Perturbation analysis for mixed least squares-total least squares problems, Numer Linear Algebra Appl., 26(4) (2019), e2239.
[31]L. Zhou, L. Lin, Y. Wei, and S. Qiao, Perturbation analysis and condition numbers of scaled total least
squares problems, Numer. Algorithms, 51(3) (2009), pp. 381–399.