Block -circulant preconditioners for parabolic optimal control problems
Abstract
In this work, we propose a class of novel preconditioned Krylov subspace methods for solving an optimal control problem of parabolic equations. Namely, we develop a family of block -circulant based preconditioners for the all-at-once linear system arising from the concerned optimal control problem, where both first order and second order time discretization methods are considered. The proposed preconditioners can be efficiently diagonalized by fast Fourier transforms in a parallel-in-time fashion, and their effectiveness is theoretically shown in the sense that the eigenvalues of the preconditioned matrix are clustered around , which leads to rapid convergence when the minimal residual method is used. When the generalized minimal residual method is deployed, the efficacy of the proposed preconditioners are justified in the way that the singular values of the preconditioned matrices are proven clustered around unity. Numerical results are provided to demonstrate the effectiveness of our proposed solvers.
Keywords: Toeplitz, skew/circulant matrices, -circulant matrices, preconditioners, parallel-in-time
AMS Subject Classification: 65F08, 65F10, 65M22, 15B05
1 Introduction
In recent years, there has been a growing interest in analyzing and solving optimization problems constrained by parabolic equations. We refer to [25, 14, 34, 6] and the references therein for a comprehensive overview.
In this work, we are interested in solving the distributed optimal control model problem. Namely, the following quadratic cost functional is minimized:
(1) |
subject to a parabolic equation with certain initial and boundary conditions
(2) |
where are the distributed control and the desired tracking trajectory, respectively, is a regularization parameter, , and and are given problem dependent functions. Under appropriate assumptions, theoretical aspects such as the existence, uniqueness, and regularity of the solution were well studied in [25]. The optimal solution of (1) & (2) can be characterized by the following system:
(3) |
where the control variable has been eliminated.
Following [36, 37], we discretize (3) using the -method for time and some space discretization, which gives
The backward Euler method corresponds to , while the Crank-Nicolson method is adopted when .
Combining the given initial and boundary conditions, one needs to solve the following linear system
(4) |
where we have , ,
(5) |
and the matrices are, respectively,
We assume that the matrix is symmetric positive definite, and the matrix is symmetric positive semi-definite. The matrices and represent the mass matrix and the stiffness matrix, respectively, if a finite element method is employed. For the finite difference method, the linear system is such that and , where is the discretization matrix of the negative Laplacian.
Following a similar idea in [21], we can further transform (5) into the following equivalent system
(6) |
where , , , , and
Note that , , is the identity matrix, and
(8) |
where is a lower triangular Toeplitz matrix whose entries are known explicitly as
Incidentally, can be expressed as the product of two Toeplitz matrices, i.e., . As will be explained in Section 2, the Toeplitz matrices and are respectively generated by the functions
(9) |
and
(10) |
In what follows, we focus on using the finite difference method to discretize the system (3), namely, and in the linear system (6). However, we point out that our proposed preconditioning methods with minimal modification are still applicable when a finite element method is used. We first develop a preconditioned generalized minimal residual (GMRES) method for a nonsymmetric equivalent system of (6), which is
(11) |
where
(12) |
For , we propose the following novel block preconditioner:
(13) |
where
(14) |
Notice that , where
and with . Clearly, both and are -circulant matrices [3, 2], so they admit the eigendecompositions
(15) |
Remark 1.
Since is an -circulant matrix, its eigenvalues can be found explicitly, i.e., . It is known that can be singular. For example, has a zero eigenvalue for even when and (i.e., ), which was discussed in [36]. When it happens, a remedy is to replace the zero eigenvalue by a nonzero real number. Thus, it can easily create a nonsingular circulant matrix such that . Hence, our preconditioning approach can work with replaced by , without the restrictive assumption needed in [36] (i.e., should be chosen odd). Therefore, for ease of exposition, we assume that is nonsingular in the rest of this work.
Remark 2.
It should be noted that our adopted -circulant preconditioning is distinct from the -circulant preconditioning that has received much attention in the literature due to its excellent performance for solving PDE problems (see, e.g, [24, 26, 33, 27]), despite these two kind of matrices have a similar decomposition like (15). A notable difference between the two is that is a complex number in general, while is only chosen real. Correspondingly, the diagonal matrix for -circulant matrices is unitary in general, while that for -circulant matrices is not.
When , becomes the block circulant based preconditioner proposed in [36]. The existing block skew-circulant based preconditioner [4] proposed only with the backward Euler method is also included in our preconditioning strategy when and . As extensively studied in [8, 9], -circulant matrices as preconditioners for Toeplitz systems can substantially outperform the Strang type preconditioners [32] (i.e., when ), especially in the ill-conditioned case. For related studies on the unsatisfactory performance of Strang preconditioners for ill-conditioned nonsymmetric (block) Toeplitz systems, we refer to [19, 18].
Since -circulant matrices can be efficiently diagonalized by the fast Fourier transforms (FFTs), which can be parallelizable over different possessors. Hence, our preconditioner is especially advantageous in a high performance computing environment.
In order to support our GMRES solver with as a preconditioner, we will show that the singular values of are clustered around unity. However, despite of its success which can be seen in the numerical experiments from Section 4, the convergence study of preconditioning strategies for nonsymmetric problems is to a great extent heuristic. As mentioned in [35, Chapter 6], descriptive convergence bounds are usually not available for GMRES or any of the other applicable nonsymmetric Krylov subspace iterative methods.
Therefore, as an alternative solver, we develop a preconditioned minimal residual (MINRES) method for the symmetric system (6), instead of (11). Notice that our proposed MINRES method is in contrast with the aforementioned GMRES solvers, such as [36, 4] where a block (skew-)circulant type preconditioner was proposed and the eigenvalues of the preconditioned matrix were shown clustered around unity. As well explained in [13], the convergence behaviour of GMRES cannot be rigorously analyzed by using only eigenvalues in general. Thus, our MINRES solver can get round these theoretical difficulties of GMRES.
Based on the spectral distribution of , we first propose the following novel SPD block diagonal preconditioner as an ideal preconditioner for :
(16) |
Despite its excellent preconditioning effect for , which will be shown in Section 3.2, the matrix is computational expensive to invert. Thus, we then propose the following parallel-in-time (PinT) preconditioner, which mimics and can be fast implemented:
(17) |
However, the preconditioner requires fast diagonalizability of in order to be efficiently implemented. When such diagonalizability is not available, we further propose the following preconditioner as a modification of :
One of our main contributions in this work is to develop a preconditioned MINRES method with the proposed preconditioners, which has theoretically guaranteed convergence based on eigenvalues.
It is worth noting that our preconditioning approaches are fundamentally different from another kind of related existing work (see, e.g., [31, 23]), which is a typical preconditioning approach in the context of preconditioning for saddle point systems. Its effectiveness is based on the approximation of Schur complements (e.g., [30, 21]) and the classical preconditioning techniques [1, 28]. Yet, for instance, our preconditioning proposal extends the MINRES preconditioning strategy proposed in [17] from optimal control of wave equations to that of parabolic equations, resulting in a clustered spectrum around . Moreover, the implementation of our preconditioners based on FFTs are parallel-in-time.
The paper is organized as follows. In Section 2, we review some preliminary results on block Toeplitz matrices. In Section 3, we provide our main results on the spectral analysis for our proposed preconditioners. Numerical examples are given in Section 4 for supporting the performance of our proposed preconditioners.
2 Preliminaries on Toeplitz matrices
In this section, we provide some useful background knowledge regarding Toeplitz matrices.
We let be the Banach space of all functions that are Lebesgue integrable over and periodically extended to the whole real line. The Toeplitz matrix generated by is denoted by , namely,
where
are the Fourier coefficients of . The function is called the generating function of . If is complex-valued, then is non-Hermitian for all sufficiently large . Conversely, if is real-valued, then is Hermitian for all . If is real-valued and nonnegative, but not identically zero almost everywhere, then is Hermitian positive definite for all . If is real-valued and even, is symmetric for all . For thorough discussions on the related properties of block Toeplitz matrices, we refer readers to [11] and references therein; for computational features see [29, 7, 12] and references there reported.
3 Main results
In this section, the main results which support the effectiveness of our proposed preconditioners are provided. Also, the implementation issue is discussed.
3.1 GMRES - block -circulant based preconditioner
Proposition 3.1.
Proof.
First, we observe that
Now, we examine via the following matrix decomposition:
From the simple structure of these matrices, it is clear that and (because ). Thus, we have , implying . Then, we have
where . ∎
As a consequence of proposition 3.1, we can show that the singular values of are clustered around unity except for a number of outliers whose size is independent of in general. From a preconditioning for nonsymmetric Toeplitz systems point of view, such a singular value cluster is often used to support the preconditioning effectiveness of for when GMRES is used. We refer to [29] for a systematic exposition of preconditioning for non-Hermitian Toeplitz systems. One could further show that the eigenvalues of are also clustered around unity, as with many existing works. However, as mentioned in Section 1, the convergence of GMRES in general cannot be rigorously analyzed by using only eigenvalues. As such, in the next subsections, we provide theoretical supports for our proposed MINRES solvers.
3.2 MINRES - ideal preconditioner
In what follows, we will show that an ideal preconditioner for is the SPD matrix defined by (16).
Proposition 3.2.
Let be defined by (1). Then, the preconditioned matrix is both (real) symmetric and orthogonal.
Proof.
Considering the singular value decomposition of , the associated decomposition of is obtained by direct computation, that is
where is orthogonal given by
with
and
It is obvious that both diagonal matrices and are invertible. Thus is well-defined.
Since both and are orthogonal, is also orthogonal. Hence, from (3.2), we have obtained an eigendecomposition of , i.e.,
Thus, we have
where is orthogonal and is a diagonal matrix containing the singular values of . Thus,
which is both symmetric and orthogonal. The proof is complete. ∎
In other words, Proposition 3.2 shows that the preconditioner can render the eigenvalues exactly at , providing a good guide to designing effective preconditioners for . As a consequence of Proposition 3.2, we conclude that the MINRES with as a preconditioner can achieve mesh-independent convergence, i.e., a convergence rate independent of both the meshes and the regularization parameter.
Despite the fact that is an ideal preconditioner, its direct application has the drawback of being computationally expensive in general. Proposition 3.2 reveals an eigendecomposition of both and , allowing us to develop preconditioners based on the spectral symbol. In what follows, we will show that defined by (17) is a good preconditioner for in the sense that the preconditioned matrix can be expressed as the sum of a Hermitian unitary matrix and a low-rank matrix.
3.3 MINRES - block -circulant based preconditioner
The following theorem accounts for the preconditioning effect of MINRES-.
Theorem 3.3.
Proof.
Let . Notice that .
Simple calculations show that
Thus, , following an argument from the proof of Proposition 3.1. Thus, we have
where .
Since is Hermitian, we have , where is a unitary matrix and is diagonal matrix containing the eigenvalues of . Correspondingly, we have , where is diagonal matrix containing the absolute value eigenvalues of . Thus, we have , which is clearly Hermitian. Also, as , we know that is unitary.
The proof is complete. ∎
As a consequence of Theorem 3.3 and [5, Corollary 3], we know that the preconditioned matrix has clustered eigenvalues at , with a number of outliers independent of in general (i.e., depending only on ). Thus, the convergence is independent of the time step in general, and we can expect that MINRES for will converge rapidly in exact arithmetic with as the preconditioner.
3.4 MINRES - modified block -circulant based preconditioner
To support the preconditioning effect of , we will show that it is spectrally equivalent to . Before proceeding, we introduce the following auxiliary matrix which is useful to showing the preconditioning effect for .
Also, the following lemma is useful.
Lemma 3.4.
Let be defined in (14). Then, is (Hermitian) positive semi-definite.
Proof.
The eigenvalues of can be found explicitly, which are
Thus, we have
which is always non-negative provided is nonsingular by assumption. The proof is complete. ∎
Denote by the spectrum of a square matrix .
Lemma 3.5.
Let and , with the involved notation defined in (14). Then,
Proof.
Let
and
Knowing that the eigendecomposition of is given by with assumed SPD, where is orthogonal and is a real-valued diagonal matrix containing the eigenvalues of , we have
and
Since and are diagonalized by the unitary matrix , they are simultaneously diagonalizable.
Since both and are invertible, they are Hermitian positive definite by construction. To examine the target spectrum of , we consider the Rayleigh quotient for complex :
By the invertibility of and , both numerator and denominator are positive.
On one hand, we estimate an upper bound for . Let and be an entry of and , respectively. We have
which implies
since is positive. Thus, we have
Therefore,
On the other hand, we estimate an lower bound for by first examining the definiteness of the matrix . Since is (Hermitian) non-negative definite by Lemma 3.4, which implies
is also non-negative definite. Thus, we also have
implying
The proof is complete. ∎
Similarly, we can show the following lemma:
Lemma 3.6.
Let and , with the involved notation defined in (14). Then,
Remark 3.
Lemma 3.8.
Let and , with the involved notation defined in (14). Then,
Proof.
Similar to the proof of Lemma 3.5, we let
and
Knowing that the eigendecomposition of is given by , where is orthogonal and is a real diagonal matrix containing the eigenvalues of , we have
and, again,
Thus, and are simultaneously diagonalized by the unitary matrix .
Since both and are invertible, they are Hermitian positive definite by construction. To examine the target spectrum of , we consider the Rayleigh quotient for complex :
For two non-negative numbers, and , it is known that
Therefore, by letting and be an entry of and , respectively, we have
The proof is complete. ∎
Similarly, we can show the following lemma and proposition.
Lemma 3.9.
Let and , with the involved notation defined in (14). Then,
Now, we are ready to show that and are spectrally equivalent in the following theorem, which explains the effectiveness of .
Remark 4.
From Remark 3, we know that when .
In what follows, we will justify the preconditioning effectiveness of for .
Lemma 3.12.
[16, Theorem 4.5.9 (Ostrowski)] Let be matrices. Suppose is Hermitian and is nonsingular. Let the eigenvalues of and be arranged in an increasing order. For each there exists a positive real number such that and
As a consequence of Theorem 3.11, Lemma 3.12, and [5, Corollary 3], we can show the following corollary accounting for the preconditioning effectiveness of :
Corollary 3.13.
Proof.
Remark 5.
When the Crank-Nicolson method is used, we can show that the eigenvalues of the matrix are contained , with a number of outliers independent of in general.
In light of the last corollary, we can expect that MINRES for will converge rapidly in exact arithmetic with as the preconditioner.
3.5 Implementation
We begin by discussing the computation of (and ) for any given vector . The computation of matrix-vector product can be computed in operations by using fast Fourier transforms, because of the fact that contains two block (dense) Toeplitz matrices. The required storage is of . In the special case when for instance, the product requires only linear complexity of since is a sparse matrix with a simple bi-diagonal Toeplitz matrix .
In each GMRES iteration, the matrix-vector product for a given vector needs to be computed. Since -circulant matrices are diagonalizable by the product of a diagonal matrix and a discrete Fourier matrix with , we can represent the matrix defined by (14) using eigendecomposition . Note that is a diagonal matrix.
Hence, we can decompose from (13) as follows:
where is an unitary matrix. Note that the matrix can be further decomposed using the following Lemma in [36].
Lemma 3.14.
([36], Lemma 2.3) Let be four diagonal matrices and . Suppose and are invertible. Then, it holds that
provided is invertible, where is the identity matrix and
Applying Lemma 3.14 to the matrix , we can further decompose as follows:
where , and the matrices and are explicitly known from Lemma 3.14.
Therefore, the computation of can be implemented by the following three steps.
-
1.
,
-
2.
,
-
3.
.
Both Steps 1 and 3 can be computed by fast Fourier transformation in . In Step 2, the shifted Laplacian systems can be efficiently solved for instance by using the multigrid method. A detailed description of this highly effective implementation can be found in [15] for example.
In each MINRES iteration, we need to compute a matrix-vector product in the form of for some given vector . The eigendecomposition of is given by with assumed SPD, where is orthogonal and is a diagonal matrix containing the eigenvalues of .
Hence, we can rewrite from (17) as follows:
where is an unitary matrix and .
Therefore, the computation of can be implemented with the following three steps.
-
1.
,
-
2.
,
-
3.
.
When the spatial grid is uniformly partitioned, the orthogonal matrix becomes the discrete sine matrix . In this case, Step 1 and 3 can be computed efficiently by fast Fourier transform and fast sine transform in operations. For step 2, the required computations take operations since the matrix involved is a simple diagonal matrix.
The product of for any vector can be implemented following the above procedures. Note that
where is an unitary matrix.
The computation of can be implemented by the following three steps.
-
1.
,
-
2.
Compute
-
3.
.
Both Steps 1 and 3 can be computed by fast Fourier transformation in . As for Step 2, again, the shifted Laplacian systems can be efficiently solved by the multigrid method. We refer to [36] for more details regarding such efficient implementation.
4 Numerical examples
In this section, we provide several numerical results to show the performance of our proposed preconditioners. All numerical experiments are carried out using MATLAB 2022b on a PC with Intel i5-13600KF CPU 3.50GHz and 32 GB RAM.
The CPU time in seconds is measured using MATLAB built-in functions . All Steps in Section 3.5 are implemented by the functions and as discrete sine transform and fast Fourier transform respectively. All Krylov subspace solvers used are implemented using the built-in functions on MATLAB. We choose a zero initial guess and a stopping tolerance of based on the reduction in relative residual norms for all Krylov subspace solvers tested unless otherwise indicated.
We adopt the notation MINRES- and MINRES- to represent the MINRES solvers with and , respectively. Also, GMRES- is used to represent the GMRES solver with the proposed preconditioner . We compare our proposed methods against the state-of-the-art solver proposed recently in [23] (denoted by PCG-), where an -circulant preconditioner was constructed. Note that we did not compare with the matching Schur complement preconditioners proposed in [30, 21]. It is expected that their effectiveness cannot surpass PCG- as studied in the numerical tests carried out in [23].
In the related tables, we denote by ’Iter’ the number of iterations for solving a linear system by an iterative solver within the given accuracy. Denote by ’DoF’, the number of unknowns in a linear system. Let and denote the approximate solution to and , respectively. Then, we define the error measure as
(21) |
The time interval and the space are partitioned uniformly with the mesh step size and , respectively, where can be found in the related tables. Also, only is used in the related tables for in the block -circulant preconditioners. It is because, after conducting extensive trials, it was consistently observed that the preconditioner corresponding to yielded the best results.
Example 1.
To support the result of Theorem 3.11, we display the eigenvalues of the matrix for various values of in Figure 1. The illustration confirms that the eigenvalues consistently fall within the interval , aligning with the expectations set forth in Remark 4. Furthermore, it is evident that as diminishes, the eigenvalues of exhibit increased clustering around one. This trend can be attributed to the fact that grows larger when is reduced, assuming the matrix size (or ) remains constant. As becomes larger, it follows from their respective definitions that becomes closer to .



Table 1 displays the iteration counts, CPU times, and errors for GMRES- and MINRES- when applying the Crank-Nicolson method with different values of . Note that MINRES- was not implemented for this example, as can already be efficiently implemented using fast sine transforms. We observe that: (i) both GMRES- and MINRES- with perform excellently and stably, considering both iteration counts and CPU times across various values of ; and (ii) the error decreases as the mesh is refined, except in the case when . In such an instance, the error exhibits only a slight decrease as the matrix size grows, which is likely due to the convergence tolerance used for MINRES not being sufficiently small to demonstrate the anticipated reduction.
In Table 2, we compare our preconditioners against PCG- from [23] with , where only the Crank-Nicolson method is considered. We report that for a larger value of , our proposed GMRES- with outperforms PCG- significantly, namely, the computational time PCG- needed for convergence is roughly two times larger. When is small, GMRES- is still highly comparable with PCG- in terms of CPU times. Overall, GMRES- is stable and robust in both iteration numbers and computational time for a wide range of .
Example 2.
In this example, we consider the following two-dimensional problem of solving (1) with a variable function , where , , , and
The analytical solution of which is given by
In the given example, the direct application of MINRES- is not feasible due to the non-diagonalizability of using fast transform methods. Consequently, we adopt MINRES- which incorporates a multigrid method. Specifically, to solve the shifted Laplacian linear system (as detailed in Subsection 3.5) and compute for any vector , we apply one iteration of the V-cycle geometric multigrid method. In this iteration, the Gauss-Seidel method is employed as the pre-smoother.
Table 3 shows the iteration numbers, CPU time, and error of GMRES- and MINRES-, respectively, when the Crank-Nicolson method is applied with a range of . This example aims to investigate the effectiveness of our solvers when in (3) is not a constant. The results show that (i) GMRES- with maintains relatively stable iteration numbers and CPU time across a wide range of ; (ii) MINRES- performs well for small , but its efficiency decreases as increases — a phenomenon that has also been observed and reported in a previous study [17]. This behavior may be attributed to the eigenvalue distribution of . Specifically, when is very small compared to , it is plausible to assume that is quite large. Indeed, as approaches , it is corroborated by [17, Corollary 3.2] that the matrix-sequence has eigenvalues relatively clustered around , thus facilitating the solving of the all-at-once system. Additionally, as discussed in the preceding example, the increasing size of results in closely resembling , which in turn leads to an improved preconditioning effect.
Example 3.
This example aims to test the robustness of our proposed method with the (homogeneous) Neumann boundary condition. We consider the following two-dimensional problem, where , , , and
The analytical solution of which is given by
Note that we use one iteration of the V-cycle geometric multigrid method with the Gauss-Seidel method as a pre-smoother to solve the shifted Laplacian linear system for GMRES-. Also, MINRES is not applicable in this case with the Neumann boundary condition, since is not symmetric. Thus, we resort to using GMRES-.
Table 4 shows the iteration numbers, CPU time, and error of GMRES- when the Crank-Nicolson method is applied with various values of . The results indicate that GMRES- maintains stable and low iteration numbers across a wide range of .
5 Conclusions
In this work, we have provided a unifying preconditioning framework for circulant-based preconditioning applied to the concerned parabolic control problem. The framework is applicable for both first order (i.e., ) and second order (i.e., ) time discretization schemes. Moreover, it encompasses both circulant (i.e., ) and skew-circulant (i.e., and ) preconditioners previously proposed in existing works. We note that it appears feasible to extend our proposed preconditioning theory to various implicit time-discretization methods, such as Backward Difference Formulas, as long as the block Toeplitz structure within the resulting all-at-once linear system remains intact. When considering more general discretization schemes, such as (semi-)implicit Runge-Kutta methods, a good starting point could be the study recently introduced in [20]. This work develops a robust preconditioning approach designed specifically for all-at-once linear systems that are derived from the Runge-Kutta time discretization of time-dependent PDEs.
Specifically, we have proposed a class of block -circulant based preconditioners for the all-at-once system of the parabolic control problem. First, when GMRES is considered, we have proposed a PinT preconditioner for the concerned system. Second, when MINRES is used for the symmetrized system, we have constructed an ideal preconditioner , which can be used as a prototype for designing efficient preconditioners based on . Then, we have designed two novel preconditioners and for the same problem, which can be efficiently implemented in a PinT manner. All proposed preconditioners have been shown effective in both numerical tests and a theoretical study. Based on our numerical tests, it has been demonstrated that our proposed solver, GMRES- with and , can achieve rapid convergence, consistently maintaining stable iteration counts across a wide range of values.
We stress that the development of our proposed MINRES approach for optimal control problems is still in its infancy. As future work, we plan at least to develop more efficient preconditioned Krylov subspace solvers by integrating with an -circulant matrix, where a small is chosen. In recent years, this approach has been shown successful for solving various PDEs (see, e.g, [24, 26, 33, 27]), achieving clustered singular values without any outliers. We will investigate whether such a combination can reduce the number of singular values/eigenvalue outliers that are present as a result from our preconditioners, which could achieve parameter-independent convergence in the MINRES framework.
Acknowledgments
The work of Sean Hon was supported in part by the Hong Kong RGC under grant 22300921 and a start-up grant from the Croucher Foundation.
DoF | GMRES- | MINRES- | ||||||
---|---|---|---|---|---|---|---|---|
Iter | CPU | Iter | CPU | |||||
61504 | 3 | 0.039 | 1.18e-9 | 3 | 0.033 | 3.18e-9 | ||
508032 | 3 | 0.35 | 1.18e-9 | 5 | 0.48 | 1.45e-9 | ||
4129024 | 3 | 3.32 | 1.04e-9 | 6 | 4.91 | 1.04e-9 | ||
33292800 | 3 | 27.96 | 4.49e-10 | 6 | 40.91 | 4.49e-10 | ||
61504 | 3 | 0.030 | 1.12e-7 | 6 | 0.045 | 1.26e-7 | ||
508032 | 3 | 0.37 | 6.71e-8 | 6 | 0.52 | 6.71e-8 | ||
4129024 | 3 | 3.30 | 1.81e-8 | 6 | 4.90 | 1.81e-8 | ||
33292800 | 3 | 27.85 | 4.53e-9 | 6 | 40.83 | 4.53e-9 | ||
61504 | 3 | 0.029 | 2.90e-6 | 6 | 0.044 | 2.90e-6 | ||
508032 | 3 | 0.34 | 7.26e-7 | 6 | 0.56 | 7.26e-7 | ||
4129024 | 3 | 3.35 | 1.81e-7 | 6 | 4.91 | 1.81e-7 | ||
33292800 | 3 | 27.76 | 4.54e-8 | 6 | 40.98 | 4.54e-8 | ||
61504 | 3 | 0.029 | 2.87e-5 | 6 | 0.045 | 2.87e-5 | ||
508032 | 3 | 0.39 | 7.19e-6 | 6 | 0.51 | 7.19e-6 | ||
4129024 | 3 | 3.34 | 1.80e-6 | 6 | 4.91 | 1.80e-6 | ||
33292800 | 3 | 27.96 | 4.49e-7 | 6 | 40.84 | 4.49e-7 | ||
61504 | 3 | 0.030 | 2.77e-4 | 6 | 0.046 | 2.77e-4 | ||
508032 | 3 | 0.37 | 6.91e-5 | 6 | 0.55 | 6.91e-5 | ||
4129024 | 3 | 3.35 | 1.73e-5 | 6 | 4.90 | 1.73e-5 | ||
33292800 | 3 | 27.80 | 4.31e-6 | 6 | 40.96 | 4.31e-6 |
DoF | PCG- | ||||
---|---|---|---|---|---|
Iter | CPU | ||||
61504 | 3 | 0.084 | 1.18e-9 | ||
508032 | 3 | 0.23 | 1.18e-9 | ||
4129024 | 4 | 2.49 | 1.04e-9 | ||
33292800 | 5 | 24.55 | 4.49e-10 | ||
61504 | 4 | 0.023 | 1.12e-7 | ||
508032 | 5 | 0.34 | 6.71e-8 | ||
4129024 | 5 | 3.01 | 1.81e-8 | ||
33292800 | 5 | 24.70 | 4.53e-9 | ||
61504 | 5 | 0.046 | 2.90e-6 | ||
508032 | 5 | 0.33 | 7.26e-7 | ||
4129024 | 6 | 3.50 | 1.81e-7 | ||
33292800 | 6 | 28.64 | 4.54e-8 | ||
61504 | 9 | 0.063 | 2.87e-5 | ||
508032 | 9 | 0.57 | 7.19e-6 | ||
4129024 | 9 | 4.95 | 1.80e-6 | ||
33292800 | 10 | 44.46 | 4.49e-7 | ||
61504 | 11 | 0.061 | 2.77e-4 | ||
508032 | 11 | 0.68 | 6.91e-5 | ||
4129024 | 11 | 5.95 | 1.73e-5 | ||
33292800 | 12 | 53.96 | 4.31e-6 |
DoF | GMRES- | MINRES- | ||||||
---|---|---|---|---|---|---|---|---|
Iter | CPU | Iter | CPU | |||||
61504 | 3 | 0.14 | 6.60e-13 | 3 | 0.11 | 2.91e-10 | ||
508032 | 3 | 0.67 | 4.68e-13 | 5 | 0.74 | 2.55e-11 | ||
4129024 | 3 | 5.35 | 7.51e-13 | 6 | 6.27 | 4.43e-13 | ||
33292800 | 3 | 48.26 | 8.40e-12 | 6 | 61.63 | 1.73e-11 | ||
61504 | 3 | 0.13 | 6.30e-11 | 6 | 0.17 | 6.29e-11 | ||
508032 | 3 | 0.64 | 5.48e-11 | 6 | 0.84 | 5.39e-11 | ||
4129024 | 3 | 5.39 | 1.13e-10 | 7 | 7.51 | 7.28e-10 | ||
33292800 | 3 | 59.97 | 1.14e-10 | 9 | 107.00 | 3.90e-11 | ||
61504 | 3 | 0.17 | 2.53e-9 | 7 | 0.24 | 2.79e-9 | ||
508032 | 3 | 0.81 | 1.26e-9 | 10 | 1.65 | 5.65e-10 | ||
4129024 | 3 | 6.71 | 1.14e-9 | 10 | 11.98 | 3.22e-10 | ||
33292800 | 3 | 74.15 | 1.14e-9 | 10 | 126.17 | 5.38e-10 | ||
61504 | 5 | 0.25 | 1.53e-7 | 14 | 0.42 | 1.53e-7 | ||
508032 | 5 | 1.24 | 3.40e-8 | 15 | 2.57 | 3.40e-8 | ||
4129024 | 5 | 10.89 | 8.51e-9 | 18 | 23.00 | 8.51e-9 | ||
33292800 | 5 | 100.60 | 2.13e-9 | 23 | 313.30 | 2.13e-9 | ||
61504 | 5 | 0.25 | 1.16e-5 | 20 | 0.57 | 1.16e-5 | ||
508032 | 5 | 1.54 | 2.90e-6 | 24 | 4.28 | 2.90e-6 | ||
4129024 | 5 | 12.99 | 7.25e-7 | 30 | 39.82 | 7.25e-7 | ||
33292800 | 5 | 121.50 | 1.81e-7 | 52 | 739.91 | 1.81e-7 |
DoF | GMRES- | ||||
---|---|---|---|---|---|
Iter | CPU | ||||
65536 | 3 | 0.23 | 1.51e-11 | ||
524288 | 3 | 1.02 | 1.39e-11 | ||
4194304 | 3 | 6.98 | 1.18e-11 | ||
33554432 | 3 | 81.18 | 4.99e-12 | ||
65536 | 3 | 0.18 | 1.43e-9 | ||
524288 | 3 | 1.07 | 7.93e-10 | ||
4194304 | 3 | 8.37 | 2.06e-10 | ||
33554432 | 3 | 94.33 | 5.05e-11 | ||
65536 | 3 | 0.22 | 3.69e-8 | ||
524288 | 3 | 1.26 | 8.56e-9 | ||
4194304 | 3 | 12.36 | 2.06e-9 | ||
33554432 | 3 | 152.13 | 5.05e-10 | ||
65536 | 3 | 0.27 | 3.73e-7 | ||
524288 | 3 | 1.70 | 8.64e-8 | ||
4194304 | 3 | 18.68 | 2.08e-8 | ||
33554432 | 3 | 249.61 | 5.10e-9 | ||
65536 | 3 | 0.35 | 4.10e-6 | ||
524288 | 3 | 2.28 | 9.51e-7 | ||
4194304 | 3 | 22.44 | 2.29e-7 | ||
33554432 | 3 | 282.79 | 5.61e-8 |
References
- [1] Owe Axelsson and Maya Neytcheva. Eigenvalue estimates for preconditioned saddle point matrices. Numerical Linear Algebra with Applications, 13(4):339-360, 2006.
- [2] Daniele Bertaccini and Michael K. Ng. Block -circulant preconditioners for the systems of differential equations. Calcolo, 40(2):71-90, 2003.
- [3] Dario A. Bini, Guy Latouche, and Beatrice Meini. Numerical Methods for Structured Markov Chains. Oxford University Press, New York, 2005.
- [4] Arne Bouillon, Giovanni Samaey, and Karl Meerbergen. On generalized preconditioners for time-parallel parabolic optimal control. arXiv preprint, arXiv:2302.06406, 2021.
- [5] Jan H. Brandts and Ricardo Reis da Silva. Computable eigenvalue bounds for rank- perturbations. Linear Algebra and Its Applications, 432(12), 3100-3116, 2010.
- [6] Alfio Borzì and Volker Schulz. Computational optimization of systems governed by partial differential equations. Society for Industrial and Applied Mathematics. 2011.
- [7] Raymond H. Chan and Michael K. Ng. Conjugate gradient methods for Toeplitz systems. SIAM Review, 38(3), 427-482, 1996.
- [8] Daniel Potts and Gabriele Steidl. Preconditioners for ill-conditioned Toeplitz matrices. BIT Numerical Mathematics, 39(3), 513-533, 1999.
- [9] Daniel Potts and Gabriele Steidl. Preconditioners for ill-conditioned Toeplitz systems constructed from positive kernels. SIAM Journal on Scientific Computing, 22(5), 1741-1761, 2001
- [10] Paola Ferrari, Isabella Furci, Sean Hon, Mohammad Ayman-Mursaleen, and Stefano Serra-Capizzano. The eigenvalue distribution of special 2-by-2 block matrix-sequences with applications to the case of symmetrized Toeplitz structures. SIAM Journal on Matrix Analysis and Applications, 40(3), 1066-1086, 2019.
- [11] Carlo Garoni and Stefano Serra-Capizzano. Generalized locally Toeplitz sequences: Theory and applications, volume 1. Springer, Cham, 2017.
- [12] Carlo Garoni and Stefano Serra-Capizzano. Generalized locally Toeplitz sequences: Theory and applications, volume 2. Springer, Cham, 2018.
- [13] Anne Greenbaum, Vlastimil Pták, and Zdeněk Strakoš. Any nonincreasing convergence curve is possible for GMRES. SIAM Journal on Matrix Analysis and Applications, 17(3):465–469, 1996.
- [14] Michael Hinze, Rene Pinnau, Michael Ulbrich, and Stefan Ulbrich. Optimization with PDE constraints (Vol. 23). Springer Science & Business Media. 2008.
- [15] Yunhui He and Jun Liu. A Vanka-type multigrid solver for complex-shifted Laplacian systems from diagonalization-based parallel-in-time algorithms. Applied Mathematics Letters, 132, 108125, 2022.
- [16] Roger A. Horn and Charles R. Johnson. Matrix Analysis. Cambridge University Press. 1990.
- [17] Sean Hon, Jiamei Dong, and Stefano Serra-Capizzano. A preconditioned MINRES method for optimal control of wave equations and its asymptotic spectral distribution theory. SIAM Journal on Matrix Analysis and Applications, 44(4):1477–1509, 2023.
- [18] Sean Hon, Po Yin Fung, Jiamei Dong, and Stefano Serra-Capizzano. A sine transform based preconditioned MINRES method for all-at-once systems from constant and variable-coefficient evolutionary PDEs. Numerical Algorithms, 2023.
- [19] Sean Hon, Stefano Serra-Capizzano, and Andy Wathen. Band-Toeplitz preconditioners for ill-conditioned Toeplitz systems. BIT Numerical Mathematics. 2021.
- [20] Santolo Leveque, Luca Bergamaschi, Ángeles Martínez, and John W. Pearson. Fast iterative solver for the all-at-once Runge–Kutta discretization. arXiv preprint, arXiv:2303.02090, 2023.
- [21] Santolo Leveque and John W. Pearson. Fast iterative solver for the optimal control of time-dependent PDEs with Crank–Nicolson discretization in time. Numerical Linear Algebra with Applications, 29:e2419, 2022.
- [22] Buyang Li, Jun Liu, and Mingqing Xiao. A new multigrid method for unconstrained heat optimal control problems. Journal of Computational and Applied Mathematics, 326, 358-373, 2017.
- [23] Xue-lei Lin and Shu-Lin Wu. A parallel-in-time preconditioner for Crank-Nicolson discretization of a parabolic optimal control problem. arXiv preprint, arXiv:2109.12524, 2023.
- [24] Xue-lei Lin and Michael Ng. An All-at-once preconditioner for evolutionary partial differential equations. SIAM Journal on Scientific Computing, 43(4), A2766–A2784, 2021.
- [25] Jacques Louis Lions. Optimal control of systems governed by partial differential equations. Springer-Verlag Berlin Heidelberg. 1971.
- [26] Jun Liu and Shu-Lin Wu. A fast block -circulant preconditioner for all-at-once systems from wave equations. SIAM Journal on Matrix Analysis and Applications. 41(4), 1912–1943, 2021.
- [27] Martin J. Gander, Jun Liu, Shu-Lin Wu, Xiaoqiang Yue, and Tao Zhou. ParaDiag: parallel-in-time algorithms based on the diagonalization technique. arXiv e-prints, arXiv:2005.09158, 2020.
- [28] Malcolm F. Murphy, Gene H. Golub, and Andrew J. Wathen. A note on preconditioning for indefinite linear systems. SIAM Journal on Scientific Computing, 21(6):1969–1972, 2000.
- [29] Michael K. Ng. Iterative Methods for Toeplitz Systems. Numerical Mathematics and Scientific Computation. Oxford University Press, New York, 2004.
- [30] John W. Pearson and Andrew J. Wathen. A new approximation of the Schur complement in preconditioners for PDE-constrained optimization. Numerical Linear Algebra and Applications, 19(5):816-829, 2012.
- [31] John W. Pearson, Martin Stoll, and Andrew J. Wathen. Regularization-robust preconditioners for time-dependent PDE-constrained optimization problems. SIAM Journal on Matrix Analysis and Applications, 33:1126-1152, 2012.
- [32] Gilbert Strang. A Proposal for Toeplitz Matrix Calculations. Studies in Applied Mathematics, 74(2), 171-176, 1986.
- [33] Yafei Sun, Shu-Lin Wu, and Yingxiang Xu. A Parallel-in-Time Implementation of the Numerov Method For Wave Equations. Journal of Scientific Computing, 90.1, 2022.
- [34] Ferdi Tröltzsch. Optimal control of partial differential equations: theory, methods, and applications (Vol. 112). American Mathematical Soc. 2010.
- [35] Andrew J. Wathen. Preconditioning. Acta Numerica. 24, 329-376, 2015.
- [36] Shu-Lin Wu and Tao Zhou. Diagonalization-based parallel-in-time algorithms for heat PDE-constrained optimization problems. ESAIM: Control, Optimisation and Calculus of Variations, 26, 88, 2020.
- [37] Shu-Lin Wu, Zhiyong Wang, and Tao Zhou. PinT Preconditioner for Forward-Backward Evolutionary Equations. SIAM Journal on Matrix Analysis and Applications, 44(4): 1771-1798, 2023.