Relaxed Fixed Point Iterations for Matrix Equations Arising in Markov Chains Modeling††thanks: The authors are partially supported by INDAM/GNCS and by the project PRA_2020_61 of the University of Pisa.
Abstract
We present some accelerated variants of fixed point iterations for computing the minimal non-negative solution of the unilateral matrix equation associated with an M/G/1-type Markov chain. These variants derive from certain staircase regular splittings of the block Hessenberg M-matrix associated with the Markov chain. By exploiting the staircase profile we introduce a two-step fixed point iteration. The iteration can be further accelerated by computing a weighted average between the approximations obtained at two consecutive steps. The convergence of the basic two-step fixed point iteration and of its relaxed modification is proved. Our theoretical analysis, along with several numerical experiments, show that the proposed variants generally outperform the classical iterations.
Keywords: M-matrix, Staircase Splitting, Nonlinear Matrix Equation, Hessenberg Matrix, Markov Chain.
1 Introduction
The transition probability matrix of an M/G/1-type Markov chain is a block Hessenberg matrix of the form
(1) |
with and and stochastic matrices.
In the sequel, given a real matrix , we write () if () for any . A stochastic matrix is matrix such that , where is the column vector having all the entries equal to 1.
In the positive recurrent case, the computation of the steady state vector of , such that
(2) |
is related with the solution of the unilateral power series matrix equation
(3) |
Under some mild assumptions this equation has a componentwise minimal non-negative solution which determines, by means of Ramaswami’s formula [16], the vector .
Among the easy-to-use, but still effective, tools for numerically solving (3), there are fixed point iterations (see [3] and the references given therein for a general review of these methods). The intrinsic simplicity of such schemes make them attractive in domains where high performance computing is crucial. But they come at a price: the convergence can become very slow especially for problems which are close to singularity. The design of acceleration methods (aka extrapolation methods) for fixed point iterations is a classical topic in numerical analysis [5]. Relaxation techniques are commonly used for the acceleration of classical stationary iterative solvers for large linear systems. In this paper we introduce some new coupled fixed point iterations for solving (3) which can be combined with relaxation techniques to speed up their convergence.
More specifically, we first observe that computing the solution of the matrix equation (3) is formally equivalent to solving a semi-infinite block Toeplitz block Hessenberg linear system. Customary block iterative algorithms applied for the solution of this system yield classical fixed point iterations. In particular, the traditional and the U-based fixed point iteration [3] originate from the block Jacobi and the block Gauss-Seidel method, respectively. Recently, in [7] the authors show that some iterative solvers based on a block staircase partitioning outperform the block Jacobi and the block Gauss-Seidel method for M-matrix linear systems in block Hessenberg form. The application of the staircase splitting to the block Toeplitz block Hessenberg linear system associated with (3) yields the new coupled fixed point iteration
(4) |
starting from an initial approximation . The contribution of this paper is aimed at highlighting the properties of (4).
We show that, if , the sequence defined in (4) converges to faster than the traditional fixed point iteration. In the case where the starting matrix of (4) is any column stochastic matrix and is also column stochastic, we prove that the sequence still converges to . Moreover, by comparison of the mean asymptotic rates of convergence, we conclude that (4) is asymptotically faster than the traditional fixed point iteration.
At each iteration the scheme (4) determines two approximations which can be combined by using the relaxation technique, that is, the approximation computed at the -th step takes the form of a weighted average between and . The modified relaxed variant is defined by the sequence
(5) |
where is the relaxation parameter. The convergence results for (4) easily extend to the modified scheme in the case of under-relaxation, that is, the parameter satisfies . Heuristically, it is argued that over-relaxation values ( can improve the convergence. If , under some assumptions a theoretical estimate of the asymptotic convergence rate of (5) is given which confirms this heuristic. Moreover, an adaptive strategy is devised which makes possible to perform over-relaxed iterations of (5) by still ensuring the convergence of the overall iterative process. The results of extensive numerical experiments confirm the effectiveness of the proposed variants, which generally outperform the -based fixed point iteration for nearly singular problems. In particular, the over-relaxed scheme (5) with , combined with the adaptive strategy for parameter estimation, is capable to significantly accelerate the convergence without increasing the computational cost.
The paper is organized as follows. In Section 2 we set up the theoretical framework, briefly recalling some preliminary properties and assumptions. In Section 3 we revisit classical fixed point iterations for solving (3) by establishing the link with the iterative solution of an associated block Toeplitz block Hessenberg linear system. In Section 4 we introduce the new fixed point iteration (4) by proving some convergence results. The relaxed variant (5), as well as the generalizations of convergence results for this variant, are described in Section 5. Section 6 deals with a formal analysis of the asymptotic convergence rate of both (4) and (5). Adaptive strategies for the choice of the relaxation parameter are discussed in Section 7, together with their cost analysis under some simplified assumptions. Finally, the results of extensive numerical experiments are presented in Section 8 whereas conclusions and future work are the subjects of Section 9.
2 Preliminaries and assumptions
Throughout the paper we assume that , , are nonnegative matrices, such that their sum is irreducible and row stochastic, that is, , . According to the results of [3, Chapter 4], such assumption implies that (3) has a unique componentwise minimal nonnegative solution ; moreover, is an invertible M-matrix and, hence, .
Furthermore, in view of the Perron Frobenius Theorem, there exists a unique vector such that and , . If the series is convergent we may define the vector . In the study of M/G/1-type Markov chains, the drift is the scalar number [14]. The sign of the drift determines the positive recurrence of the M/G/1-type Markov chain [3].
3 Nonlinear matrix equations and structured linear systems
In this section we reinterpret classical fixed point iterations for solving the matrix equation (3) in terms of iterative methods for solving a structured linear system.
Formally, the power series matrix equation (3) can be rewritten as the following block Toeplitz block Hessenberg linear system
The above linear system can be expressed in compact form as
(6) |
where , is the matrix obtained from in (1) by removing its first block row and block column and . Classical fixed point iterations for solving (3) can be interpreted as iterative methods for solving (6), based on suitable partitionings of the matrix .
For instance, from the partitioning , where and , we find that the block vector is a solution of the fixed point problem
(7) |
From this equation we may generate the sequence of block vectors
such that
We may easily verify that the sequence coincides with the sequence generated by the so called natural fixed point iteration , , applied to (3).
Similarly, the Jacobi partitioning, where and , which leads to the sequence
corresponds to the traditional fixed point iteration
(8) |
The anti-Gauss-Seidel partitioning, where is the block upper triangular part of and , determines the -based fixed point iteration
(9) |
The convergence properties of these three fixed point iterations are analyzed in [3]. Among the three iterations (9) is the fastest and also the most expensive since it requires the solution of a new linear system (with multiple right hand sides) at each iteration. Moreover, it turns out that fixed-point iterations exhibit arbitrarily slow convergence for problems which are close to singularity. In particular, for positive recurrent Markov chains having a drift close to zero the convergence slows down and the number of iterations becomes arbitrarily large. In the next sections we present some new fixed point iterations which offer several advantages in terms of computational efficiency and convergence properties when compared with (9).
4 A new fixed point iteration
Recently in [7] a comparative analysis has been performed for the asymptotic convergence rates of some regular splittings of a non-singular block upper Hessenberg M-matrix of finite size. The conclusion is that the staircase splitting is faster than the anti-Gauss-Seidel splitting, that in turn is faster than the Jacobi splitting. The second result is classical, while the first one is somehow surprising since the matrix in the staircase splitting is much more sparse than the corresponding matrix in the anti-Gauss-Seidel partitioning and the splittings are not comparable. Inspired from these convergence properties, we introduce a new fixed point iteration for solving (3), based on the staircase partitioning of , namely,
(10) |
The splitting has attracted interest for applications in parallel computing environments [12, 10]. In principle the alternating structure of the matrix in (10) suggests several different iterative schemes.
From one hand, the odd block entries of the system yield the traditional fixed point iteration. On the other hand, the even block entries lead to the implicit scheme recently introduced in [4]. Differently, by looking at the structure of the matrix on the whole, we introduce the following composite two-stage iteration:
(11) |
or, equivalently,
(12) |
starting from an initial approximation . At each step , this scheme consists of a traditional fixed point iteration that computes from , followed by a cheap correction step for computing the new approximation .
As for classical fixed point iterations, the convergence is guaranteed when .
Proposition 1.
Assume that . Then the sequence generated by (12) converges monotonically to .
Proof.
We show by induction on that for any . For we verify easily that
which gives . Suppose now that , . We find that
and
By multiplying both sides by the inverse of we obtain that . This also implies that and therefore . Since
we prove similarly that . It follows that is convergent, the limit solves (3) by continuity and, hence, the limit coincides with the matrix , since is the minimal nonnegative solution. ∎
A similar result also holds for the case where is a stochastic matrix, assuming that [A1] holds, so that is also stochastic.
.
Proposition 2.
Assume that condition [A1] is fulfilled and that is a stochastic matrix. Then, the sequence generated by (12) converges to .
Proof.
From (11) we obtain that
which gives that and , for any , since . By assuming that , we may easily show by induction that for any . Therefore, all the matrices and , , are stochastic. Let be the sequence generated by (12) with . We can easily show by induction that for any . Since , then any convergent subsequence of converges to a stochastic matrix such that . Since is also stochastic, it follows that and therefore, by compactness, we conclude that the sequence is also convergent to . ∎
5 A relaxed variant
At each iteration, the scheme (12) determines two approximations which can be combined by using a relaxation technique, that is, the approximation computed at the -th step takes the form of a weighted average between and ,
In matrix terms, the resulting relaxed variant of (12) can be written as
(13) |
If , for , the relaxed scheme reduces to the traditional fixed point iteration (8). If , for , the relaxed scheme coincides with (12). Values of greater than 1 can speed up the convergence of the iterative scheme.
Concerning convergence, the proof of Proposition 1 can immediately be generalized to show that the sequence defined by (13), with , converges for any , , such that . Moreover, let and be the sequences generated by (13) for and with , respectively. It can be easily shown that for any and, hence, that the iterative scheme (12) converges faster than (13) if .
The convergence analysis of the modified scheme (13) for is much more involved since the choice of a relaxation parameter can destroy the monotonicity and the nonnegativity of the approximation sequence, which is at the core of the proofs of Proposition 1 and Proposition 2 . In order to maintain the convergence properties of the modified scheme we introduce the following definition.
Definition 3.
The sequence is eligible for the scheme (13) if , , and the following two conditions are satisfied:
(14) |
and
(15) |
It is worth noting that condition (14) is implicit since the construction of also depends on the value of . By replacing in (14) with the expression in the right-hand side of (13) we obtain a quadratic inequality with matrix coefficients in the variable . Obviously , , , are eligible sequences.
The following generalization of Proposition 1 holds.
Proposition 4.
Set and let condition [A1] be satisfied. If is eligible then the sequence generated by (13) converges monotonically to .
Proof.
We show by induction that . For we have
which gives immediately . Moreover, . Suppose now that , . We find that
from which it follows . This also implies that . From (15) it follows that for all and therefore the sequence of approximations is upper bounded and it has a finite limit . By continuity we find that solves the matrix equation (3) and . Since is the unique stochastic solution, then . ∎
Remark 5.
As previously mentioned, condition (14) is implicit, since also depends on . An explicit condition can be derived by noting that
with . There follows that (14) is fulfilled whenever
which can be reduced to a linear inequality in over a fixed search interval. Let be such that
(16) |
Then we can impose that
(17) |
From a computational viewpoint the strategy based on (16) and (17) for the choice of the value of can be too much expensive and some weakened criterion should be considered (compare with Section 7 below).
In the following section we perform a convergence analysis to estimate the convergence rate of (13) in the stationary case , , as a function of the relaxation parameter.
6 Estimate of the convergence rate
Relaxation techniques are usually aimed at accelerating the convergence speed of frustratingly slow iterative solvers. Such inefficient behavior is typically exhibited when the solver is applied to a nearly singular problem. Incorporating some relaxation parameter into the iterative scheme (3) can greatly improve its convergence rate. Preliminary insights on the effectiveness of relaxation techniques applied for the solution of the fixed point problem (7) come from the classical analysis for stationary iterative solvers and are developed in Section 6.1. A precise convergence analysis is presented in Section 6.2.
6.1 Finite dimensional convergence analysis
Suppose that in (7) is block tridiagonal of finite size , even. We are interested in comparing the iterative algorithm based on the splitting (10) with other classical iterative solvers for the solution of a linear system with coefficient matrix . As usual, we can write , where is block diagonal, while and are staircase matrices with zero block diagonal. The eigenvalues of the Jacobi iteration matrix satisfy
Let us consider a relaxed scheme where the matrix is obtained from (10) by multiplying the off-diagonal blocks by . The eigenvalues of the iteration matrix associated with the relaxed staircase regular splitting are such that
and, equivalently,
By using a similarity transformation induced by the matrix we find that
There follows that
whenever fulfills
Therefore, the eigenvalues of the Jacobi and relaxed staircase regular splittings are related by
For the staircase splitting reduces to the Jacobi partitioning. For we find that which yields the classical relation between the spectral radii of Jacobi and Gauss-Seidel methods. Observe that it is well known that the asymptotic convergence rates of Gauss-Seidel and the staircase iteration coincide when applied to a block tridiagonal matrix [1]. For the spectral radius of the relaxed staircase scheme can be significantly less than the spectral radius of the same scheme for . In Figure 1 we illustrate the plot of the function
for a fixed .

For the best choice of we find .
6.2 Asymptotic Convergence Rate
A formal analysis of the asymptotic convergence rate of the relaxed variants (13) can be carried out by using the tools described in [11]. In this section we relate the approximation error at two subsequent steps and we provide an estimate of the asymptotic rate of convergence, expressed as the spectral radius of a suitable matrix depending on .
Hereafter it is assumed that assumption [A1] is verified.
6.2.1 The case
Let us introduce the error matrix , where is generated by (13) with . We also define , for . Suppose that
-
C0.
is an eligible sequence according to Definition 3.
Under this assumption from Proposition 4 the sequence converges monotonically to and , . Since and , we analyze the convergence of the vector , .
We have
(18) |
Similarly, for the second equation of (13), we find that
which gives
(19) |
Denote by the matrix on the right hand side of (18), i.e.,
Since , equation (6.2.1), together with the monotonicity, yields
(20) |
Observe that , where
hence
(21) |
where
(22) |
The matrix can be written as where
and
Let us assume the following condition holds:
-
C1.
The relaxation parameter satisfies with such that
Assumption [C1] ensures that and, therefore and is a regular splitting of . If [C1] is satisfied at each iteration of (13), then from (21) we obtain that
Therefore, the asymptotic rate of convergence, defined as
where is any vector norm, is such that
The above properties can be summarized in the following result, that gives a convergence rate estimate for iteration (13).
Proposition 6.
When , we find that , where . According to Theorem 4.14 in [3], is a nonsingular M-matrix and therefore, since and , is a regular splitting. Hence, the spectral radius of is less than 1. More generally, under Assumption [C1] since
from characterization F20 in [15] we find that is a nonsingular M-matrix and is a regular splitting. Hence, we deduce that . The following result gives an estimate of by showing its dependence as function of the relaxation parameter.
Proposition 7.
Let be such that and condition [C1] holds. Assume that the Perron eigenvector of is positive. Then we have
(23) |
where and , with . Moreover, .
Proof.
Observe that, in the QBD case, where for , we have and from the proof above . Therefore, we have and, hence, . In particular, linearly decreases with , and . In the general case, inequality (23) shows that the upper bound to linearly decreases as a function of . Therefore the choice gives the fastest convergence rate.
Remark 8.
For the sake of illustration we consider a quadratic equation associated with a block tridiagonal Markov chain taken from [9]. We set , where and has zero diagonal entries and all off-diagonal entries equal to a given value determined so that is stochastic. We find that for . In Figure 2 we plot the spectral radius of . The linear plot is in accordance with Theorem 7.

6.2.2 The Case stochastic
In this section we analyze the convergence of the iterative method (13) starting with a stochastic matrix , that is, and . Eligible sequences are such that for any . This happens for , . Suppose that:
-
S0.
The sequence in (13) is determined so that and for any .
Observe that the property , is automatically satisfied. Hence, under assumption [S0], all the approximations generated by the iterative scheme (13) are stochastic matrices and therefore, Proposition 2 can be extended in order to prove that the sequence is convergent to .
The analysis of the speed of convergence follows from relation (6.2.1). Let us denote as the vector obtained by stacking the columns of the matrix on top of one another. Recall that for any . By using this property we can rewrite (6.2.1) as follows. We have:
(24) |
for . The convergence of depends on the choice of , . Suppose that for any and [S0] holds. Then (24) can be rewritten in a compact form as
where and
It can be shown that the asymptotic rate of convergence satisfies
In the sequel we compare the cases , which corresponds with the traditional fixed point iteration (8), and which reduces to the staircase fixed point iteration (12).
For , , we find that
which means that
Let be the Schur form of and set . Then
which means that is similar to the matrix on the right hand-side. There follows that the eigenvalues of belong to the set
with eigenvalue of . Since is stochastic we have . Thus, from
we conclude that in view of the Wielandt theorem [13].
A similar analysis can be performed in the case , . We find that
By the same arguments as above we find that the eigenvalues of belong to the set
with eigenvalue of , and
Since
we conclude that
Therefore, in the application of (12) we expect a faster convergence when is a stochastic matrix, rather than . Indeed, numerical results shown in Section 5 exhibit a very rapid convergence profile when is stochastic, even better than the one predicted by . This might be explained with the dependence of the asymptotic convergence rate on the second eigenvalue of the corresponding iteration matrices as reported in [11].
7 Adaptive Strategies and Efficiency Analysis
The efficiency of fixed point iterations depends on both speed of convergence and complexity properties. Markov chains are generally defined in terms of sparse matrices. To take into account this feature we assume that , , multiplications/divisions are sufficient to perform the following tasks:
-
1.
to compute a matrix multiplication of the form , where ;
-
2.
to solve a linear system of the form , where .
We also suppose that the transition matrix in (1) is banded, hence for . This is always the case in numerical computations where the matrix power series has to be approximated by some finite partial sum . Under these assumptions we obtain the following cost estimates per step:
-
1.
the traditional fixed point iteration (8) requires multiplicative operations;
-
2.
the -based fixed point iteration (9) requires multiplicative operations;
-
3.
the staircase-based (S-based) fixed point iteration (12) requires multiplicative operations.
Observe that the cost of the S-based fixed point iteration is comparable with the cost of the -based iteration, which is the fastest among classical iterations [3]. Therefore, in the cases where the /S-based fixed point iterative schemes require significantly less iterations to converge, these algorithms are more efficient than the traditional fixed point iteration.
Concerning the relaxed versions (13) of the S-based fixed point iteration for a given fixed choice of we get the same complexity of the unmodified scheme (12) obtained with . The adaptive selection of exploited in Proposition 4 and Remark 5 with requires more care.
The strategy (16) is computationally unfeasible since it needs the additional computation of . To approximate this quantity we recall that
Let be such that
Then condition (16) can be replaced with
(25) |
The iterative scheme (13) complemented with the strategy based on (25) for the selection of the parameter requires no more than multiplicative operations. The efficiency of this scheme will be investigated experimentally in the next section.
8 Numerical Results
In this section we present the results of some numerical experiments which confirm the effectiveness of our proposed schemes. All the algorithm have been implemented in Matlab and tested on a PC i9-9900K CPU 3.60GHz8. Our test suite includes:
-
1.
Synthetic Examples:
-
(a)
The block tridiagonal Markov chain of Remark 8. Observe that the drift of the Markov chain is exactly equal to .
-
(b)
A numerical example given in [2] and considered in [8] for testing a suitable modification –named Algorithm 1– of the U-based fixed point iteration. The Markov chain of the M/G/1 type is given by
The Markov chain is positive recurrent, null recurrent or transient according as , or , respectively. In our computations we have chosen different values of , in the range and the matrices are treated as zero matrix for .
-
(c)
Synthetic examples of M/G/1-type Markov chains described in [4]. These examples are constructed in such a way that the drift of the associated Markov chain is close to a given nonnegative value. We do not describe in detail the construction, as it would take some space, but refer the reader to [4, Sections 7.1].
-
(a)
-
2.
Application Examples:
-
(a)
Some examples of PH/PH/1 queues collected in [4, Sections 7.1] for testing purposes. The construction depends on a parameter with . In this case the drift is .
-
(b)
The Markov chain of M/G/1 type associated with the infinitesimal generator matrix from the queuing model described in [6]. This is a complex queuing model, a BMAP/PHF/1/N model with retrial system with finite buffer of capacity and non-persistent customers. For the construction of the matrix we refer the reader to [6, Sections 4.3 and 4.5].
-
(a)
8.1 Synthetic Examples
The first test concern with the validation of the analysis performed above, regarding the convergence of fixed point iterations. In Table 1 we show the number of iterations required by different iterative schemes on Example 1.a with . Specifically we compare the traditional fixed point iteration, the -based fixed point iteration, the S-based fixed point iteration (12) and the relaxed fixed point iterations (13). For the latter case we consider the Sω-based iteration where is fixed a priori and the Sω(k)-based iteration where the value of is dynamically adjusted at any step according to the strategy (25) complemented with condition (15). The relaxed stationary iteration is applied for . The relaxed adaptive iteration is applied with . The iterations are stopped when the residual error is smaller than .
trad. | -based | S-based | Sω-based | Sω(k)-based | |
---|---|---|---|---|---|
1.0e-2 | 1447 | 731 | 724 | 515 | 65 |
496 | |||||
479 | |||||
1.0e-4 | 84067 | 42046 | 42037 | 30023 | 11771 |
28976 | |||||
28027 | |||||
1.0e-6 | 2310370 | 1154998 | 1155352 | 825121 | 329843 |
796693 | |||||
770250 |
The first four columns of Table 1 1 confirms the theoretical comparison of asymptotic convergence rates of classical fixed point iterations applied to a block tridiagonal matrix. Specifically, the -based and the S-based iterations are twice faster than the traditional iteration. Also, the relaxed stationary variants greatly improve the convergence speed. An additional remarkable improvement is obtained by adjusting dynamically the value of the relaxation parameter. Also notice that the Sω(k)-based iteration is guaranteed to converge differently to the stationary Sω-based iteration.
The superiority of the adaptive implementation over the other fixed point iterations is confirmed by numerical results on Example 1.b. In Table 2 for different values of we show the number of iterations required by different iterative schemes including also Algorithm 1 implemented in [8]. For comparison with the results in [8] here we set in the stopping criterion.
trad. | -based | S-based | Algorithm 1 | Sω(k)-based | |
---|---|---|---|---|---|
0.3 | 14 | 11 | 10 | 12 | 9 |
0.48 | 122 | 84 | 91 | 85 | 72 |
0.5 | 7497 | 5000 | 5622 | 5001 | 4374 |
0.55 | 53 | 37 | 39 | 39 | 32 |
Finally, we compare the convergence speed of the traditional, -based, S-based and Sω(k)-based fixed point iterations applied on the synthetic examples of M/G/1-type Markov chains described in [4]. In Figure 3 we report the semilogarithmic plot of the residual error in the infinity norm generated by the four fixed point iterations for two different values of the drift.


Observe that the adaptive relaxed iteration is about twice faster than the traditional fixed point iteration. The observation is confirmed in in Table 3 where we indicate the speed-up in terms of CPU-time with respect to the traditional fixed point iteration for different values of the drift .
-based | S-based | Sω(k)-based | |
---|---|---|---|
-0.1 | 1.6 | 1.5 | 2.1 |
-0.05 | 1.5 | 1.4 | 2.0 |
-0.01 | 1.6 | 1.5 | 2.1 |
-0.005 | 1.6 | 1.5 | 2.1 |
-0.001 | 1.6 | 1.4 | 2.2 |
-0.0005 | 1.6 | 1.5 | 2.2 |
-0.0001 | 1.7 | 1.6 | 2.4 |
In Figure 4 we repeat the set of experiments of Figure 3 with a starting stochastic matrix . Here the adaptive strategy is basically the same as used before where we select in the interval as the maximum value which maintains the nonnegativity of .


In this case the adaptive strategy seems to be quite effective in reducing the number of iterations. Other results are not so favorable and we believe that in the stochastic case the design of a general efficient strategy for the choice of the relaxation parameter is still an open problem.
8.2 Application Examples
The first set 2.a of examples from applications includes several cases of PH/PH/1 queues collected in [4]. The construction of the Markov chain depends on a parameter with and two integers which specify the PH distributions of the model. The Markov chain generates in this way is denoted as Example . Its drift is . In Tables 4 5 and 6 we compare the number of iterations for different values of . Here and hereafter the relaxed stationary Sω-based iteration is applied with .
trad. | -based | S-based | Sω-based | Sω(k)-based | |
---|---|---|---|---|---|
0.99 | 6708 | 3666 | 3638 | 1061 | 1069 |
0.999 | 53900 | 30018 | 29271 | 8559 | 8609 |
0.9999 | 386332 | 215660 | 209975 | 61608 | 65209 |
trad. | -based | S-based | Sω-based | Sω(k)-based | |
---|---|---|---|---|---|
0.99 | 3735 | 3241 | 2225 | 1582 | 1158 |
0.999 | 29162 | 25241 | 17460 | 12469 | 9331 |
0.9999 | 209275 | 181139 | 125689 | 89965 | 68967 |
trad. | -based | S-based | Sω-based | Sω(k)-based | |
---|---|---|---|---|---|
0.99 | 11372 | 9679 | 9416 | 8031 | 8204 |
0.999 | 87566 | 74611 | 72595 | 61992 | 63346 |
0.9999 | 597330 | 509108 | 495942 | 424141 | 433710 |
For comparison in Table 7 we show the number of iterations on Example of Table 5 starting with a stochastic matrix. We compare the traditional, -based and S-based iterations. We observe a rapid convergence profile and the fact that the number of iterations is independent of the drift value.
trad. | -based | S-based | |
---|---|---|---|
0.99 | 52 | 45 | 31 |
0.999 | 51 | 45 | 30 |
0.9999 | 51 | 45 | 30 |
For a more challenging example from applications in 2.b we consider the generator matrix from the queuing model described in [6]. In Table 8 we indicate the number of iterations for different values of the capacity .
trad. | -based | S-based | Sω-based | Sω(k)-based | |
---|---|---|---|---|---|
20 | 4028 | 3567 | 3496 | 3089 | 2826 |
30 | 4028 | 3567 | 3496 | 3089 | 2826 |
40 | 4028 | 3567 | 3496 | 3089 | 2826 |
9 Conclusion and Future Work
In this paper we have introduced a novel fixed point iteration for solving M/G/1-type Markov chains. It is shown that this iteration complemented with suitable adaptive relaxation techniques is generally more efficient than other classical iterations. Incorporating relaxation techniques into other different inner-outer iterative schemes as the ones introduced in [4] is an ongoing research.
References
- [1] P. Amodio and F. Mazzia. A parallel Gauss-Seidel method for block tridiagonal linear systems. SIAM J. Sci. Comput., 16(6):1451–1461, 1995.
- [2] Z. Z Bai. A class of iteration methods based on the Moser formula for nonlinear equations in Markov chains. Linear Algebra Appl., 266:219–241, 1997.
- [3] D. A. Bini, G. Latouche, and B. Meini. Numerical methods for structured Markov chains. Numerical Mathematics and Scientific Computation. Oxford University Press, New York, 2005. Oxford Science Publications.
- [4] D. A. Bini, G. Latouche, and B. Meini. A family of fast fixed point iterations for M/G/1-type Markov chains. IMA Journal of Numerical Analysis, 42(2):1454–1477, 02 2021.
- [5] C. Brezinski and M. Redivo Z. Extrapolation methods, volume 2 of Studies in Computational Mathematics. North-Holland Publishing Co., Amsterdam, 1991. Theory and practice, With 1 IBM-PC floppy disk (5.25 inch).
- [6] S. Dudin, A. Dudin, O. Kostyukova, and O. Dudina. Effective algorithm for computation of the stationary distribution of multi-dimensional level-dependent Markov chains with upper block-Hessenberg structure of the generator. J. Comput. Appl. Math., 366:112425, 17, 2020.
- [7] L. Gemignani and F. Poloni. Comparison theorems for splittings of M-matrices in (block) Hessenberg form. BIT Numerical Mathematics, 2022.
- [8] C. H. Guo. On the numerical solution of a nonlinear matrix equation in Markov chains. Linear Algebra Appl., 288(1-3):175–186, 1999.
- [9] G. Latouche and V. Ramaswami. A logarithmic reduction algorithm for quasi-birth-death processes. J. Appl. Probab., 30(3):650–674, 1993.
- [10] H. Lu. Stair matrices and their generalizations with applications to iterative methods. I. A generalization of the successive overrelaxation method. SIAM J. Numer. Anal., 37(1):1–17, 1999.
- [11] B. Meini. New convergence results on functional iteration techniques for the numerical solution of type Markov chains. Numer. Math., 78(1):39–58, 1997.
- [12] G. Meurant. Domain decomposition preconditioners for the conjugate gradient method. Calcolo, 25(1-2):103–119 (1989), 1988.
- [13] C. Meyer. Matrix analysis and applied linear algebra. Society for Industrial and Applied Mathematics (SIAM), Philadelphia, PA, 2000. With 1 CD-ROM (Windows, Macintosh and UNIX) and a solutions manual (iv+171 pp.).
- [14] M. F. Neuts. Matrix-geometric solutions in stochastic models, volume 2 of Johns Hopkins Series in the Mathematical Sciences. Johns Hopkins University Press, Baltimore, Md., 1981. An algorithmic approach.
- [15] R. J. Plemmons. -matrix characterizations. I. Nonsingular -matrices. Linear Algebra Appl., 18(2):175–188, 1977.
- [16] V. Ramaswami. A stable recursion for the steady state vector in Markov chains of type. Comm. Statist. Stochastic Models, 4(1):183–188, 1988.