Multilevel Richardson-Romberg and Importance Sampling in Derivative Pricing
Abstract
In this paper, we propose and analyze a novel combination of multilevel Richardson-Romberg (ML2R) and importance sampling algorithm, with the aim of reducing the overall computational time, while achieving desired root-mean-squared error while pricing. We develop an idea to construct the Monte-Carlo estimator that deals with the parametric change of measure. We rely on the Robbins-Monro algorithm with projection, in order to approximate optimal change of measure parameter, for various levels of resolution in our multilevel algorithm. Furthermore, we propose incorporating discretization schemes with higher-order strong convergence, in order to simulate the underlying stochastic differential equations (SDEs) thereby achieving better accuracy. In order to do so, we study the Central Limit Theorem for the general multilevel algorithm. Further, we study the asymptotic behavior of our estimator, thereby proving the Strong Law of Large Numbers. Finally, we present numerical results to substantiate the efficacy of our developed algorithm.
Keywords: Multilevel Monte-Carlo Estimator; Richardson-Romberg Extrapolation; Importance Sampling; Path-dependent Functional; Option Pricing
1 Introduction
At the core of financial engineering lie the fundamental questions of pricing, portfolio management, and risk mitigation, which for all practical purposes, necessitates resorting to computational techniques, particularly, the Monte Carlo simulation approach. The emergence of this technique, as the key computational approach in the finance industry, can be attributed to the requirement of simulating high-dimensional stochastic models, which in turn is a result of the growth in computational complexity, pertaining to the dimension of the problem itself. More specifically, the main purpose of the usage of this procedure is to numerically approximate the expected value , where is a functional of the random variable , with (in case of the concerned applications) being driven by a stochastic differential equation (SDE).
The increasing interest and application of the Monte Carlo approach in finance industry, can also be attributed to the development of Multilevel Monte Carlo (MLMC) technique, due to Giles [1], as it delivers remarkable improvement in computational complexity, over the standard Monte Carlo technique, in the biased framework. The interested readers may refer to the website of Giles ( http://people.maths.ox.ac.uk/~gilesm/mlmc_community.html) for an exhaustive and detailed listing of the flow of research concerning MLMC, both theoretical as well as application. The technique of MLMC draws its foundations from the technique of parametric integration, developed in [2], which was extended in [1], in order to construct a MLMC path method. While the control variate viewpoint used in [1] is similar to that of [2], however, unlike in case of [2], for the construction in [1], the random variable is infinite dimensional, with respect to the Brownian path, and there is no parametric integration that is involved. MLMC, as the name suggests, makes use of various levels of resolution, from the coarsest to the finest, where, from the perspective of the control variate, the simulation executed using the coarsest path, is used as a control variate, while carrying out the finer path based estimation. The introduction to MLMC was accomplished through the Euler-Maruyama discretization, a thorough analysis of which suggested that in order to estimate the variance of MLMC, one has to examine the strong convergence properties.
In another 2008 article [3], Giles presented a Milstein scheme for discretization of an SDE, an approach that led to improvement in the estimation of the variance. Further, it was demonstrated that, it could be more prudent to make use of different estimators for the coarser level and the finer level. The numerical results presented, as well as the thorough numerical analysis carried out in [4], demonstrates the efficacy of the method in case of option pricing problems. With the improvement in variance convergence, resulting from the usage of the Milstein scheme, the extension of the scheme in case of higher dimensional SDEs (involving the concept of Levy area) can be accomplished. However, in practice, there is no efficient way to simulate the Levy area, and the interested reader may refer to [5] for more discussion on this. In order to tackle this issue, the authors in [6], introduced the antithetic MLMC estimator, based on the classical antithetic variance reduction technique, wherein the idea of antithetic MLMC exploits the flexibility of the general MLMC estimator.
Giles and Waterhouse [7], published an extension of MLMC using Quasi-Monte Carlo (QMC) sample paths instead of Monte Carlo sample paths. The numerical results presented showed the effectiveness of QMC for SDE applications. However, the results presented in the paper were not supported by any theoretical development in this context. Nevertheless, recently, there has been some promising contribution towards the theoretical development of Quasi MLMC (QMLMC), for which one may refer to [8] and references therein. It has been shown that, under certain conditions, the QMLMC leads to multilevel methods, with a complexity which is with . In [9], Rhee and Glynn introduced a new approach of constructing an unbiased estimator, given a family of the biased estimators. The idea presented by them is closely related to that of MLMC, wherein the finest level of estimation is chosen, contingent on the level of accuracy. The results presented in [9] demonstrated a significant improvement in the computational cost over the standard MLMC. Also, they proved the square root convergence of the estimator, given that the strong order of convergence is greater than for the path functionals.
2 Preliminaries
In this section we discuss the necessary preliminaries and assumptions, as a prelude to the comprehensive study of the algorithm developed in this work. We present a brief outline of Multilevel Richardson-Romberg estimator, introduced in [10] and recall the importance sampling algorithm in the context of the standard Monte-Carlo algorithm.
2.1 Multilevel Richardson-Romberg
In [1], Giles explored the Richardson extrapolation in the context of both the MLMC and the standard Monte Carlo. The MLMC on its own was significantly better than the Richardson extrapolation. However, taken together, they worked even better. Lemaire and Pages [10] took this approach and undertook further comprehensive error analysis. They combined the method developed in [1] and Multistep Richardson extrapolation in order to minimize the cost of simulation.
Suppose, we are interested in estimating the expected payoff value i.e., in an option pricing problem, with finite time horizon , where is a process with values in , governed by the following SDE,
| (1) |
where, is a -dimensional Brownian motion on a filtered probability space
, with and being the functions satisfying the following condition:
| (A.1) |
The assumption (A.1) ensures the existence and uniqueness [11] of the strong solution of (1). Now, in order to estimate one should be able to simulate . However, except for a handful of cases, where one can devise an analytical or semi-analytical solution, we must resort to discretization schemes to perform the simulation. As discussed above, we use the Euler-Maruyama discretization scheme in order to simulate equation (1).
Let, be the refinement factor and let , where is the level refinement. Let be the time step size on level . For example, the Euler scheme on level is given by,
where, . We approximate by , where is the finest level. In [10] the authors proposed a methodology combining the order bias cancellation of Multilevel Richardson Romberg (ML2R), with variance control of MLMC, to solve the problem of improving the computational complexity, along with determining the optimal parameters in order to achieve the desired root-mean-squared error, with minimum computational effort. The study was based on weak error and strong error assumptions, on the sequence , where . The assumptions [10, 11, 12] are stated as follows:
-
(A)
Weak Error:
(2) -
(B)
Strong Error:
(3)
With the consideration of the above assumptions, the ML2R estimator is defined as,
| (4) |
where are the optimal parameters obtained as the solution to,
| (5) |
where the cost function is given by [12],
Further, in the above equation, is the number of sample paths on level , and are the weights given by,
| (6) |
where is the solution to the Vandermonde system , with the Vandermonde matrix being defined by,
| (7) |
Here, is the weak error rate as defined above. The interested reader may refer to [10] for the construction of the optimal parameters, as the closed solution to equation (5). Here we tabulate the explicit values of these parameters required to achieve the root-mean-squared error , with the following constants being used:
-
(A)
.
-
(B)
.
| , | |
The above estimator was highly effective whenever there is a strong order of convergence i.e., , as it achieves for and for , contrary to and , respectively, which is achieved by the standard MLMC.
2.2 Importance Sampling
Now following the idea of [13], we consider a parametric family of stochastic process , with , governed by the following SDE,
| (8) |
By the Girsanov’s Theorem, we know that there exists a probability measure equivalent to such that,
| (9) |
under which, the process is a Brownian motion. Therefore,
| (10) |
The efficiency of the Monte Carlo method is considerably dependent on how small the variance is in the process of estimation. Among the many variance reduction techniques, available in literature (for improvement of efficiency of the Monte Carlo method), importance sampling is one such variance reduction technique, which is known for its efficiency [14]. In general, the idea of parametric importance sampling is to introduce the parametric transformation such that ,
More specifically, we consider the general problem of estimating , where is a -dimensional random variable. Also if is the multivariate density, then,
where . Now the idea of importance sampling Monte Carlo method is to estimate , where is given by,
| (11) |
In order to solve the above problem one can use the so-called Robbins-Monro algorithm that deals with a sequence of random variable , which approximates accurately. However, the convergence of this algorithm requires restrictive conditions known as the non explosion condition given by [15],
To overcome these restrictive conditions, a truncation based procedure was introduced by Chen in [16, 17] which was furthered in [18, 19]. This was then proposed in the context of importance sampling [13]. An unconstrained procedure to approximate , by using the regularity of the involved density extensively, was introduced in [20] along with the proof of the convergence of the algorithm.
The authors in [15] studied importance sampling in the context of statistical Romberg integration, which is a one-step generalization of the standard Monte Carlo integration. The paper used both constrained, as well as unconstrained optimization routines to approximate . Further, they proved the almost sure convergence of both constrained and unconstrained versions of the optimization algorithm, towards the optimal parameter , under discretized diffusion setting. The idea was then extended by [21] and [22] in the context of MLMC. The study carried out in [21] is the extension of [15] in the paradigm of MLMC. However [22] used sample average approximation to approximate . Also, their algorithm performed an approximation procedure on each level of resolution, thus further enhancing the overall performance of the adaptive algorithm.
In this paper, we develop a hybrid algorithm incorporating ML2R and adaptive importance sampling procedure developed in [13] in the context of standard Monte Carlo in order to reduce the overall computational time, while achieving the desired accuracy. In [15], the authors studied the discretized version of the asymptotic variance obtained in the central limit theorem associated with the statistical Romberg method. The study was performed to determine the optimal parameter , to minimize the variance of the Monte Carlo estimator. In [22], the authors extended the algorithm developed in [15] in the context of MLMC, building upon the Central Limit Theorem for the Euler MLMC studied in [23]. As discussed in [23], the optimal value of is obtained by using a constrained and unconstrained version of the stochastic Robbins-Monro algorithm, whereas in [22] it is determined using deterministic Newton-Raphson’s method. However, the algorithm was developed in the context of Euler discretization, and only European option pricing was studied, with no reference to path-dependent functionals. In this paper, we develop upon the Central Limit Theorem, for the general multilevel approach studied in [12], that incorporates the results for path-dependent functionals as well. We look at the stochastic optimization approach to approximate . Also, we use the Milstein scheme, as well as Euler in order to discretize the underlying SDE. In order to highlight the novelty and contribution of this work, we present a comparison of the work accomplished, with the existing literature, in Table 2
ML2R IS Hybrid Algorithm: (IS+MC Version) Euler Scheme Milstein Scheme Vanilla Option Exotic Option Reference Yes No No Yes No Yes Yes [10] No Yes Yes ( With Standard Monte carlo) Yes No Yes No [13] No Yes Yes (With MLMC) Yes No Yes No [22] No Yes Yes (With Statistical Romberg Integration) Yes No Yes No [15] No Yes Yes (Adaptive MLMC) Yes No Yes No [21] Yes Yes Yes Yes Yes Yes Yes Present Work
3 Algorithm
In this paper, we follow the idea presented in [22] to develop our coupling algorithm, which we will refer to as AISML2R:
| (12) | |||||
By applying Monte Carlo method to each level with samples in equation (12), we get,
| (13) | |||||
In the above estimator, the value of on level , is calculated using the values of the payoff, and obtained in the previous path simulation, using the stochastic approximation algorithm studied in Section 3.2 below, thus describing the adaptive nature of the algorithm.
3.1 Optimization Problem
In this subsection, we deal with the problem of finding the optimal value of , to minimize the variance on level . For the sake of brevity, we let,
Further let,
We start our discussion by recalling the Central Limit Theorems for ML2R from [12],
Theorem 1 (Central Limit Theorem, ).
Assume strong error for and that is uniformly integrable. We set:
with,
Assuming weak error for , we have,
Theorem 2 (Central Limit Theorem, ).
Assume strong error for and that is uniformly integrable. Assume furthermore , we set:
Assuming weak error for , we have,
Based of the two theorems stated above we develop the optimization algorithm for and .
3.1.1
In Theorem 1, let
where,
and therefore,
From the practical point of view, it is necessary to use the truncated version of the above summation. In the thorough study carried out in [12], it was proven that for ,
Therefore, owing to the above result and motivated by the analysis pertaining to the Central Limit Theorem carried out in [12], we can formulate the problem for as,
| (14) | ||||
where,
| (15) |
Using the Girsanov’s theorem, we can see that,
| (16) |
where,
Similarly for , we have,
| (17) |
3.1.2 Case II:
Based on Theorem 2, we define be the level approximation of , where,
Therefore, we have,
| (18) |
We follow the similar line of development as for to formulate the problem for . As one can observe from (18) that, in order to minimize on level , we only need to minimize . Therefore, we formulate the optimization problem as follows ,
| (19) |
3.2 Stochastic Algorithm
In this section we outline the stochastic approximation algorithm to estimate . Based on the discussion in the previous section, we perform the discretization, such as Euler or Milstein, of the underlying SDE to approximate ’s. Whereas for stochastic approximation we resort to the Robbins-Monro algorithm, to approximate the value of . The aim is to construct a sequence of , such that, .
Let be a compact convex subset such that . Then the sequence is constructed recursively as follows,
| (20) |
where . Also is a decreasing sequence of positive real numbers satisfying,
| (21) |
where for ,
| (22) |
and for ,
| (23) |
It may be noted that in this paper, we use the constrained version of the Robbins-Monro algorithm in order to estimate the approximate value of for various level of resolutions.
4 Analysis
In this section we study our algorithm in detail. We start our discussion by establishing the existence and uniqueness of the for various level of resolution. Further we carry out the asymptotic analysis of our AISML2R estimator, thereby establishing the Strong Law of Large Numbers.
4.1 Existence and Uniqueness of .
Before, establishing the existence and uniqueness of , we recall results related to the weights , from [12] necessary for our study. Accordingly, let us define,
The following result was proved in [12],
Lemma 1.
Let . and the associated weights , be as given in (6).
-
(a)
and .
-
(b)
The weights are uniformly bounded,
-
(c)
For every ,
-
(d)
Let be a bounded sequence of positive real numbers. Let and assume that when . Then the following limits hold:
With above results at our disposal, we present a series of results establishing the existence and uniqueness of the optimal parameter on various level of resolution. For the most part the proof follows line of argument similar to that presented in [22, 15, 13]. We start our discussion bu proving the following lemma, necessary for the existence and uniqueness results.
Lemma 2.
Let and assume the -strong error rate assumption, i.e.,
| (24) |
Then, for and some constant .
Proof.
We prove the above lemma for . The proof for follows a similar line of argument and is relatively easy to prove. Consider,
Then, for ,
Taking expectation on both sides and applying Hölder’s inequality for all , we get,
It is clear that is bounded. As for we fallback to the -strong error assumption. Therefore, we have,
| (25) |
where,
| (26) |
The boundedness of can be derived from the definition of and from (b) and (c) of Lemma 1. Hence, from the above analysis, we can conclude that, is bounded for . ∎
We state now, the result pertaining to the existence and uniqueness of optimal parameters on various level of resolutions.
Proposition 1.
Suppose and satisfies assumption (A.1). Let be such that,
Further, if satisfies the following conditions,
-
(1)
.
-
(2)
.
Then the function is and strictly convex with for all . Moreover, there exists an unique such that .
Proof.
To prove the proposition, we refer to the proof in [15, 22], in our context. Here we discuss the proof for and . For and the proofs are relatively easy and can be reproduced in the similar way. To begin with, one can observe that is a continuously infinitely differentiable function with respect to , and . Therefore, the first derivative of the map is equal to . As we have already seen in Lemma 2 that the is bounded, therefore by Lebesgue’s theorem we can conclude that is , with for all . A similar line of argument also proves that is . The Hessian of is given as follows,
Since, , therefore, for all , . Hence, we can conclude that is strictly convex. As a consequence, there exists a minimum , such that . Further,the unique minimum is attained for a finite value of , it is enough to prove that . This can be proved using Fatou’s lemma as,
This completes the proof. ∎
As the above proposition guarantees the existence and uniqueness of the for various level of resolutions, we now state the main result proving the convergence of the sequence , as , under the assumptions of the Lemma 2 and Proposition 1.
Theorem 3.
If is such that, and , then , as .
Proof.
To prove the above result we follow the assertion made in Theorem A.1 [24] that suggest that in order to prove , where the sequence is constructed through a constrained version of the Robbins Monro, we need to verify two conditions, namely,
-
(1)
.
-
(2)
Non explosion condition: such that .
As we know, and in the previous proposition it was proven that is convex, therefore as a consequence, we prove that,
| (27) |
For the non-explosion condition, we use the Cauchy-Schwarz inquality,
| (28) |
Under the assumption and following the similar line of argument of Lemma 2, it is easy to prove that there exists a constant , such that,
| (29) |
Further using the fact the , we can deduce that,
| (30) |
which in turns concludes the non explosion condition. This proves the almost sure convergence of to as . ∎
4.2 Main Result
Theorem 4.
Let be a family of sequence converging to as . Moreover for assume weak error for all and . Furthermore assume the -strong error rate assumption,i.e.,
| (31) |
If is a sequence of positive real numbers such that , then the AISML2R estimator satisfies,
| (32) |
We will assume the following notation,
where we set,
| (33) |
and
| (34) |
Therefore, we have,
A thorough analysis carried out in Section 4.2 of [12] shows that the last term in the equation converges to zero as . We start our discussion by proving the following lemma.
Lemma 3.
Let and . Then there exist a positive constant such that,
Proof.
By Minkowski’s Inequality, we have,
| (35) | ||||
In order to bound of (35), we apply the Holder’s Inequality,
| (36) | ||||
Therefore, from above analysis, we have,
| (37) |
Further for of (35), the assumption (31) yields,
| (38) |
Now from (35) and (37), we get,
| (39) |
Combining above inequality and (38) yields,
| (40) |
where,
∎
Proposition 2.
Let and for and . Then there exists a constant , such that,
| (41) |
Proof.
We start our discussion by observing that as , there exists a positive constant such that . Now, we define the following filtration , where . With this filtration, one can readily observe that is a martingale with respect to . Now consider the following definition,
| (42) |
Since is martingale, therefore by Rosenthal’s inequality [25], we have,
| (43) |
Now, from the previous Lemma, one can easily conclude that,
| (44) |
As for the first term, we use Theorem A.8 of [25], to obtain,
| (45) | ||||
where the last inequality is the consequence of the previous Lemma. Therefore, we have for ,
| (46) |
Now, let . Therefore, we have,
| (47) |
Now, consider the following,
| (48) |
As one can observe that, are independent random variable in , therefore, by a version of Rosenthal’s inequality [25], we have,
| (49) | ||||
For , we have,
Therefore,
| (50) |
We know that, , and as a result we have,
Further, owing to the expression for , we have,
Since, [12], therefore, combining everything we get,
| (51) |
where, . Now as a consequence of Lemma 4.5 in [12], with , we have,
| (52) |
Moreover, it is easy to prove that,
| (53) |
With all the preceding discussion, we have,
| (54) |
where,
| (55) |
∎
With above results in our hand we are ready to prove the Strong Law of Large Numbers.
Proof of Theorem 4.
We start our proof by proving, as . Clearly, as a consequence of our assumption on , we have,
Hence, by Beppo-Levi’s Theorem, , and as an implication we have as . We now turn our attention to proving as . It is clear that as , , which in turns implies . As one can observe that, is martingale. Therefore, as an application of Rosenthal’s inequality for , we have,
| (56) | ||||
where the last equality is the consequence of being independent of and being measurable. Further, due to to Grisanov’s theorem, we introduce a couple of random variable and independent of , justifying the last equality. Further, as , therefore there exists a such that for . As a consequence, . Now for ,
Therefore, as consequence of Theorem 3 and Lebesgue theorem, we obtain that,
Now as the application of Cesaro’s lemma in equation (55) one can easily conclude that, as . ∎
5 Numerical Results
In this section we will illustrate the efficacy of the algorithm introduced, through a couple of examples. In order to keep the things simple we restrict ourselves to one dimensional problems, arising in the realm of mathematical finance. More specifically, we look at the results in the case of European call and lookback call options. We start our computation by determining the optimal parameters required to perform the simulations. In order to do so, we perform a pre-simulation to approximate the value of , and , necessary for the computation of the optimal parameters. For computing we refer to the formula presented in [10], i.e.,
| (57) |
Here we set . As for , we perform a small prior simulations and empirically calculate the variance. Further, the value of is calculated as . We use the above parameters to perform the simulation of ML2R. As for performing the adaptive simulation, we calculate , and , to compute the necessary parameters. To compute these structural parameters, we perform a dummy optimization procedure based on the Robbins-Monro algorithm, to calculate the value of . We calculate the value of using the same formula used for the calculation of . We again run a small simulation as before, to calculate the value of , further calculating . Based on these structural parameters, and using the formulas from Table 1, we calculate the parameters to perform the adaptive simulation. In order to perform the simulation for both the adaptive and the non-adaptive algorithm, we use the Milstein scheme, unlike in the original study carried out in [22], which uses Euler-Maruyama scheme, to discretize the underlying SDE. The choice of the scheme is due to the establishment of the Central Limit Theorem in [12] for , which provides us with the freedom to use discretization scheme of higher order. In order to determine these structural parameters, we input the value of , the desired root-mean-squared error. For, the purpose of our numerical experiments, we set , where .
It may be noted that in order to simulate in the probability space i.e., for adaptive simulation, the calculation of and are rather sub-optimal. This is because the formulas associated with these calculations do not consider the number of iterations required to estimate , on various level of resolutions. However, while using Milstein discretization in the parametric probability space, it was observed that the values for and , calculated using the procedure described above is sufficient to achieve an accuracy comparable to that achieved by ML2R, whereas, under the Euler discretization, twice the number of paths generated through the formula are used. As we will see, even by increasing the number of sample paths the adaptive algorithm outperforms standard ML2R, thereby achieving the desired level of accuracy. We now briefly describe the numerical scheme used to perform simulation in either probability space. Consider the general one-dimensional SDE,
| (58) |
The Milstein discretization of the above equation is given as,
| (59) |
In the above equation, is the uniform time-step, , and , with . However, under the parametric change of measure, with as the parameter, we have the following SDE,
| (60) |
where, . Therefore, we have the following discretization:
As for the Euler scheme, we have,
| (61) |
We use the above discretization schemes, in order to simulate the SDE in the probability space . Further, for the purpose of our numerical experiments, we use geometric Brownian motion to determine the value of the asset at time . As the process is the solution to the SDE,
therefore, we can use the Euler or the Milstein scheme described above, in order to perform the Monte-Carlo simulations. In the above process, denotes the risk-less interest rate, denotes the volatility and is a standard Brownian motion defined on the probability space .
For the purpose of estimating optimal for each level of resolutions, we use the algorithm described in Section 3.2. In all the numerical experiments carried out below, we consider and . For the practical purpose, we stop the stochastic approximation procedure after finite iterations. Further we use Rupert and Poliak method for stabilizing the convergence of the described algorithm, i.e., instead of using , we use , on level . In order to compare the results generated through ML2R and AISML2R, we define the following improvement factor (if) as follows,
| (62) |
5.1 European option
The payoff function in the case of European call option is described as below,
| (63) |
For the purpose of the practical implementation, we have considered , , , and [10]. With these parameters, the price obtained through closed form solution is, . Also, we have chosen for the Milstein scheme and for the Euler scheme. Further, under the Milstein discretization, we stop the stochastic algorithm after iterations, whereas, we under go only iteration under the Euler scheme.
Table 3 and Table 4 demonstrate potency of the AISML2R over ML2R, under Milstein and Euler discretization, respectively. Figure 1 and Figure 2 graphically represents the performance of both the estimators under Milstein and Euler schemes, respectively. It is quite evident from the tabulated results, as well as the graphical representation, that the effectiveness of the AISML2R over ML2R increases with the requirement of greater accuracy. The value of goes from , for the worst case to in the best case, while simulating under the Milstein scheme. On the other hand, under the Euler scheme, the value of goes from in the best case to in the best case. These values suggest that AISML2R can achieve desired root-mean-squared error much faster as compared to ML2R.
| AISML2R | |||||
| k | N | variance | bias | rmse | time |
| 3 | 1.05E+05 | 3.75E-03 | 4.93E-02 | 6.18E-02 | 5.12E+00 |
| 4 | 4.19E+05 | 7.33E-04 | 2.21E-02 | 2.72E-02 | 8.51E+00 |
| 5 | 1.68E+06 | 1.95E-04 | 1.17E-02 | 1.43E-02 | 2.12E+01 |
| 6 | 6.70E+06 | 4.54E-05 | 5.68E-03 | 7.01E-03 | 6.81E+01 |
| 7 | 2.68E+07 | 6.07E-06 | 2.14E-03 | 2.48E-03 | 2.57E+02 |
| 8 | 1.05E+08 | 2.28E-06 | 1.14E-03 | 1.51E-03 | 1.02E+03 |
| 9 | 4.20E+08 | 5.69E-07 | 5.99E-04 | 7.55E-04 | 4.14E+03 |
| ML2R | |||||
| k | N | variance | bias | rmse | time |
| 3 | 1.17E+05 | 1.53E-02 | 1.07E-01 | 1.24E-01 | 1.04E+00 |
| 4 | 4.68E+05 | 4.41E-03 | 5.47E-02 | 6.66E-02 | 3.52E+00 |
| 5 | 1.87E+06 | 9.93E-04 | 2.60E-02 | 3.17E-02 | 1.44E+01 |
| 6 | 7.48E+06 | 1.43E-04 | 1.05E-02 | 1.26E-02 | 5.65E+01 |
| 7 | 2.99E+07 | 4.39E-05 | 5.29E-03 | 6.62E-03 | 2.28E+02 |
| 8 | 1.17E+08 | 1.45E-05 | 2.96E-03 | 3.82E-03 | 9.01E+02 |
| 9 | 4.69E+08 | 4.28E-06 | 1.61E-03 | 2.14E-03 | 3.73E+03 |
| AISML2R | |||||
| k | N | variance | bias | rmse | time |
| 3 | 9.28E+04 | 1.99E-02 | 4.39E-02 | 1.48E-01 | 3.69E+00 |
| 4 | 3.71E+05 | 3.15E-03 | 1.25E-02 | 5.75E-02 | 7.12E+00 |
| 5 | 1.48E+06 | 1.58E-03 | 2.39E-03 | 3.98E-02 | 2.10E+01 |
| 6 | 5.94E+06 | 3.23E-04 | 6.25E-03 | 1.90E-02 | 7.83E+01 |
| 7 | 2.92E+07 | 9.99E-05 | 4.76E-04 | 1.00E-02 | 4.51E+02 |
| 8 | 1.17E+08 | 1.50E-05 | 6.49E-05 | 3.87E-03 | 1.85E+03 |
| 9 | 4.67E+08 | 5.40E-06 | 4.10E-04 | 2.36E-03 | 7.79E+03 |
| ML2R | |||||
| N | variance | bias | rmse | time | |
| 3 | 1.45E+05 | 2.44E-02 | 5.25E-03 | 1.56E-01 | 1.53E+00 |
| 4 | 5.80E+05 | 5.71E-03 | 5.18E-03 | 7.57E-02 | 5.83E+00 |
| 5 | 2.32E+06 | 1.19E-03 | 8.79E-03 | 3.55E-02 | 2.37E+01 |
| 6 | 9.28E+06 | 4.93E-04 | 2.27E-03 | 2.23E-02 | 9.64E+01 |
| 7 | 4.54E+07 | 1.14E-04 | 1.52E-03 | 1.08E-02 | 5.68E+02 |
| 8 | 1.81E+08 | 2.47E-05 | 8.38E-05 | 4.97E-03 | 2.35E+03 |
| 9 | 7.26E+08 | 8.37E-06 | 6.35E-04 | 2.96E-03 | 1.03E+04 |
5.2 Lookback Option
In this section we consider a partial lookback call option, whose payoff is defined as,
| (64) |
The parameters for the purpose of the practical implementation, taken from [10], are , , and . Also, the value of is set as . Further, the refinement factor used to perform the simulation at various level of dicretization, is set as , for either case. With these parameters, the price obtained through closed form solution is . The simulation is carried out using both Milstein and Euler scheme, as described above. In order to exploit the full essence of Milstein discretization, we use the formula described below in order to simulate the value of on various refinement level, [3].
| (65) |
Further, we set . In this case we stop the stochastic algorithm after iterations for either case.
The results summarized in Table 5 and Table 6 shows the efficacy of AISML2R over ML2R in the case of both Euler and Milstein discretization scheme, respectively. Figure 3 and Figure 4 graphically represents the performance of both estimators under Milstein and Euler schemes, respectively. Similar to the case of European Option, we can easily observe that the efficiency of AISML2R over ML2R increases as more accuracy is required. The value of if goes from in the worst case to in the best case, while simulating under Milstein scheme. Under the Euler scheme, the value goes from in the worst case to in the best case.
| AISML2R | |||||
| k | N | variance | bias | rmse | time |
| 3 | 2.86E+03 | 1.44E-02 | 2.41E-02 | 1.22E-01 | 7.99E-01 |
| 4 | 1.14E+04 | 3.51E-03 | 2.13E-03 | 5.93E-02 | 1.03E+00 |
| 5 | 4.58E+04 | 7.14E-04 | 8.31E-03 | 2.80E-02 | 1.67E+00 |
| 6 | 1.83E+05 | 1.85E-04 | 9.92E-03 | 1.68E-02 | 3.78E+00 |
| 7 | 7.33E+05 | 6.73E-05 | 9.88E-03 | 1.28E-02 | 1.32E+01 |
| 8 | 3.39E+06 | 1.45E-05 | 8.88E-03 | 9.66E-03 | 7.14E+01 |
| 9 | 1.36E+07 | 5.06E-06 | 8.81E-03 | 9.09E-03 | 2.77E+02 |
| ML2R | |||||
| k | N | variance | bias | rmse | time |
| 3 | 6.43E+03 | 1.29E-02 | 1.45E-02 | 1.15E-01 | 2.97E-01 |
| 5 | 2.57E+04 | 4.37E-03 | 1.17E-03 | 6.61E-02 | 6.39E-01 |
| 5 | 1.03E+05 | 8.24E-04 | 1.19E-02 | 3.11E-02 | 1.91E+00 |
| 6 | 4.12E+05 | 2.43E-04 | 8.33E-03 | 1.77E-02 | 7.42E+00 |
| 7 | 1.65E+06 | 6.59E-05 | 9.35E-03 | 1.24E-02 | 3.02E+01 |
| 8 | 7.93E+06 | 1.39E-05 | 9.05E-03 | 9.79E-03 | 1.76E+02 |
| 9 | 3.17E+07 | 2.62E-06 | 9.08E-03 | 9.22E-03 | 7.10E+02 |
| AISML2R | |||||
| k | N | variance | bias | rmse | time |
| 3 | 1.59E+03 | 1.29E-02 | 7.52E-02 | 1.36E-01 | 7.71E-01 |
| 4 | 6.38E+03 | 4.31E-03 | 2.92E-02 | 7.19E-02 | 9.21E-01 |
| 5 | 2.55E+04 | 1.01E-03 | 2.99E-02 | 4.36E-02 | 1.30E+00 |
| 6 | 1.02E+05 | 2.27E-04 | 2.64E-02 | 3.04E-02 | 2.52E+00 |
| 7 | 4.08E+05 | 7.20E-05 | 2.43E-02 | 2.57E-02 | 7.40E+00 |
| 8 | 1.61E+06 | 1.14E-05 | 2.36E-02 | 2.38E-02 | 3.27E+01 |
| 9 | 6.44E+06 | 4.40E-06 | 2.27E-02 | 2.28E-02 | 1.17E+02 |
| ML2R | |||||
| k | N | variance | bias | rmse | time |
| 3 | 5.16E+03 | 1.24E-02 | 2.65E-02 | 1.14E-01 | 2.93E-01 |
| 5 | 2.06E+04 | 3.00E-03 | 2.76E-02 | 6.13E-02 | 5.02E-01 |
| 5 | 8.25E+04 | 8.71E-04 | 2.90E-02 | 4.14E-02 | 1.33E+00 |
| 6 | 3.30E+05 | 4.09E-04 | 2.34E-02 | 3.09E-02 | 4.93E+00 |
| 7 | 1.32E+06 | 7.61E-05 | 2.48E-02 | 2.63E-02 | 1.99E+01 |
| 8 | 5.21E+06 | 1.71E-05 | 2.39E-02 | 2.42E-02 | 8.31E+01 |
| 9 | 2.08E+07 | 4.02E-06 | 2.26E-02 | 2.27E-02 | 3.30E+02 |
6 Conclusion
This paper presents a novel combination of ML2R and importance sampling algorithm developed upon the studies carried out in [10, 13, 15, 21, 22]. As evident from the numerical results presented in Section 5, the adaptive estimator outperforms AISML2R whenever a high degree of accuracy is required. One of the critical features of this, presented in the paper, is the use of a higher order scheme for discretizing the underlying SDE. Therefore, accommodating the antithetic feature developed by Giles [6] to simulate higher dimensional SDEs, using the Milstein scheme, in the realm of the above-developed algorithm, can substantially affect the overall computational time. Further, the completely automated nature of the algorithm gives its edge over the already existing estimators. However, it may be pointed out that a small amount of fine-tuning may be required while dealing with the Euler discretization scheme. In this paper, besides establishing the existence and uniqueness of the optimal parameter on various levels of resolutions (Section 4.1), we have also proved the Strong Law of Large Numbers in Section 4.2, proving the convergence of the adaptive estimator to the standard expectation. However, proving the Central Limit theorem still requires further detailed study of our estimator.
References
- [1] M. B. Giles, “Multilevel monte carlo path simulation,” Operations research, vol. 56, no. 3, pp. 607–617, 2008.
- [2] S. Heinrich, “Multilevel monte carlo methods,” in International Conference on Large-Scale Scientific Computing, pp. 58–67, Springer, 2001.
- [3] M. Giles, “Improved multilevel monte carlo convergence using the milstein scheme,” in Monte Carlo and Quasi-Monte Carlo Methods 2006, pp. 343–358, Springer, 2008.
- [4] M. B. Giles, K. Debrabant, and A. Rössler, “Analysis of multilevel monte carlo path simulation using the milstein discretisation,” Discrete & Continuous Dynamical Systems-B, vol. 24, no. 8, p. 3881, 2019.
- [5] M. Giles and L. Szpruch, “Multilevel monte carlo methods for applications in finance,” Recent Developments in Computational Finance: Foundations, Algorithms and Applications, pp. 3–47, 2013.
- [6] M. B. Giles and L. Szpruch, “Antithetic multilevel monte carlo estimation for multidimensional sdes,” in Monte Carlo and Quasi-Monte Carlo Methods 2012, pp. 367–384, Springer, 2013.
- [7] M. B. Giles and B. J. Waterhouse, “Multilevel quasi-monte carlo path simulation,” in Advanced financial modelling, pp. 165–182, De Gruyter, 2009.
- [8] M. B. Giles, “Multilevel monte carlo methods,” Acta Numerica, vol. 24, pp. 259–328, 2015.
- [9] C.-h. Rhee and P. W. Glynn, “Unbiased estimation with square root convergence for sde models,” Operations Research, vol. 63, no. 5, pp. 1026–1043, 2015.
- [10] V. Lemaire and G. Pagès, “Multilevel richardson–romberg extrapolation,” Bernoulli, vol. 23, no. 4A, pp. 2643–2692, 2017.
- [11] P. E. Kloeden and E. Platen, “Stochastic differential equations,” in Numerical Solution of Stochastic Differential Equations, pp. 103–160, Springer, 1992.
- [12] D. Giorgi, V. Lemaire, and G. Pagès, “Limit theorems for weighted and regular multilevel estimators,” Monte Carlo Methods and Applications, vol. 23, no. 1, pp. 43–70, 2017.
- [13] B. Arouna, “Adaptative monte carlo method, a variance reduction technique,” 2004.
- [14] P. Glasserman, Monte Carlo methods in financial engineering, vol. 53. Springer, 2004.
- [15] M. B. Alaya, K. Hajji, and A. Kebaier, “Importance sampling and statistical romberg method,” Bernoulli, vol. 21, no. 4, pp. 1947–1983, 2015.
- [16] H.-F. Chen, L. Guo, and A.-J. Gao, “Convergence and robustness of the robbins-monro algorithm truncated at randomly varying bounds,” Stochastic Processes and their Applications, vol. 27, pp. 217–231, 1987.
- [17] H.F.Chen and Y. Zhu, “Stochastic approximation procedures with randomly varying truncations,” Science in China, Ser. A, 1986.
- [18] C. Andrieu, É. Moulines, and P. Priouret, “Stability of stochastic approximation under verifiable conditions,” SIAM Journal on control and optimization, vol. 44, no. 1, pp. 283–312, 2005.
- [19] J. Lelong, “Almost sure convergence of randomly truncated stochastic algorithms under verifiable conditions,” Statistics & Probability Letters, vol. 78, no. 16, pp. 2632–2636, 2008.
- [20] V. Lemaire and G. Pagès, “Unconstrained recursive importance sampling,” The Annals of Applied Probability, vol. 20, no. 3, pp. 1029–1067, 2010.
- [21] M. B. Alaya, K. Hajji, and A. Kebaier, “Improved adaptive multilevel monte carlo and applications to finance,” arXiv preprint arXiv:1603.02959, 2016.
- [22] A. Kebaier and J. Lelong, “Coupling importance sampling and multilevel monte carlo using sample average approximation,” Methodology and Computing in Applied Probability, vol. 20, no. 2, pp. 611–641, 2018.
- [23] M. B. Alaya and A. Kebaier, “Central limit theorem for the multilevel monte carlo euler method,” The Annals of Applied Probability, vol. 25, no. 1, pp. 211–234, 2015.
- [24] S. Laruelle, C.-A. Lehalle, et al., “Optimal posting price of limit orders: learning by trading,” Mathematics and Financial Economics, vol. 7, no. 3, pp. 359–403, 2013.
- [25] P. Hall and C. C. Heyde, Martingale limit theory and its application. Academic press, 2014.