This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

On Data-Driven Drawdown Control
with Restart Mechanism in Trading

Chung-Han Hsieh Department of Quantitative Finance,
National Tsing Hua University, Hsinchu 30044, Taiwan R.O.C.
(e-mail: [email protected]).
Abstract

This paper extends the existing drawdown modulation control policy to include a novel restart mechanism for trading. It is known that the drawdown modulation policy guarantees the maximum percentage drawdown no larger than a prespecified drawdown limit for all time with probability one. However, when the prespecified limit is approaching in practice, such a modulation policy becomes a stop-loss order, which may miss the profitable follow-up opportunities, if any. Motivated by this, we add a data-driven restart mechanism into the drawdown modulation trading system to auto-tune the performance. We find that with the restart mechanism, our policy may achieve a superior trading performance to that without the restart, even with a nonzero transaction costs setting. To support our findings, some empirical studies using equity ETF and cryptocurrency with historical price data are provided.

keywords:
Control applications, stochastic systems, algorithmic trading, financial engineering, drawdown control.
thanks: This paper was supported in part by Ministry of Science and Technology, R.O.C. Taiwan, under Grant: MOST 111–2221–E–007–124– .

1 Introduction

Starting from the pioneering work by Markowitz (1952, 1959), the portfolio optimization problem is often solved by the mean-variance approach. That is, the trader seeks an optimal trade-off between payoff and risk measured by the portfolio returns’ mean and variance. While the variance is widely used as a standard risk metric in finance, it is more to the dispersion risk, which treats both positive and negative deviation from the mean as equally risky; see Fabozzi et al. (2007) for a good introduction to this topic.

1.1 Downside Risks

To remedy the equal riskiness on the dispersion risks, many surrogate risk measures are proposed to pay attention to the downside risks. This includes Value at Risk (VaR); see Jorion (2000), Conditional Value at Risk (CVaR); see Rockafellar and Uryasev (2000), absolute drawdown; see Magdon-Ismail and Atiya (2004); Hayes (2006), conditional expected drawdown (CED), the tail mean of maximum drawdown distributions, see Goldberg and Mahmoud (2017), and the more general coherent risk that axiomatize the risk measures; see Luenberger (2013); Shapiro et al. (2021). See also Korn et al. (2022) for an empirical study of comparing various drawdown-based risk metrics. In this paper, we focus on a more practical drawdown measure, the maximum percentage drawdown, the maximum percentage drops in wealth over time, as the risk measure.

1.2 Control of Drawdown

Different types of drawdown and methodologies are studied extensively in the existing literature. For example, optimal drawdown control problem in a continuous-time setting are studied in Grossman and Zhou (1993); Cvitanic and Karatzas (1994); Chekhlov et al. (2005); Malekpour and Barmish (2012, 2013). See also Boyd et al. (2017) for a study on multiperiod portfolio optimization problems involving drawdown as a constraint in a discrete-time setting. A recent study that uses deep reinforcement learning to address practical drawdown issues can be found in Wu et al. (2022).

Among all the existing papers, the prior works in Hsieh and Barmish (2017a, b) are the most closely related papers, where the key result, the so-called drawdown modulation lemma, is proved. Roughly speaking, it shows a necessary and sufficient conditions for a broad class of control policies that guarantees an almost sure maximum percentage drawdown protection for all time. However, Hsieh and Barmish (2017a) indicate that, in practice, when the prespecified drawdown limit is approaching, the trading policy may behave like a stop-loss order, and the trading may be stopped; see also Hsieh (2022a) for a study on a class of affine policies with a stop-loss order. To this end, we extend the drawdown modulation policy with a novel restart mechanism to remedy the stop-loss phenomenon.

1.3 Contributions of the Paper

This paper extends the existing drawdown modulation policy with a novel restart mechanism. The preliminaries are provided in Section 2. We formulate an optimal drawdown control problem for a two-asset portfolio with one risky and one riskless asset. Then we extend the existing drawdown modulation theory so that the riskless asset is explicitly involved; see Lemma 3.1. The necessary and sufficient conditions for a broad class of control policy which we call the drawdown modulation policy are provided. Then, the modulation policy with a restart mechanism is discussed in Section 3. The idea of the restart is simple: When the percentage drawdown up-to-date is close to the prespecified drawdown limit, we restart the trades with an updated policy. We also provide numerical examples involving ETF and cryptocurrency historical price data to support our findings; see Section 4.

2 Preliminaries

We now provide some useful preliminaries for the sections to follow.

2.1 Asset Trading Formulation

Fix an integer N>0.N>0. For stage k=0,1,,Nk=0,1,\dots,N, we let S(k)>0S(k)>0 denote the prices of the underlying financial asset at stage kk. The associated per-period returns are given by X(k):=S(k+1)S(k)S(k)X(k):=\frac{S(k+1)-S(k)}{S(k)} and the returns are assumed to be bounded, i.e., XminX(k)XmaxX_{\min}\leq X(k)\leq X_{\max} with XminX_{\min} and XmaxX_{\max} being points in the support, denoted by 𝒳\mathcal{X}, and satisfying 1<Xmin<0<Xmax.-1<X_{\min}<0<X_{\max}. For the money market asset, e.g., bond or bank account, we use rf(k)r_{f}(k) to denote the interest rate at stage kk.

Remark 2.1

Note that the returns considered here are not necessarily independent and can have an arbitrary but bounded distribution with the bounds XminX_{\min} and XmaxX_{\max}.

2.2 Account Value Dynamics

Beginning at some initial account value V(0)>0V(0)>0, consider a portfolio consisting of two assets, with one being risky and the other being a riskless asset with interest rate rf(k)[0,Xmax]r_{f}(k)\in[0,X_{\max}] for all kk almost surely. For stage k=0,1,k=0,1,\dots, we let V(k)V(k) denote the account value at stage kk. Then the evolution of the account value dynamics is described by the stochastic recursion

V(k+1)=V(k)+u(k)X(k)+(V(k)u(k))rf(k).\displaystyle V(k+1)=V(k)+u(k)X(k)+(V(k)-u(k))r_{f}(k).

Given a prespecified drawdown limit dmax(0,1)d_{\max}\in(0,1), we focus on conditions on selecting a policy u(k)u(k) under which satisfaction of the constraint d(k)dmaxd(k)\leq d_{\max} is assured for all kk with probability one where d(k)d(k) is the percentage drawdown up to date kk, which is defined below.

Definition 2.1 (Maximum Percentage Drawdown)

For k=0,1,,Nk=0,1,\dots,N, the percentage drawdown up to date kk, denoted by d(k)d(k), is defined as

d(k):=Vmax(k)V(k)Vmax(k)d(k):=\frac{V_{\max}(k)-V(k)}{V_{\max}(k)}

where Vmax(k):=max0ikV(i).V_{\max}(k):=\max_{0\leq i\leq k}V(i). The maximum percentage drawdown, call it dd^{*}, is then defined as

d:=max0kNd(k).d^{*}:=\max_{0\leq k\leq N}d(k).
Remark 2.2

It is readily seen that the percentage drawdown satisfies d(k)[0,1]d(k)\in[0,1] for all kk with probability one.

3 Drawdown Modulation with Restart

According to Hsieh and Barmish (2017a), it states a necessary and sufficient condition on any trading policy u(k)u(k) that guarantees the percentage drawdown up to date d(k)d(k) is no greater than a given level dmaxd_{\max} for all kk with probability one. Below, we extend the result to include a riskless asset.

Lemma 3.1 (Drawdown Modulation)

Let dmax(0,1)d_{\max}\in(0,1) be given. An trading policy u()u(\cdot) guarantees prespecified drawdown limit satisfying d(k)dmaxd(k)\leq d_{\max} for all kk with probability one if and only if for all kk, the condition

M(k)+rf(k)Xmaxrf(k)V(k)u(k)M(k)+rf(k)|Xmin|+rf(k)V(k)-\frac{M(k)+r_{f}(k)}{X_{\max}-r_{f}(k)}V(k)\leq u(k)\leq\frac{M(k)+r_{f}(k)}{|X_{\min}|+r_{f}(k)}V(k)

is satisfied along all sample paths where

M(k):=dmaxd(k)1d(k).M(k):=\frac{d_{\max}-d(k)}{1-d(k)}.

Proof. The idea of the proof is similar to that of Hsieh and Barmish (2017a). However, for the sake of completeness, we provide full proof here. To prove necessity, assuming that d(k)dmaxd(k)\leq d_{\max} for all kk and all sequences of returns, we must show the condition on u(k)u(k) holds along all sequences of returns. Fix kk. Since both d(k)dmaxd(k)\leq d_{\max} and d(k+1)dmaxd(k+1)\leq d_{\max} for all sequences of returns, we claim this forces the required inequalities on u(k)u(k). Without loss of generality, we prove the right-hand inequality for the case u(k)0u(k)\geq 0 and note that an almost identical proof also works for u(k)<0u(k)<0. To establish the condition on u(k)u(k) for all sequences of returns, it suffices to consider the path with the worst loss |Xmin|u(k)|X_{\min}|u(k). In this case, we have Vmax(k+1)=Vmax(k)V_{\max}(k+1)=V_{\max}(k). Hence,

d(k+1)\displaystyle d(k+1)
=Vmax(k+1)V(k+1)Vmax(k+1)\displaystyle=\frac{V_{\max}(k+1)-V(k+1)}{V_{\max}(k+1)}
=Vmax(k)V(k+1)Vmax(k)\displaystyle=\frac{V_{\max}(k)-V(k+1)}{V_{\max}(k)}
=Vmax(k)V(k)+u(k)|Xmin|(V(k)u(k))rf(k)Vmax(k)\displaystyle=\frac{V_{\max}(k)-V(k)+u(k)|X_{\min}|-(V(k)-u(k))r_{f}(k)}{V_{\max}(k)}
=d(k)+u(k)|Xmin|(V(k)u(k))rf(k)Vmax(k)dmax\displaystyle=d(k)+\frac{u(k)|X_{\min}|-(V(k)-u(k))r_{f}(k)}{V_{\max}(k)}\leq d_{\max}

We now substitute Vmax(k)=V(k)1d(k)>0{V_{\max}}(k)=\frac{{V(k)}}{{1-d(k)}}>0 into the inequality above and obtain

|Xmin|u(k)(V(k)u(k))rf(k)M(k)V(k),|X_{\min}|u(k)-(V(k)-u(k))r_{f}(k)\leq M(k)V(k),

where M(k)=dmaxd(k)1d(k)M(k)=\frac{d_{\max}-d(k)}{1-d(k)}. This implies that

(|Xmin|+rf(k))u(k)(M(k)+rf(k))V(k).(|X_{\min}|+r_{f}(k))u(k)\leq\left(M(k)+r_{f}(k)\right)V(k).

Or equivalently,

u(k)M(k)+rf(k)|Xmin|+rf(k)V(k).u(k)\leq\frac{M(k)+r_{f}(k)}{|X_{\min}|+r_{f}(k)}V(k).

To prove sufficiency, assuming that the stated bounds on u(k)u(k) hold along all sequences of returns, we must show d(k)dmaxd(k)\leq d_{\max} for all kk and all sequences of returns. Proceeding by induction, for k=0k=0, we trivially have d(0)=0dmaxd(0)=0\leq d_{\max}. To complete the inductive argument, we assume that d(k)dmaxd(k)\leq d_{\max} for all sequences of returns, and must show d(k+1)dmaxd(k+1)\leq d_{\max} for all sequences of returns. Without loss of generality, we again provide a proof for the case u(k)0u(k)\geq 0 and note that a nearly identical proof is used for u(k)<0u(k)<0. Indeed, by noting that

d(k+1)\displaystyle d(k+1) =1V(k+1)Vmax(k+1),\displaystyle=1-\frac{{V(k+1)}}{{{V_{\max}}(k+1)}},

and Vmax(k)Vmax(k+1)V_{\max}(k)\leq V_{\max}(k+1) for all sequences of returns, we split the argument into two cases: If Vmax(k)<Vmax(k+1)V_{\max}(k)<V_{\max}(k+1), then Vmax(k+1)=V(k+1)V_{\max}(k+1)=V(k+1) and we have d(k+1)=0dmax.d(k+1)=0\leq d_{\max}. On the other hand, if Vmax(k)=Vmax(k+1)V_{\max}(k)=V_{\max}(k+1), with the aid of the dynamics of account value, we have

d(k+1)\displaystyle d(k+1) =1V(k)+u(k)X(k)+(V(k)u(k))rf(k)Vmax(k)\displaystyle=1-\frac{V(k)+u(k)X(k)+(V(k)-u(k))r_{f}(k)}{V_{\max}(k)}
1+V(k)(1+rf(k))+u(k)(|Xmin|+rf(k))Vmax(k).\displaystyle\leq 1+\frac{-V(k)(1+r_{f}(k))+u(k)(|X_{\min}|+r_{f}(k))}{V_{\max}(k)}.

Using the upper bound on u(k)u(k); i.e.,

u(k)M(k)+rf(k)|Xmin|+rf(k)V(k)u(k)\leq\frac{M(k)+r_{f}(k)}{|X_{\min}|+r_{f}(k)}V(k)

and Vmax(k)=V(k)1d(k),{V_{\max}}(k)=\frac{{V(k)}}{{1-d(k)}}, we obtain

d(k+1)\displaystyle d(k+1) 1+(1+M(k))(1d(k))\displaystyle\leq 1+(-1+M(k))(1-d(k))
=d(k)+M(k)(1d(k)).\displaystyle=d(k)+M(k)(1-d(k)).

Using the definition of M(k)=dmaxd(k)1d(k)M(k)=\frac{d_{\max}-d(k)}{1-d(k)}, it follows that d(k+1)dmax,d(k+1)\leq d_{\max}, and the proof is complete. ∎

3.1 Drawdown Modulation Policy

Consistent with Hsieh and Barmish (2017a), fix the prespecified drawdown limit dmax(0,1)d_{\max}\in(0,1). With the aid of Lemma 3.1, one can readily obtain a class of policy functions u(k)u(k) expressed as a linear time-varying feedback controller parameterized by a gain γ\gamma, leading to the satisfaction of the drawdown specification. Specifically, we express u(k)u(k) in the feedback form

u(k):=K(k)V(k)\displaystyle u(k):=K(k)V(k) (1)

with K(k):=γM(k)K(k):=\gamma M(k) and

γΓ:=[1Xmaxmaxkrf(k),1|Xmin|+maxkrf(k)].\gamma\in\Gamma:=\left[\frac{-1}{X_{\max}-\max_{k}r_{f}(k)},\,\frac{1}{|X_{\min}|+\max_{k}r_{f}(k)}\right].

Equation (1) is called the drawdown modulation policy, which is parameterized by the two parameters (γ,dmax)(\gamma,d_{\max}).

Remark 3.1

(i)(i) It is readily verified that the drawdown modulation policy (1) satisfies Lemma 3.1. (ii)(ii) To link back to finance concepts, the special case of buy-and-hold is obtained when K(k)1K(k)\equiv 1. Note that u(k)<0u(k)<0 stands for short selling.111Short selling a stock means that a trader borrows the stocks from someone who owns it and selling it with the hope that the prices of the stock will drop in the near future; see Luenberger (2013). (iii)(iii) Instead of using a fixed feasible set Γ\Gamma, it is also possible to allow a time-varying feasible set, say Γk\Gamma_{k}, to reflect the time dependency of the returns.

Corollary 3.1 (Maximum Drawdown Protection)

With the drawdown modulation policy (1), the maximum percentage drawdown satisfies ddmaxd^{*}\leq d_{\max}.

Proof. Since the drawdown modulation policy satisfies Lemma 3.1, it assures d(k)dmaxd(k)\leq d_{\max} for all kk with probability one. Therefore, it follows that

d=max0kNd(k)dmax.d^{*}=\max_{0\leq k\leq N}d(k)\leq d_{\max}.\qed

3.2 Optimal Drawdown Control Problem

Having obtained the drawdown modulation policy (1), a natural question arises of how to select an “optimal” γ.\gamma. To this end, we define the total return up to terminal stage NN as a ratio

Rγ(N):=V(N)V(0)R_{\gamma}(N):=\frac{V(N)}{V(0)}

where the subscript of Rγ()R_{\gamma}(\cdot) is used to emphasize the dependence on the gain γ\gamma. Define J(γ):=𝔼[Rγ(N)]J(\gamma):=\mathbb{E}[R_{\gamma}(N)]. Then, we consider a multiperiod drawdown-based stochastic optimization problem

J:=maxγΓJ(γ)\displaystyle J^{*}:=\max_{\gamma\in\Gamma}J(\gamma)

subject to

V(k+1)\displaystyle V(k+1) =V(k)+u(k)X(k)+(V(k)u(k))rf(k)\displaystyle=V(k)+u(k)X(k)+(V(k)-u(k))r_{f}(k)
=[1+rf(k)+γM(k)(X(k)rf(k))]V(k).\displaystyle=[1+r_{f}(k)+\gamma M(k)(X(k)-r_{f}(k))]V(k).

It is readily verified that

V(N)V(0)=k=0N1[1+rf(k)+γM(k)(X(k)rf(k))].\frac{V(N)}{V(0)}=\prod_{k=0}^{N-1}[1+r_{f}(k)+\gamma M(k)(X(k)-r_{f}(k))].

Therefore, we rewrite the problem as the following equivalent form

maxγΓ𝔼[Rγ(N)]\displaystyle\max_{\gamma\in\Gamma}\mathbb{E}[R_{\gamma}(N)]
=maxγΓ𝔼[k=0N1[1+rf(k)+γM(k)(X(k)rf(k))]].\displaystyle=\max_{\gamma\in\Gamma}\mathbb{E}\left[\prod_{k=0}^{N-1}[1+r_{f}(k)+\gamma M(k)(X(k)-r_{f}(k))]\right]. (2)

In the sequel, we shall use γ\gamma^{*} to denote a maximizer of the optimization problem above. In practice, if one view that the optimum γ\gamma^{*} obtained may be too aggressive, a practical but arguably suboptimal way is to introduce an additional fraction, call it α\alpha, that is used to shrink the investment size; see Maclean et al. (2010) for a similar idea. That is, instead of working with γ\gamma^{*}, one may work with αγ\alpha\gamma^{*} where α(0,1]\alpha\in(0,1].

Remark 3.2 (Non-Convexity)

It is important to note that solving Problem (3.2) is challenging since the modulation function M(k)M(k) depends on γ\gamma and the history of X(0),,X(k1)X(0),\dots,X(k-1), which in general yields a nonconvex problem; e.g., see Figure 4 in Section 4 for an illustration of non-convexity nature. Therefore, Monte-Carlo simulations are often needed to obtain the optimum.

3.3 Modulation with Restart

As mentioned in Section 1, while the derived drawdown modulation policy provides almost sure drawdown protection, it may incur a stop-loss behavior. To remedy this, we now introduce a restart mechanism into the modulation policy. Specifically, let ε(0,dmax)\varepsilon\in(0,d_{\max}) be a prespecified threshold parameter. Then we set the threshold for restarting the trade by

d(k)+ε>dmax\displaystyle d(k)+\varepsilon>d_{\max} (3)

If, at some stage k=k0k=k_{0}, Inequality (3) is satisfied, the trading is restarted by re-initializing d(k0):=0d(k_{0}):=0 and reset the time-varying gain function K(k)K(k) of the modulation policy u(k)=K(k)V(k)u(k)=K(k)V(k) at that stage k=k0k=k_{0} as

K(k0):=γαek0/NM(k0)\displaystyle K(k_{0}):=\gamma^{*}\alpha e^{-k_{0}/N}M(k_{0}) (4)

where αek0/N\alpha e^{-k_{0}/N} represents a forgetting factor with fraction α(0,1]\alpha\in(0,1] mentioned previously. Then we continue the trade until the next restart stage or to the terminal stage NN.

Remark 3.3

(i)(i) The forgetting factor in Equation (4) reflects the idea that the trading size should be shrunk after the restart. Said another way, if the trades are approaching the prespecified drawdown limit dmaxd_{\max}, the follow-up trades should be more conservative after the restart. (ii)(ii) Note that after the restart, the control policy satisfies |u(k0)||γ|αdmax|u(k_{0})|\leq|\gamma^{*}|\alpha d_{\max} since M(k)dmaxM(k)\leq d_{\max} for all kk with probability one. (iii)(iii) While it does not consider in this paper, we should note that the optimal γ\gamma^{*} in Equation (4) can also be re-calculated at each restart time by using the previous k0mk_{0}-m stages information for some integer m>1m>1; see also Wang and Hsieh (2022) for a similar idea for obtaining a data-driven log-optimal portfolio via a sliding-window approach.

4 Illustrative Examples

In this section, two trading examples are proposed to support our theory. The first example is trading with ETF and riskless asset. The second example is trading with Bitcoin and a riskless asset. For the sake of simplicity, we take a constant daily interest rate rf(k):=0.01/365r_{f}(k):=0.01/365 for all kk, which corresponds to a 1%1\% annual rate. While our theory allows leveraging, in the sequel, we impose an additional cash-financing condition by imposing that |u(k)|V(k)|u(k)|\leq V(k) for all kk, which corresponds to |K(k)|1.|K^{*}(k)|\leq 1.

4.1 Trading with ETF and Riskless Asset

Consider the Vanguard Total World Stock Index Fund ETF (Ticker: VT)222VT invests in both foreign and U.S. stocks; hence it can be viewed as a good representative of the global stock market to be the risky asset covering a one-year in-sample period for optimization from January 02, 2019 to January 02, 2020, and the out-of-sample period from January 02, 2020 to September 20, 2022, which contains a total of N=684N=684 trading days; see Figure 2 where the cyan colored trajectory is used for in-sample optimization and the blue colored trajectory is used for the out-of-sample trading test. It should be noted that due to the COVID-19 pandemic, the considered prices covering the 2020 stock market crash from February 20, 2020 to April 7, 2020, and a recovery period after the crash. Thus, we view this dataset as an excellent back-testing case for our proposed drawdown modulation policy with the restart.

Refer to caption
Figure 1: Stock Prices of VT

Without loss of generality, consider the initial account value to be V(0):=$1V(0):=\$1. To implement the drawdown modulation policy with restart, we set dmax:=0.1d_{\max}:=0.1 and restart threshold ε:=dmax/10\varepsilon:=d_{\max}/10. With the data collected in the in-sample period, the corresponding feasible set is Γ[30.79, 34.9]\Gamma\approx[-30.79,\;34.9]. Then we numerically solve the optimization problem (3.2) via Monte-Carlo simulations. It follows that any γ(8.5,34.9)\gamma\in(8.5,34.9) share an almost identical optimal value when the cash-financing condition is imposed. For the sake of risk-averseness, we pick the infimum of the candidates, i.e., γ:=8.5\gamma^{*}:=8.5. Using this γ\gamma, we obtain the drawdown modulation policy u(k)=γM(k)V(k).u^{*}(k)=\gamma^{*}M(k)V(k).

The corresponding trading performance is shown in Figure 2, where the green dots indicate that the trade is restarted. In the same figure, we also compare it with the standard buy-and-hold strategy333Here, we mean buy and hold on the risky asset VT with K(k)=1K(k)=1. and the modulation policy without restart. We see clearly that the modulation policy with restart leads to superior performance. Some key performance metrics, including maximum drawdown and cumulative rate returns, and NN-period Sharpe ratio444The per-period Sharpe ratio is SR:=μrfσSR:=\frac{\mu-r_{f}}{\sigma} where μ\mu is the per-period sample mean return, σ\sigma is the per-period sample standard deviation, and rfr_{f} is the per-period riskless return. are reported in Table 1.

Refer to caption
Figure 2: Drawdown Modulation with/without Restart (Green Dots Indicate a Restart)
Table 1: Performance Metrics for Trading VT and Riskless Asset
Trading Performance with dmax=0.1d_{\max}=0.1 and ε=dmax/10\varepsilon=d_{\max}/10
Trading Policy Buy and Hold Modulation without Restart Modulation with Restart
Maximum percentage drawdown dd^{*} 34.24%34.24\% 10.00%10.00\% 10.00%\textbf{10.00}\%
Cumulative rate of return V(N)V(0)V(0)\frac{V(N)-V(0)}{V(0)} 10.64%10.64\% 8.422%-8.422\% 20.97%\textbf{20.97}\%
NN-period Sharpe ratio NSR\sqrt{N}\cdot SR 0.40820.4082 1.3375-1.3375 1.0462

4.2 Trading with Cryptocurrency and Riskless Asset

As a second example, we consider a portfolio consisting of cryptocurrency BTC-USD and a riskless asset. The BTC-USD asset covers the same in-sample and out-of-sample periods described in Example 4.1. From January 02, 2020 to September 20, 2022, it has a total of N=993N=993 trading days555It is worth noting that the cryptocurrency is typically traded at 24 hours a day, seven days a week. Therefore, it has longer tradings days than that trades with VT in Section 4.1.; see Figure 3. The corresponding feasible set for γ\gamma is Γ=[5.761, 7.083]\Gamma=[-5.761,\;7.083].

Refer to caption
Figure 3: Prices of BTC-USD (In-Sample and Out-of-Sample)

Take dmax:=0.2d_{\max}:=0.2 and restart threshold ε:=dmax/10\varepsilon:=d_{\max}/10. By solving the optimal drawdown control problem (3.2), we obtain γ5.138\gamma^{*}\approx 5.138; see Figure 4 for J(γ)J(\gamma) versus γΓ\gamma\in\Gamma. Note that the J(γ)J(\gamma) is clearly not concave for γΓ\gamma\in\Gamma. To consider the volatile nature of cryptocurrency and unforeseen estimation error, we consider using a fractional γ\gamma^{*} by αγ\alpha\gamma^{*} with α=1/2\alpha=1/2. That is, u(k)=K(k)V(k)u^{*}(k)=K^{*}(k)V(k) with K(k)=γ2M(k)V(k)K^{*}(k)=\frac{\gamma^{*}}{2}M(k)V(k). The trading performances using drawdown modulation policy with and without restart, and buy-and-hold strategy are shown in Figure 5, where the green dots indicate that the trades were restarted. Some performance metrics are summarized in Table 2 where we see that the modulation with restart provides the highest Sharpe ratio among all the other strategies.

Refer to caption
Figure 4: Seeking the Optimum: J(γ)J(\gamma) Versus γΓ\gamma\in\Gamma
Refer to caption
Figure 5: Trading Performance of BTC-USD with dmax:=0.2d_{\max}:=0.2 and ε:=dmax/10\varepsilon:=d_{\max}/10.
Table 2: Performance Metrics for Trading BTC-USD and Riskless Asset
Trading Performance with dmax=0.2d_{\max}=0.2 and ε=dmax/10\varepsilon=d_{\max}/10
Trading Policy Buy and Hold Modulation without Restart Modulation with Restart
Maximum percentage drawdown dd^{*} 72.12%72.12\% 20.00%20.00\% 20.18%\textbf{20.18}\%
Cumulative rate of return V(N)V(0)V(0)\frac{V(N)-V(0)}{V(0)} 170.43%170.43\% 17.41%17.41\% 101.82%\textbf{101.82}\%
NN-period Sharpe ratio NSR\sqrt{N}\cdot SR 1.44241.4424 0.69700.6970 2.1363

4.2.1 Transaction Costs Considerations.

Traditionally, with cryptocurrency as the underlying asset, most exchanges typically charge transaction fees from 0%0\% to 1%1\%, depending on the trading volumes. While we are not developing the drawdown modulation theory involving the transaction costs, in this example, we perform backtests to see the effects with percentage costs of 0.1%0.1\% per trade.666 This cost level is typical in common cryptocurrency exchanges such as Binance; see Hsieh (2022b). Take dmax=0.2d_{\max}=0.2, ε=dmax/10\varepsilon=d_{\max}/10 and γ=2.69\gamma^{*}=2.69, the trading performances are shown in Figure 6. From the figure, we see that, even with transaction costs, the proposed modulation policy with a restart mechanism is still superior to that without a restart. The corresponding Sharpe ratio is about 1.5211.521 compared with that without restart, where the Sharpe ratio is down to 0.043-0.043.

Refer to caption
Figure 6: Trading Performance with and without Transaction Costs

5 Concluding Remarks

This paper extended the existing drawdown modulation policy with a novel data-driven restart mechanism. We first derived a drawdown modulation policy by proving a lemma for a two-asset portfolio involving one risky and riskless asset. Then we modified the derived modulation policy by adding the restart mechanism. Subsequently, we showed an auto-tuned trading performance using historical price data for an ETF and cryptocurrency within a duration where the prices are volatile. Overall, our modulation policy with restart dynamically controls the drawdown and maintains the profits level.

Some possible future research directions might include the multi-asset problem, which will be considered in the subsequent journal version of this paper. The other possibility is to consider the consumption-investment problem; i.e.,

supγ𝔼[k=0T1U1(c(k))+U2(V(N))]\sup_{\gamma}\mathbb{E}\left[\sum_{k=0}^{T-1}U_{1}(c(k))+U_{2}(V(N))\right]

where U1U_{1} and U2U_{2} are two utility functions with c(k)c(k) being the amount consumed at time kk. See Cvitanic and Zapatero (2004). Lastly, one might consider a more general portfolio optimization problem that seeks γ\gamma maximizing supγ𝔼[U(Vγ(N))]\sup_{\gamma}\mathbb{E}\left[U(V_{\gamma}(N))\right] where U:U:\mathbb{R}\to\mathbb{R} is a strictly increasing and concave utility function. One example is to consider a general hyperbolic absolute risk aversion (HARA) class of utilities; i.e.,

U(x)=1θθ(ax1θ+b)θU(x)=\frac{1-\theta}{\theta}\left(\frac{ax}{1-\theta}+b\right)^{\theta}

for a,b>0a,b>0 and θ<1\theta<1. The HARA utility functions include the quadratic, exponential, and the isoelastic utility function as special cases.

Acknowledgment

The author thanks Chia-Yin Lee for coding and running some preliminary numerical examples on the early draft of this paper.

References

  • Boyd et al. (2017) Boyd, S., Busseti, E., Diamond, S., Kahn, R.N., Koh, K., Nystrup, P., Speth, J., et al. (2017). Multi-Period Trading via Convex Optimization. Foundations and Trends® in Optimization, 3(1), 1–76.
  • Chekhlov et al. (2005) Chekhlov, A., Uryasev, S., and Zabarankin, M. (2005). Drawdown Measure in Portfolio Optimization. International Journal of Theoretical and Applied Finance, 8(1), 13–58.
  • Cvitanic and Karatzas (1994) Cvitanic, J. and Karatzas, I. (1994). On Portfolio Optimization Under “Drawdown” Constraints. IMA Volumes in Mathematics and its Applications, 65, 35–46.
  • Cvitanic and Zapatero (2004) Cvitanic, J. and Zapatero, F. (2004). Introduction to the Economics and Mathematics of Financial Markets. MIT press.
  • Fabozzi et al. (2007) Fabozzi, F.J., Focardi, S.M., Kolm, P.N., and Pachamanova, D.A. (2007). Robust Portfolio Optimization and Management. John Wiley & Sons.
  • Goldberg and Mahmoud (2017) Goldberg, L.R. and Mahmoud, O. (2017). Drawdown: From Practice to Theory and Back Again. Mathematics and Financial Economics, 11(3), 275–297.
  • Grossman and Zhou (1993) Grossman, S.J. and Zhou, Z. (1993). Optimal Investment Strategies for Controlling Drawdowns. Mathematical Finance, 3(3), 241–276.
  • Hayes (2006) Hayes, B.T. (2006). Maximum Drawdowns of Hedge Funds with Serial Correlation. The Journal of Alternative Investments, 8(4), 26–38.
  • Hsieh (2022a) Hsieh, C.H. (2022a). Generalization of Affine Feedback Stock Trading Results to Include Stop-Loss Orders. Automatica, 136, 110051.
  • Hsieh (2022b) Hsieh, C.H. (2022b). On Robustness of Double Linear Trading with Transaction Costs. To appear in the IEEE Control Systems Letters.
  • Hsieh and Barmish (2017a) Hsieh, C.H. and Barmish, B.R. (2017a). On Drawdown-Modulated Feedback Control in Stock Trading. IFAC-PapersOnLine, 50(1), 952–958.
  • Hsieh and Barmish (2017b) Hsieh, C.H. and Barmish, B.R. (2017b). On Inefficiency of Markowitz-Style Investment Strategies when Drawdown is Important. In Proceedings of the IEEE Conference on Decision and Control (CDC), 3075–3080.
  • Jorion (2000) Jorion, P. (2000). Value at Risk.
  • Korn et al. (2022) Korn, O., Möller, P.M., and Schwehm, C. (2022). Drawdown Measures: Are They All the Same? The Journal of Portfolio Management, 48(5), 104–120.
  • Luenberger (2013) Luenberger, D.G. (2013). Investment Science. Oxford University Press.
  • Maclean et al. (2010) Maclean, L.C., Thorp, E.O., and Ziemba, W.T. (2010). Long-Term Capital Growth: the Good and Bad Properties of the Kelly and Fractional Kelly Capital Growth Criteria. Quantitative Finance, 10(7), 681–687.
  • Magdon-Ismail and Atiya (2004) Magdon-Ismail, M. and Atiya, A.F. (2004). Maximum Drawdown. Risk Magazine, 17(10), 99–102.
  • Malekpour and Barmish (2012) Malekpour, S. and Barmish, B.R. (2012). How Useful Are Mean-Variance Considerations in Stock Trading via Feedback Control? In Proceedings of the IEEE Conference on Decision and Control (CDC), 2110–2115.
  • Malekpour and Barmish (2013) Malekpour, S. and Barmish, B.R. (2013). A Drawdown Formula for Stock Trading via Linear Feedback in a Market Governed by Brownian Motion. In Proceedings of the European Control Conference (ECC), 87–92.
  • Markowitz (1952) Markowitz, H.M. (1952). Portfolio Selection. The Journal of Finance, 7, 77–91.
  • Markowitz (1959) Markowitz, H.M. (1959). Portfolio Selection: Efficient Diversification of Investments, volume 16. John Wiley New York.
  • Rockafellar and Uryasev (2000) Rockafellar, R.T. and Uryasev, S. (2000). Optimization of Conditional Value-at-Risk. Journal of Risk, 2, 21–42.
  • Shapiro et al. (2021) Shapiro, A., Dentcheva, D., and Ruszczynski, A. (2021). Lectures on Stochastic Programming: Modeling and Theory. SIAM.
  • Wang and Hsieh (2022) Wang, P.T. and Hsieh, C.H. (2022). On Data-Driven Log-Optimal Portfolio: A Sliding Window Approach. To appear in the IFAC-PapersOnline.
  • Wu et al. (2022) Wu, J.M.T., Lin, S.H., Syu, J.H., and Wu, M.E. (2022). Embedded Draw-Down Constraint Reward Function for Deep Reinforcement Learning. Applied Soft Computing, 125, 109150.