This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Preference Robust Modified Optimized Certainty Equivalent

Qiong Wu111Department of Systems Engineering and Engineering Management, The Chinese University of Hong Kong, Hong Kong. Email: [email protected]   and   Huifu Xu222Department of Systems Engineering and Engineering Management, The Chinese University of Hong Kong, Hong Kong. Email: [email protected]
Abstract

Ben-Tal and Teboulle [6] introduce the concept of optimized certainty equivalent (OCE) of an uncertain outcome as the maximum present value of a combination of the cash to be taken out from the uncertain income at present and the expected utility value of the remaining uncertain income. In this paper, we consider two variations of the OCE. First, we introduce a modified OCE by maximizing the combination of the utility of the cash and the expected utility of the remaining uncertain income so that the combined quantity is in a unified utility value. Second, we consider a situation where the true utility function is unknown but it is possible to use partially available information to construct a set of plausible utility functions. To mitigate the risk arising from the ambiguity, we introduce a robust model where the modified OCE is based on the worst-case utility function from the ambiguity set. In the case when the ambiguity set of utility functions is constructed by a Kantorovich ball centered at a nominal utility function, we show how the modified OCE and the corresponding worst case utility function can be identified by solving two linear programs alternatively. We also show the robust modified OCE is statistically robust in a data-driven environment where the underlying data are potentially contaminated. Some preliminary numerical results are reported to demonstrate the performance of the modified OCE and the robust modified OCE model.

Keywords. Robust modified optimized certainty equivalent, ambiguity of utility function, Kantorovich ball, piecewise linear approximation, error bounds, statistical robustness

1 Introduction

Let (Ω,,)(\Omega,\mathcal{F},\mathbb{P}) be a probability space with σ\sigma algebra \mathcal{F} and probability measure \mathbb{P} and ξ:(Ω,,)IR\xi:(\Omega,\mathcal{F},\mathbb{P})\to{\rm I\!R} be a random variable representing future income of a decision maker (DM). The optimized certainty equivalent of ξ\xi is defined as

(OCE)Su(ξ):=supxIR{x+𝔼P[u(ξx)]},{\rm(OCE)}\quad\quad\displaystyle{S_{u}(\xi):=\sup_{x\in{\rm I\!R}}\;\;\{x+{\mathbb{E}}_{P}[u(\xi-x)]\}}, (1.1)

where u:IRIRu:{\rm I\!R}\to{\rm I\!R} is the decision maker’s utility function and P:=ξ1P:=\mathbb{P}\circ\xi^{-1} is the probability measure on IR{\rm I\!R} induced by ξ\xi. The concept is first introduced by Ben-Tal and Teboulle [6] and closely related to other notions of certainty equivalent and risk measures, see [7] for a comprehensive discussion. The economic interpretation of this notion is that the decision maker may need to consume part of ξ\xi at present, denoted by xx, the sure present value of ξ\xi under the consumption plan becomes x+𝔼P[u(ξx)]x+{\mathbb{E}}_{P}[u(\xi-x)], and the optimized certainty equivalent Su(ξ)S_{u}(\xi) gives rise to the optimal allocation of the consumption which maximizes the sure present value of ξ\xi. As a measure, it enjoys a number of nice properties including constancy (Su(C)=CS_{u}(C)=C for constant CC), risk aversion (Su(ξ)𝔼P[ξ]S_{u}(\xi)\leq{\mathbb{E}}_{P}[\xi]) and translation invariance (Su(ξ+C)=Su(ξ)+CS_{u}(\xi+C)=S_{u}(\xi)+C). In particular, if uu is a normalized exponential utility function, it coincides with the classical certainty equivalent u1(𝔼P[u(ξ)])u^{-1}({\mathbb{E}}_{P}[u(\xi)]) in the literature of economics. Moreover, if u(t)=1α(t)+u(t)=-\frac{1}{\alpha}(-t)_{+} where α(0,1)\alpha\in(0,1) and (t)+=max{t,0}(t)_{+}=\max\{t,0\} for tIRt\in{\rm I\!R}, then the OCE effectively recovers the conditional value-at-risk (CVaR):

Su(ξ)\displaystyle\displaystyle S_{u}(\xi) =\displaystyle= supxIR{x1α𝔼P[(ξ+x)+]}\displaystyle\sup_{x\in{\rm I\!R}}\;\;\left\{x-\frac{1}{\alpha}{\mathbb{E}}_{P}[(-\xi+x)_{+}]\right\} (1.2)
=\displaystyle= infxIR{x+1α𝔼P[(ξx)+]}= CVaRα(ξ).\displaystyle-\inf_{x\in{\rm I\!R}}\;\;\left\{x+\frac{1}{\alpha}{\mathbb{E}}_{P}[(-\xi-x)_{+}]\right\}=-\text{ CVaR}_{\alpha}(-\xi).

The last equality is Rockafellar and Uryasev’s formulation of CVaR, see [7, 41]. Since CVaR is average of quantile, it is also known as average value-at-risk (AVaR), tail value-at-risk (TVaR) and expected shortfall, see [38, 37].

In this paper, we revisit the subject OCE from two perspectives. One is to consider a modified version of the optimized certainty equivalent

(MOCE)Mu(ξ):=supxIR{u(x)+𝔼P[u(ξx)]}.{\rm(MOCE)}\quad\quad\displaystyle{M_{u}(\xi):=\sup_{x\in{\rm I\!R}}\;\;\{u(x)+{\mathbb{E}}_{P}[u(\xi-x)]\}}. (1.3)

The modification is motivated to align the sure present value of ξ\xi to the expected utility theory [35] by considering the utility of present consumption u(x)u(x) instead of the monetary value xx. Recall that in Von Neumann-Morgenstern expected utility theory [35], the utility function is used to represent the decision maker’s preference relation over a prospect space including both random and deterministic prospects, and such representation is unique up to positive linear transformation. This means we can use both uu and 100u100u to represent the DM’s preference. However, the two utility functions would lead to completely different optimal values and optimal solutions in the OCE model. In contrast, the optimal solution is not affected in the MOCE model, and the optimal value is only affected by the same scale of the utility function. In our view, this kind of “invariance” of the optimal allocation xx^{*} and “scalability” w.r.t. the utility function is important because the optimal decision on the allocation/consumption xx should be determined by the DM’s risk preference irrespective of its equivalent representations.

The modified OCE model may be regarded as a special case of the well known consumption/investment models in economics [35, 19, 12] where u(x)u(x) is the utility of the current consumption/investment whereas 𝔼[u(ξx)]{\mathbb{E}}[u(\xi-x)] is the expected utility of the remaining asset to be consumed/invested in future. In these models, the utility functions for the current consumption and future consumption are identical. It is also possible to use different utility functions when the consumption at present is used for a new investment or production.

The other is to consider a situation where the decision maker’s utility function u()u(\cdot) is ambiguous, in other words, there is incomplete information to identify a utility function uu which captures the decision maker’s true utility preference. Consequently we propose to consider a robust optimized certainty equivalent measure

(RMOCE)R(ξ):=supxIRinfu𝒰{u(x)+𝔼P[u(ξx)]},{\rm(RMOCE)}\quad\quad\displaystyle{R(\xi):=\sup_{x\in{\rm I\!R}}\inf_{u\in{\cal U}}\;\;\{u(x)+{\mathbb{E}}_{P}[u(\xi-x)]\}}, (1.4)

where 𝒰{\cal U} is a set of plausible utility functions consistent with the observed utility preferences of the decision maker. The definition is in line with the philosophy of robust optimization where the optimized certainty equivalent value is based on the worst case utility function from set 𝒰{\cal U} to mitigate the risk arising from potential inaccurate use or misuse of the utility function. By convention, we call 𝒰{\cal U} the ambiguity set. In the case that 𝒰{\cal U} is a singleton, RMOCE reduces to MOCE. Note that RMOCE should be differentiated from the distributionally robust formulation of OCE by Wisemann et al. [50] where the focus is on the ambiguity of PP. In decision analysis, PP is known as a decision maker’s belief of the state of nature whereas uu characterizes the decision maker’s taste for risk/utility. The RMOCE model concerns the ambiguity of decision maker’s taste rather than belief.

Ambiguity of utility preference is a well discussed topic in behavioural economics. For instances, Thurstone [44] regards such ambiguity as a lack of accurate description of human behaviour. Karmarkar [27] and Weber [49] ascribe the ambiguity to cognitive difficulty and incomplete information. The ambiguity may also arise in the decision making problems which involve several stakeholders who fail to reach a consensus. Parametric and non-parametric approaches have subsequently been proposed to assess the true utility function, including discrete choice models (Train [45]), standard and paired gambling approaches for preference comparisons and certainty equivalence (Farquhar [15]), we refer readers to Hu et al. [23] for an excellent overview on this.

In decision making under uncertainty, a decision maker may choose the worst case utility function among a set of plausible utility functions representing his/her risk preference to mitigate the overall risk. This kind of research may be traced back to Maccheroni [33]. Cerreia-Vioglio et al. [10] seem to be the first to investigate ambiguity of decision maker’s utility function in the certainty equivalent model u1(𝔼[u(ξ)])u^{-1}({\mathbb{E}}[u(\xi)]) by considering the worst-case certainty equivalent from a given set of utility functions in their cautious expected utility model. They show that the DM’s risk preference can be represented by a worst-case certainty equivalent if and only if they are given by a binary relation satisfying the weak order, continuity, weak monotonicity and negative certainty independence (NCI) (NCI states that if a sure outcome is not enough to compensate the DM for a risky prospect, then its mixture with another lottery which reduces the certainty appeal, will not be more attractive than the same mixture of the risky prospect and the lottery).

Armbruster and Delage [3] give a comprehensive treatment of the topic from minimax preference robust optimization (PRO) perspective. Specifically, they propose to use available information of the decision maker’s utility preference such as preferring certain lotteries over other lotteries and being risk averse, SS-shaped or prudent to construct an ambiguity set of plausible utility functions and then base the optimal decision on the worst case utility function from the ambiguity set. Hu and Mehrotra [24] consider a probabilistic representation of the class of increasing concave utility functions by confining them to a compact interval and normalizing them with range [0,1][0,1]. In doing so, they propose a moment-type framework for constructing the ambiguity set of the decision maker’s utility preference which covers a number of important approaches such as the certainty equivalent and pairwise comparison. Hu and Stepanyan [25] propose a so-called reference-based almost stochastic dominance method for constructing a set of utility functions near a reference utility which satisfies certain stochastic dominance relationship and use the set to characterize the decision maker’s preference. Over the past few years, the research on PRO has received increasing attentions in the communities of stochastic/robust optimization and risk management, see for instances [22, 21, 14, 52, 31, 32] and references therein.

In both (MOCE) and (RMOCE) models, the true probability distribution PP is assumed to be known. In the data driven problems, the true PP is unknown but it is possible to use empirical data to construct an approximation of PP. Unfortunately, such data may be contaminated and consequently we may be concerned by the quality of the MOCE values calculated as such. This kind of issue is well studied in robust statistics [26] and can be traced down to earlier work of Hample [20]. Cont et al. [13] first study the quality of the plug-in estimators of law invariant risk measures using Hampel’s classical concept of qualitative robustness [20], that is, the plug-in estimator of a risk functional is said to be qualitatively robust if it is insensitive to the variation of sampling data. According to Hampel’s theorem, Cont et al. [13] demonstrate that the qualitative robustness of a plug-in estimator is equivalent to the weak continuity of the risk functional and that value at risk (VaR) is qualitatively robust whereas conditional value at risk (CVaR) is not. Krätschmer et al. [30] argue that the use of Hampel’s classical concept of qualitative robustness may be problematic because it requires the risk measure essentially to be insensitive with respect to the tail behaviour of the random variable and the recent financial crisis shows that a faulty estimate of tail behaviour can lead to a drastic underestimation of the risk. Consequently, they propose a refined notion of qualitative robustness that applies also to tail-dependent statistical functionals and that allows one to compare statistical functionals in regards to their degree of robustness. The new concept captures the trade-off between robustness and sensitivity and can be quantified by an index of qualitative robustness. Guo and Xu [17] take a step forward by deriving quantitative statistical robustness of PRO models. Xu and Zhang [51] extend the analysis to distributionally robust optimization models.

In this paper, we consider a situation where the decision maker has a nominal utility function but is short of complete information as to whether it is the true. Consequently we propose to use the Kantorovich ball centered at the nominal utility function as the ambiguity set. We begin with piecewise linear utility (PLU) functions defined over a convex and closed interval of IR{\rm I\!R} and show that the inner minimization problem in the definition of RMOCE can be reformulated as a linear program when ξ\xi has a finite discrete distribution. We then propose an iterative algorithm to compute the RMOCE by solving the inner minimization problem and outer maximization problem alternatively.

To extend the scope of the proposed computational method, we extend the discussion to the cases that the utility functions are not necessarily piecewise linear and the domain of the utility function is unbounded. We derive error bounds arising from using PLU-based RMOCE to approximate the general RMOCE. Since our numerical scheme for computing the RMOCE is based on the samples of ξ\xi, we study statistical robustness of the sample-based RMOCE to address the case that the sample data of ξ\xi are potentially contaminated. Finally we carry out some numerical tests on the proposed computational schemes for concave utility functions.

The rest of the paper are organized as follows. Section 2 discusses the basic properties of MOCE and RMOCE. Section 3 presents numerical schemes for computing the RMOCE when the utility functions in the ambiguity set are piecewise linear. Section 4 details approximation of the ambiguity set of general utility functions by the ambiguity set of piecewise linear utility functions and its effect on RMOCE. Section 5 discusses the RMOCE model with utility function having unbounded domain and streamlines the potential extensions of the MOCE model to multi-attribute decision making. Section 6 discusses statistical robustness of RMOCE when it is calculated with contaminated data. Section 7 reports numerical results and finally Section 8 concludes with a brief summary of the main contributions of the paper.

2 Properties of MOCE and RMOCE

We begin by discussing the well-definedness of MOCE and RMOCE. Let Lp(Ω,,)L_{p}(\Omega,{\cal F},\mathbb{P}) denote the space of random variables mapping from (Ω,,)(\Omega,{\cal F},\mathbb{P}) to IR{\rm I\!R} with finite pp-th order moments and ξLp(Ω,,)\xi\in L_{p}(\Omega,{\cal F},\mathbb{P}). Let 𝒰:IRIR\mathscr{U}:{\rm I\!R}\to{\rm I\!R} be the set of nondecreasing concave utility functions. Throughout this paper, we make a blanket assumption to ensure the well-definedness of the expected utility in the definitions of MOCE and RMOCE.

Assumption 2.1

There exist gauge functions ϕ1:IRIR\phi_{1}:{\rm I\!R}\to{\rm I\!R} and ϕ2:IRIR\phi_{2}:{\rm I\!R}\to{\rm I\!R} parameterized by xx satisfying 𝔼P[ϕi(ξ)]<{\mathbb{E}}_{P}[\phi_{i}(\xi)]<\infty for i=1,2i=1,2 such that

|u(ξx)|ϕ1(ξ)andsupu𝒰|u(ξx)|ϕ2(ξ),x,ξIR.|u(\xi-x)|\leq\phi_{1}(\xi)\quad\text{and}\quad\sup_{u\in{\mathscr{U}}}|u(\xi-x)|\leq\phi_{2}(\xi),\forall x,\xi\in{\rm I\!R}.

The condition stipulates the interaction between the tails distribution of ξ\xi and tails of the utility function. We refer readers to Guo and Xu [18] for more detailed discussions on this. To facilitate the forthcoming discussions, we let 𝒫(Ξ)\mathscr{P}(\Xi) denote the set of probability measures on ΞIR\Xi\subset{\rm I\!R}, and for each fixed xx, define

ϕi:={P𝒫(Ξ):𝔼P[ϕi(ξ)]<}{\cal M}^{\phi_{i}}:=\{P\in\mathscr{P}(\Xi):{\mathbb{E}}_{P}[\phi_{i}(\xi)]<\infty\}

for i=1,2i=1,2. Let 𝒞Ξϕi{\cal C}_{\Xi}^{\phi_{i}} denote the class of continuous functions h:ΞIRh:\Xi\to{\rm I\!R} such that |h(t)|C(ϕi(t)+1)|h(t)|\leq C(\phi_{i}(t)+1) for all tΞt\in\Xi. The ϕi\phi_{i}-topology, denoted by τϕi\tau_{\phi_{i}}, is the coarsest topology on ϕi{\cal M}^{\phi_{i}} for which the mapping gh:=Ξh(z)P(dz),h𝒞IRϕig_{h}:=\int_{\Xi}h(z)P(dz),\;h\in{\cal C}_{\rm I\!R}^{\phi_{i}} is continuous. A sequence {PN}ϕi\{P_{N}\}\subset{\cal M}^{\phi_{i}} is said to converge ϕi\phi_{i}-weakly to PϕiP\in{\cal M}^{\phi_{i}} written PNϕiP{P_{N}}\xrightarrow[]{\phi_{i}}P if it converges w.r.t. τϕi\tau_{\phi_{i}}. Note that in the case that when the support set of ξ\xi is a compact set in IR{\rm I\!R}, then the ϕi\phi_{i}-topology reduces to ordinary topology of weak convergence.

Our first technical result is on the attainability of the optimum in the definition of MOCE.

Proposition 2.1

Assume: (a) 2.1 holds, (b) there exists α\alpha such that {xIR:u(x)+𝔼P[u(ξx)]α}\{x\in{\rm I\!R}:u(x)+{\mathbb{E}}_{P}[u(\xi-x)]\geq\alpha\} is a compact set, (c) the support set of ξ\xi, denoted by Ξ=[ξmin,ξmax]\Xi=[\xi_{\min},\xi_{\max}], is bounded, (d) uu is strictly concave over Ξ\Xi. Then for Pϕ1P\in{\cal M}^{\phi_{1}},

Mu(ξ)=supx[ξmin/2,ξmax/2]{u(x)+𝔼P[u(ξx)]}.M_{u}(\xi)=\sup_{x\in[\xi_{\min}/2,\;\xi_{\max}/2]}\{u(x)+{\mathbb{E}}_{P}[u(\xi-x)]\}. (2.5)

Moreover, if {PN}𝒫(Ξ)\{P_{N}\}\subset\mathscr{P}(\Xi) and PNP_{N} converges weakly to δξ^\delta_{\hat{\xi}}, the Dirac probability measure at ξ^\hat{\xi}, then Mu(ξN)M_{u}(\xi_{N}) converges to 2u(ξ^/2)2u(\hat{\xi}/2).

Proof. Since uu is a strictly concave function, (1.3) is a convex optimization problem. Condition (b) ensures existence of an optimal solution, denoted by xx^{*}. Following a similar analysis to the proof of [7, Lemma 2.1], we can write down the first order optimality condition of the program at xx^{*},

0u(x)+𝔼P[u(ξx)],0\in\partial u(x^{*})+\partial{\mathbb{E}}_{P}[u(\xi-x^{*})], (2.6)

where u\partial u denotes convex subdifferential [40]. Since u(x)=[u+(x),u(x)]\partial u(x)=[u_{+}^{\prime}(x),u_{-}^{\prime}(x)] for any xIRx\in{\rm I\!R}, where u,u+u^{\prime}_{-},u^{\prime}_{+} denote the left derivative and right derivative of uu at xx and

𝔼P[u(ξx)]=𝔼P[u(ξx)],\partial{\mathbb{E}}_{P}[u(\xi-x^{*})]=-{\mathbb{E}}_{P}[\partial u(\xi-x^{*})],

where the expectation/integration at the right hand side is in the sense of Aumann [4]. Consequently we can rewrite (2.6) as

0\displaystyle 0 \displaystyle\in [u+(x),u(x)]𝔼P[[u+(ξx),u(ξx)]]\displaystyle[u_{+}^{\prime}(x^{*}),u_{-}^{\prime}(x^{*})]-{\mathbb{E}}_{P}\left[[u_{+}^{\prime}(\xi-x^{*}),u_{-}^{\prime}(\xi-x^{*})]\right] (2.7)
=\displaystyle= [u+(x),u(x)][𝔼P[u+(ξx)],𝔼P[u(ξx)]],\displaystyle[u_{+}^{\prime}(x^{*}),u_{-}^{\prime}(x^{*})]-\left[{\mathbb{E}}_{P}[u_{+}^{\prime}(\xi-x^{*})],{\mathbb{E}}_{P}[u_{-}^{\prime}(\xi-x^{*})]\right],

which yields

u+(x)𝔼P[u(ξx)]0u(x)𝔼P[u+(ξx)].u^{\prime}_{+}(x^{*})-{\mathbb{E}}_{P}[u^{\prime}_{-}(\xi-x^{*})]\leq 0\leq u^{\prime}_{-}(x^{*})-{\mathbb{E}}_{P}[u^{\prime}_{+}(\xi-x^{*})].

Since uu^{\prime}_{-} and u+u^{\prime}_{+} are non-increasing, the inequality above implies

u+(x)𝔼P[u(ξx)]u(ξminx)u^{\prime}_{+}(x^{*})\leq{\mathbb{E}}_{P}[u^{\prime}_{-}(\xi-x^{*})]\leq u^{\prime}_{-}(\xi_{\min}-x^{*}) (2.8)

and

u(x)𝔼P[u+(ξx)]u+(ξmaxx).u^{\prime}_{-}(x^{*})\geq{\mathbb{E}}_{P}[u^{\prime}_{+}(\xi-x^{*})]\geq u^{\prime}_{+}(\xi_{\max}-x^{*}). (2.9)

Moreover, since u(t)>u+(t′′)u_{-}^{\prime}(t^{\prime})>u_{+}^{\prime}(t^{\prime\prime}) for any t<t′′t^{\prime}<t^{\prime\prime}, then inequalities (2.8)-(2.9) imply

xξminxandxξmaxx,x^{*}\geq\xi_{\min}-x^{*}\quad{\rm and}\quad x^{*}\leq\xi_{\max}-x^{*},

and hence (2.5).

The second part of the claim follows directly from the first part in that the interval [ξminN/2,ξmaxN/2][\xi_{\min}^{N}/2,\\ \xi_{\max}^{N}/2] converges to a single point ξ^/2\hat{\xi}/2 and τϕ1\tau_{\phi_{1}}-convergence coincides with the weak convergence because of the restriction of the range of ξ\xi to a compact subset of IR{\rm I\!R}.  

Ben-Tal and Teboulle [7] derive a similar result to the first part of the proposition for the optimized certainty equivalent and demonstrate that Su(ξ)[ξmin,ξmax]S_{u}(\xi)\in[\xi_{\min},\xi_{\max}] under the conditions that uu is concave and 1u(0)1\in\partial u(0) rather than strictly concave. Strict concavity is needed to ensure the optimum in (2.5) to be achieved in [ξmin/2,ξmax/2][\xi_{\min}/2,\xi_{\max}/2]. We can find a counter example otherwise, see 2.3.

Like the optimized certainty equivalent, the newly defined modified optimized certainty equivalent enjoys a number of properties as stated in the next proposition.

Proposition 2.2 (Properties of MOCE)

Let u:IR(,+)u:{\rm I\!R}\rightarrow(-\infty,+\infty) be a closed proper function. Under 2.1, the following assertions hold.

  • (i)

    MuM_{u} is law invariant.

  • (ii)

    (Monotonicity) For any ξ1ξ2Lp(Ω,,)\xi_{1}\leq\xi_{2}\in L_{p}(\Omega,{\cal F},\mathbb{P}), with respective distributions (push-forward probabilities) P1,P2ϕ1P_{1},P_{2}\in{\cal M}^{\phi_{1}}, Mu(ξ1)Mu(ξ2)M_{u}(\xi_{1})\leq M_{u}(\xi_{2}).

  • (iii)

    (Risk aversion) If u(t)tu(t)\leq t for all tIRt\in{\rm I\!R}, then Mu(ξ)𝔼P[ξ]M_{u}(\xi)\leq{\mathbb{E}}_{P}[\xi] for any random variable ξ\xi.

  • (iv)

    (Second-order stochastic dominance) Let ξ1,ξ2\xi_{1},\xi_{2} be random variables with compact support. Then for any concave utility function uu,

    Mu(ξ1)Mu(ξ2)Cu(ξ1)Cu(ξ2),M_{u}(\xi_{1})\geq M_{u}(\xi_{2})\Longleftrightarrow C_{u}(\xi_{1})\geq C_{u}(\xi_{2}),

    where Cu(ξ):=u1(𝔼P[u(ξ)])C_{u}(\xi):=u^{-1}({\mathbb{E}}_{P}[u(\xi)]) is the classical certainty equivalent.

  • (v)

    (Concavity and positive subhomogeneity) If uu is concave, then Mu()M_{u}(\cdot) is also concave. Moreover, if u(0)0u(0)\geq 0, then

    Mu(δξ)δMu(ξ),δ[1,)andMu(δξ)δMu(ξ),δ[0,1].M_{u}(\delta\xi)\leq\delta M_{u}(\xi),\;\forall\delta\in[1,\infty)\quad{\text{and}}\quad M_{u}(\delta\xi)\geq\delta M_{u}(\xi),\;\forall\delta\in[0,1]. (2.10)

Proof. Parts (i)-(iii) follow straightforwardly from the definitions, we prove the rest.

Part (iv). “\Longleftarrow”. By the definition of certainty equivalent, Cu(ξ1)Cu(ξ2)C_{u}(\xi_{1})\geq C_{u}(\xi_{2}) implies 𝔼P[u(ξ1)]𝔼P[u(ξ2)]{\mathbb{E}}_{P}[u(\xi_{1})]\geq{\mathbb{E}}_{P}[u(\xi_{2})] for all concave utility functions. The latter implies ξ1\xi_{1} dominates ξ2\xi_{2} in second order, which in turn guarantees ξ1x\xi_{1}-x dominates ξ2x\xi_{2}-x in second order for any fixed xIRx\in{\rm I\!R}. Consequently 𝔼P[u(ξ1x)]𝔼P[u(ξ2x)]{\mathbb{E}}_{P}[u(\xi_{1}-x)]\geq{\mathbb{E}}_{P}[u(\xi_{2}-x)] for any xIRx\in{\rm I\!R}. Adding both sides of the inequality by u(x)u(x) and taking the maximum, we obtain Mu(ξ1)Mu(ξ2)M_{u}(\xi_{1})\geq M_{u}(\xi_{2}).

\Longrightarrow”. Let x1,x2x_{1},x_{2} be the points where the supremum of Mu(ξ1)M_{u}(\xi_{1}) and Mu(ξ2)M_{u}(\xi_{2}) are attained. Then

Mu(ξ1)=u(x1)+𝔼P[u(ξ1x1)]\displaystyle M_{u}(\xi_{1})=u(x_{1})+{\mathbb{E}}_{P}[u(\xi_{1}-x_{1})] \displaystyle\geq Mu(ξ2)=u(x2)+𝔼P[u(ξ2x2)]\displaystyle M_{u}(\xi_{2})=u(x_{2})+{\mathbb{E}}_{P}[u(\xi_{2}-x_{2})]
\displaystyle\geq u(x1)+𝔼P[u(ξ2x1)],\displaystyle u(x_{1})+{\mathbb{E}}_{P}[u(\xi_{2}-x_{1})],

which yields 𝔼P[u(ξ1x1)]𝔼P[u(ξ2x1)]{\mathbb{E}}_{P}[u(\xi_{1}-x_{1})]\geq{\mathbb{E}}_{P}[u(\xi_{2}-x_{1})]. The latter implies 𝔼P[u(ξ1)]𝔼P[u(ξ2)]{\mathbb{E}}_{P}[u(\xi_{1})]\geq{\mathbb{E}}_{P}[u(\xi_{2})].

Part (v). First we prove the concavity of MuM_{u}, i.e. for λ(0,1)\lambda\in(0,1) and any random variables ξ1\xi_{1}, ξ2\xi_{2},

Mu(λξ1+(1λ)ξ2))λMu(ξ1)+(1λ)Mu(ξ2).M_{u}(\lambda\xi_{1}+(1-\lambda)\xi_{2}))\geq\lambda M_{u}(\xi_{1})+(1-\lambda)M_{u}(\xi_{2}).

Since uu is concave, the function f(z,x):=u(x)+u(zx)f(z,x):=u(x)+u(z-x) is joint concave over IR×IR{\rm I\!R}\times{\rm I\!R}. Therefore, for any x1,x2IRx_{1},x_{2}\in{\rm I\!R}, with xλ:=λx1+(1λ)x2x_{\lambda}:=\lambda x_{1}+(1-\lambda)x_{2} and ξλ:=λξ1+(1λ)ξ2\xi_{\lambda}:=\lambda\xi_{1}+(1-\lambda)\xi_{2}, one has

𝔼[f(ξλ,xλ)]λ𝔼[f(ξ1,x1)]+(1λ)𝔼P[f(ξ2,x2)].{\mathbb{E}}[f(\xi_{\lambda},x_{\lambda})]\geq\lambda{\mathbb{E}}[f(\xi_{1},x_{1})]+(1-\lambda){\mathbb{E}}_{P}[f(\xi_{2},x_{2})].

Since Mu(ξλ)=Mu(λξ1+(1λ)ξ2))=supxIR𝔼P[f(ξλ,x)]M_{u}(\xi_{\lambda})=M_{u}(\lambda\xi_{1}+(1-\lambda)\xi_{2}))=\sup_{x\in{\rm I\!R}}{\mathbb{E}}_{P}[f(\xi_{\lambda},x)], it follows that

Mu(ξλ)\displaystyle M_{u}(\xi_{\lambda}) \displaystyle\geq supx1,x2{λ𝔼P[f(ξ1,x1)]+(1λ)𝔼P[f(ξ2,x2)]}=λMu(ξ1)+(1λ)Mu(ξ2).\displaystyle\sup_{x_{1},x_{2}}\left\{\lambda{\mathbb{E}}_{P}[f(\xi_{1},x_{1})]+(1-\lambda){\mathbb{E}}_{P}[f(\xi_{2},x_{2})]\right\}=\lambda M_{u}(\xi_{1})+(1-\lambda)M_{u}(\xi_{2}).

Next, we turn to prove the subhomogeneity of MuM_{u}. Let s(δ):=1δMu(δξ)s(\delta):=\frac{1}{\delta}M_{u}(\delta\xi), for δ>0\delta>0. Then

s(δ)=supxIR{1δu(δx)+𝔼P[1δu(δ(ξx))]}.s(\delta)=\sup_{x\in{\rm I\!R}}\left\{\frac{1}{\delta}u(\delta x)+{\mathbb{E}}_{P}\left[\frac{1}{\delta}u(\delta(\xi-x))\right]\right\}. (2.11)

Let δ2>δ1>0\delta_{2}>\delta_{1}>0. By the concavity of uu,

u(δ2t)u(0)δ20u(δ1t)u(0)δ10.\frac{u(\delta_{2}t)-u(0)}{\delta_{2}-0}\leq\frac{u(\delta_{1}t)-u(0)}{\delta_{1}-0}.

Since u(0)0u(0)\geq 0, the above inequality implies

1δ2u(δ2t)1δ1u(δ1t),tIR.\frac{1}{\delta_{2}}u(\delta_{2}t)\leq\frac{1}{\delta_{1}}u(\delta_{1}t),\;\forall t\in{\rm I\!R}. (2.12)

Inequality (2.12) also implies

𝔼P[1δ2u(δ2(ξx))]𝔼P[1δ1u(δ1(ξx))].{\mathbb{E}}_{P}\left[\frac{1}{\delta_{2}}u(\delta_{2}(\xi-x))\right]\leq{\mathbb{E}}_{P}\left[\frac{1}{\delta_{1}}u(\delta_{1}(\xi-x))\right].

A combination of the two inequalities implies the objective function in (2.11) is non-increasing in δ\delta and hence s(δ)s(\delta). By setting δ1\delta_{1} and δ2\delta_{2} to 11 respectively in the inequality above, we obtain (2.10).  

Next, we discuss how the utility function uu may be recovered from a given modified certainty equivalent Mu(ξ)M_{u}(\xi), which is an important property enjoyed by the OCE. Let

ξp={zwith probabilityp,0with probability 1p,\xi_{p}=\left\{\begin{array}[]{ll}z&\text{with probability}\;p,\\ 0&\text{with probability}\;1-p,\end{array}\right.

where 0<p<10<p<1 and z>0z>0. For a concave utility function uu, the modified optimized certainty equivalent Mu(ξp)M_{u}(\xi_{p}) can be written as

Mu[z,p]:=sup0xz/2{u(x)+pu(zx)+(1p)u(x)}.M_{u}[z,p]:=\sup_{0\leq x\leq z/2}\{u(x)+pu(z-x)+(1-p)u(-x)\}. (2.13)
Proposition 2.3

If uu is a strong risk averse utility function, i.e., u(t)<tu(t)<t for all t0t\neq 0, and u(0)=0u(0)=0, then limp0+Mu[z,p]p=u(z).\lim_{p\rightarrow 0^{+}}\frac{M_{u}[z,p]}{p}=u(z).

Proof. Observe that x=0x^{*}=0 is the optimal solution of problem (2.13) if and only if

u(x)+pu(zx)+(1p)u(x)\displaystyle u(x)+pu(z-x)+(1-p)u(-x) \displaystyle\leq u(0)+pu(z0)+(1p)u(0)\displaystyle u(0)+pu(z-0)+(1-p)u(-0) (2.14)
=\displaystyle= pu(z),x[0,z/2].\displaystyle pu(z),\forall x\in[0,z/2].

The inequality above can be equivalently written as

p[u(zx)u(x)u(z)]u(x)u(x),x[0,z/2].p[u(z-x)-u(-x)-u(z)]\leq-u(-x)-u(x),\forall x\in[0,z/2]. (2.15)

Since uu is strongly risk averse, that is, u(t)<tu(t)<t for all t0t\neq 0, then u(x)u(x)>0-u(-x)-u(x)>0 and hence inequality (2.15) holds for pp sufficiently small. This in turn shows that inequality (2.15) holds and hence x=0x^{*}=0 is the optimal solution of problem (2.13) for all pp sufficiently small. Thus we have Mu[z,p]=pu(z)M_{u}[z,p]=pu(z) and the conclusion follows.  

Example 2.1

We give a few examples which illustrate how MOCE can be calculated in a closed form and their difference in comparison with OCE. Let Mu(ξ)=2(1(𝔼P[eξ])1/2)M_{u}(\xi)=2(1-\left({\mathbb{E}}_{P}[e^{-\xi}]\right)^{1/2}). Then Mu[z,p]=22(pez+(1p))1/2M_{u}[z,p]=2-2(pe^{-z}+(1-p))^{1/2} and

u(z)=limp0+Mu[z,p]p=limp0+22(pez+1p)1/2p=limp0+1ez(pez+1p)1/2.\displaystyle u(z)=\lim_{p\rightarrow 0^{+}}\frac{M_{u}[z,p]}{p}=\lim_{p\rightarrow 0^{+}}\frac{2-2(pe^{-z}+1-p)^{1/2}}{p}=\lim_{p\rightarrow 0^{+}}\frac{1-e^{-z}}{(pe^{-z}+1-p)^{1/2}}.

Hence, the recovered utility function is u(z)=1ezu(z)=1-e^{-z}.

Example 2.2 (Exponential Utility Function)

Let u(t)=1etu(t)=1-e^{-t}, tIRt\in{\rm I\!R}. It is easy to derive that the optimal solution of problem (1.3) is x=12ln𝔼P[eξ]x^{*}=-\frac{1}{2}\ln{\mathbb{E}}_{P}\left[e^{-\xi}\right] and the modified optimized certainty equivalent is Mu(ξ)=2(1(𝔼P[eξ])1/2).M_{u}(\xi)=2\left(1-\left({\mathbb{E}}_{P}[e^{-\xi}]\right)^{1/2}\right). On the other hand, it follows from [7] that Su(ξ)=ln𝔼P[eξ].S_{u}(\xi)=-\ln{\mathbb{E}}_{P}[e^{-\xi}]. Since u(t)tu(t)\leq t for all tIRt\in{\rm I\!R}, then we can deduce from the definitions that Mu(ξ)Su(ξ)M_{u}(\xi)\leq S_{u}(\xi). Indeed the strict inequality holds in that u(t)=tu(t)=t only at t=0t=0.

Example 2.3 (Piecewise Linear Utility Function)

Let

u(t)={γ2tift0,γ1tift>0,u(t)=\left\{\begin{array}[]{ll}\gamma_{2}t&\text{if}\;\;t\leq 0,\\ \gamma_{1}t&\text{if}\;\;t>0,\end{array}\right.

where 0γ1<1γ20\leq\gamma_{1}<1\leq\gamma_{2}. Then the utility function uu can be written as u(t)=γ1(t)+γ2(t)+u(t)=\gamma_{1}(t)_{+}-\gamma_{2}(-t)_{+} and the modified optimized certainty equivalent is

Mu(ξ)=supxIR{γ1(x)+γ2(x)+γ2𝔼P[(xξ)+]+γ1𝔼P[(ξx)+]}.M_{u}(\xi)=\sup_{x\in{\rm I\!R}}\{\gamma_{1}(x)_{+}-\gamma_{2}(-x)_{+}-\gamma_{2}{\mathbb{E}}_{P}[(x-\xi)_{+}]+\gamma_{1}{\mathbb{E}}_{P}[(\xi-x)_{+}]\}. (2.16)

Compared to optimized certainty equivalent (see [7])

Su(ξ)=supxIR{xγ2𝔼P[(xξ)+]+γ1𝔼P[(ξx)+]},S_{u}(\xi)=\sup_{x\in{\rm I\!R}}\{x-\gamma_{2}{\mathbb{E}}_{P}[(x-\xi)_{+}]+\gamma_{1}{\mathbb{E}}_{P}[(\xi-x)_{+}]\},

we can also conclude that Mu(ξ)Su(ξ)M_{u}(\xi)\leq S_{u}(\xi) because u(t)tu(t)\leq t for all tIRt\in{\rm I\!R}.

It might be interesting to see where the optimum in (2.16) is achieved. We consider the case that PP follows a Dirac distribution at point t>0t>0, that is, [ξmin,ξmax]={t}[\xi_{\min},\xi_{\max}]=\{t\}. Consequently

u(x)+𝔼P[u(ξx)]={(γ2γ1)x+tγ1ifx0,tγ1if 0<xt,(γ1γ2)x+tγ2ifxt.\displaystyle u(x)+{\mathbb{E}}_{P}[u(\xi-x)]=\left\{\begin{array}[]{ll}(\gamma_{2}-\gamma_{1})x+t\gamma_{1}&\;{\rm if}\;x\leq 0,\\ t\gamma_{1}&\;{\rm if}\;0<x\leq t,\\ (\gamma_{1}-\gamma_{2})x+t\gamma_{2}&\;{\rm if}\;x\geq t.\end{array}\right. (2.20)

The set of optimal solutions is [0,t][0,t], which is not contained in [0,t/2][ξmin/2,ξmax/2]={t/2}[0,t/2]\not\subset[\xi_{\min}/2,\xi_{\max}/2]=\{t/2\}. This explains that (2.5) may fail to hold without strict concavity of uu.

We now move on to discuss the properties of the robust modified optimized certainty equivalent.

Proposition 2.4 (Properties of RMOCE)

Let u:IR[,+)u:{\rm I\!R}\rightarrow[-\infty,+\infty) be a closed proper function. Under 2.1, the following assertions hold.

  • (i)

    R(ξ)R(\xi) is law invariant.

  • (ii)

    (Monotonicity) For any ξ1ξ2Lp(Ω,,)\xi_{1}\leq\xi_{2}\in L_{p}(\Omega,{\cal F},\mathbb{P}), with respective distributions (push-forward probabilities) P1,P2ϕ2P_{1},P_{2}\in{\cal M}^{\phi_{2}}, R(ξ1)R(ξ2)R(\xi_{1})\leq R(\xi_{2}).

  • (iii)

    (Risk aversion) If u(t)tu(t)\leq t, for all tIRt\in{\rm I\!R} and u𝒰u\in\mathscr{U}, then R(ξ)𝔼P[ξ]R(\xi)\leq{\mathbb{E}}_{P}[\xi], for any random variable ξ\xi.

  • (iv)

    (Second-order stochastic dominance) Let ξ1,ξ2\xi_{1},\xi_{2} be random variables with compact support. Then for any concave utility function uu,

    Cu(ξ1)Cu(ξ2)R(ξ1)R(ξ2),C_{u}(\xi_{1})\geq C_{u}(\xi_{2})\Longrightarrow R(\xi_{1})\geq R(\xi_{2}),

    where Cu(ξ)C_{u}(\xi) is the classical certainty equivalent.

  • (v)

    (Concavity and positive subhomogeneity) If uu is concave, then R()R(\cdot) is also concave. Moreover, if u(0)0u(0)\geq 0, then

    R(δξ)δR(ξ),δ[1,)andR(δξ)δR(ξ),δ[0,1].R(\delta\xi)\leq\delta R(\xi),\;\forall\delta\in[1,\infty)\quad{\text{and}}\quad R(\delta\xi)\geq\delta R(\xi),\;\forall\delta\in[0,1]. (2.21)

Proof. Parts (i)-(iii) are obvious.

Part (iv). Following a similar argument to the proof of part (iv) of 2.2, we can show that Cu(ξ1)Cu(ξ2)C_{u}(\xi_{1})\geq C_{u}(\xi_{2}) implies 𝔼P[u(ξ1)]𝔼P[u(ξ2)]{\mathbb{E}}_{P}[u(\xi_{1})]\geq{\mathbb{E}}_{P}[u(\xi_{2})] and 𝔼P[u(ξ1x)]𝔼P[u(ξ2x)]{\mathbb{E}}_{P}[u(\xi_{1}-x)]\geq{\mathbb{E}}_{P}[u(\xi_{2}-x)] for any fixed xIRx\in{\rm I\!R} and hence

u(x)+𝔼P[u(ξ1x)]u(x)+𝔼P[u(ξ2x)].u(x)+{\mathbb{E}}_{P}[u(\xi_{1}-x)]\geq u(x)+{\mathbb{E}}_{P}[u(\xi_{2}-x)].

Taking infimum on both sides w.r.t. uu over 𝒰{\cal U} and then supremum w.r.t. xx, we obtain R(ξ1)R(ξ2)R(\xi_{1})\geq R(\xi_{2}).

Part (v). Let gu(δ,x):=1δu(δx)+𝔼P[1δu(δ(ξx))].g_{u}(\delta,x):=\frac{1}{\delta}u(\delta x)+{\mathbb{E}}_{P}\left[\frac{1}{\delta}u(\delta(\xi-x))\right]. We can show as in the proof of 2.2 (v) that gu(,x)g_{u}(\cdot,x) is non-increasing over IR{\rm I\!R}. This property is preserved after taking the infimum in uu over 𝒰{\mathcal{U}} and then supremum in xx over IR{\rm I\!R}.  

Before concluding this section, we remark that it is possible to use a different utility function vv for the present consumption xx, i.e.,

(RMOCE)Mu,v(ξ):=supxIR{v(x)+𝔼P[u(ξx)]}.{\rm(RMOCE^{\prime})}\quad\quad\displaystyle{M_{u,v}(\xi):=\sup_{x\in{\rm I\!R}}\;\{v(x)+{\mathbb{E}}_{P}[u(\xi-x)]\}}. (2.22)

In that case, some of the properties of MOCE may be retained. For example, law invariance, monotonicity, risk aversion, concavity, positive subhomogeneity and second-order stochastic dominance are all satisfied when vv enjoys the same property as uu. 2.3 also holds when vv satisfies the same property as uu. However, the change will have an effect on 2.1, in which case it will be difficult to estimate the interval containing the optimal solution.

3 Computation of RMOCE

Having investigated the properties of MOCE and RMOCE in the previous section, we move on to discuss numerical schemes for computing RMOCE in this section. To this end, we need to have a concrete structure of the ambiguity set. As reviewed in the introduction, various approaches have been proposed for constructing an ambiguity set of utility functions in the literature of preference robust optimization depending on the availability of information. Here we consider a situation where the decision maker has a nominal utility function obtained from empirical data or subjective judgement but lacks of complete information to identify whether it is the true utility function which captures precisely the decision maker’s preference. Consequently we may construct a ball of utility functions centered at the nominal utility function under some appropriate metrics. Here we concentrate on the Kantorovich metric.

3.1 Kantorovich ball of piecewise linear utility functions

We begin by considering a ball of utility function centered at a piecewise linear utility function under the the Kantorovich metric. In practice, decision maker’s utility preferences are often elicited through questionnaires. For example, a customer’s utility preference may be elicited via the customer’s willingness to pay at certain price points [43, 32]. From computational point of view, piecewise linear utility function may bring significant convenience to calculation of OCE, see Nouiehed et al. [36].

Let t1<<tNt_{1}<\cdots<t_{N} be an ordered sequence of points in [a,b][a,b] and T:={t1,,tN}T:=\{t_{1},\cdots,t_{N}\} with t1=at_{1}=a and tN=bt_{N}=b. Let 𝒰N\mathscr{U}_{N} be a class of continuous, non-decreasing, concave, piecewise linear functions defined over an interval [a,b][a,b] with kinks on TT, as well as Lipschitz condition with modulus LL and normalized conditions u(a)=0u(a)=0 and u(b)=1u(b)=1. Let uN,uN0𝒰Nu_{N},u_{N}^{0}\in\mathscr{U}_{N}, we consider a ball in 𝒰N\mathscr{U}_{N} with the Kantorovich metric

𝔹K(uN0,r)={uN𝒰N|𝖽𝗅K(uN,uN0)r},\mathbb{B}_{K}(u_{N}^{0},r)=\left\{u_{N}\in\mathscr{U}_{N}|\mathsf{d\kern-0.70007ptl}_{K}(u_{N},u_{N}^{0})\leq r\right\}, (3.23)

where the subscript KK represents the Kantorovich metric and

𝖽𝗅K(u,v):=supg𝒢K|g,ug,v|=supg𝒢K{IRg(t)𝑑u(t)IRg(t)𝑑v(t)}\displaystyle{\mathsf{d\kern-0.70007ptl}_{K}(u,v):=\sup_{g\in\mathscr{G}_{K}}|\langle g,u\rangle-\langle g,v\rangle|=\sup_{g\in\mathscr{G}_{K}}\left\{\int_{{\rm I\!R}}g(t)du(t)-\int_{{\rm I\!R}}g(t)dv(t)\right\}} (3.24)

and

𝒢K:={g:IRIR|g(t)g(t)||tt|,t,tIR}.\mathscr{G}_{K}:=\{g:{\rm I\!R}\rightarrow{\rm I\!R}\mid|g(t)-g(t^{\prime})|\leq|t-t^{\prime}|,\forall t,t^{\prime}\in{\rm I\!R}\}. (3.25)

Note that piecewise linear utility functions are used to approximate general utility functions in the utility preference robust optimization model [18]. The difference is that here we use the Kantorovich ball to construct the ambiguity set of DM’s utility function whereas the authors use pairwise comparison approach to elicit the DM’s utility preferences in [18]. The next proposition states that 𝖽𝗅K(uN,uN0)\mathsf{d\kern-0.70007ptl}_{K}(u_{N},u_{N}^{0}) may be computed by solving a linear program.

Proposition 3.1

The Kantorvich distance 𝖽𝗅K(uN,uN0)\mathsf{d\kern-0.70007ptl}_{K}(u_{N},u_{N}^{0}) is the optimal value of the following linear program:

maxy1,,yN1z1,,zN\displaystyle\displaystyle\max_{\begin{subarray}{c}y_{1},\cdots,y_{N-1}\\ z_{1},\cdots,z_{N}\end{subarray}} j=2N(βj1βj10)yj1\displaystyle\sum_{j=2}^{N}(\beta_{j-1}-\beta_{j-1}^{0})y_{j-1} (3.26a)
s.t.\displaystyle{\rm s.t.}~{}~{}~{} yj1zj1(tjtj1)+12(tjtj1)2,j=2,,N,\displaystyle y_{j-1}\leq z_{j-1}(t_{j}-t_{j-1})+\frac{1}{2}(t_{j}-t_{j-1})^{2},j=2,\cdots,N, (3.26e)
yj1zj1(tjtj1)+12(tjtj1)2,j=2,,N,\displaystyle-y_{j-1}\leq-z_{j-1}(t_{j}-t_{j-1})+\frac{1}{2}(t_{j}-t_{j-1})^{2},j=2,\cdots,N,
yj1zj(tjtj1)+12(tjtj1)2,j=2,,N,\displaystyle y_{j-1}\leq z_{j}(t_{j}-t_{j-1})+\frac{1}{2}(t_{j}-t_{j-1})^{2},j=2,\cdots,N,
yj1zj(tjtj1)+12(tjtj1)2,j=2,,N.\displaystyle-y_{j-1}\leq-z_{j}(t_{j}-t_{j-1})+\frac{1}{2}(t_{j}-t_{j-1})^{2},j=2,\cdots,N.

Proof. Let g𝒢Kg\in\mathscr{G}_{K}. By definition,

abg(t)𝑑uN(t)=j=2Nβj1tj1tjg(t)𝑑t,\int_{a}^{b}g(t)du_{N}(t)=\sum_{j=2}^{N}\beta_{j-1}\int_{t_{j-1}}^{t_{j}}g(t)dt,

where βj\beta_{j} denotes the slope of uNu_{N} at interval [tj1,tj][t_{j-1},t_{j}]. Since for each g𝒢Kg\in\mathscr{G}_{K}, g𝒢K-g\in\mathscr{G}_{K},

𝖽𝗅K(uN,uN0)=supg𝒢Kj=2N(βj1βj10)tj1tjg(t)𝑑t,\mathsf{d\kern-0.70007ptl}_{K}(u_{N},u_{N}^{0})=\sup_{g\in\mathscr{G}_{K}}\sum_{j=2}^{N}(\beta_{j-1}-\beta_{j-1}^{0})\int_{t_{j-1}}^{t_{j}}g(t)dt,

where βj10\beta_{j-1}^{0} denotes the slope of u0u^{0} at interval [tj1,tj][t_{j-1},t_{j}]. Note that in this formulation, 𝖽𝗅K(uN,uN0)\mathsf{d\kern-0.70007ptl}_{K}(u_{N},u^{0}_{N}) depends on the slopes of uN,uN0u_{N},u_{N}^{0} rather than their function values. Let yj1:=tj1tjg(t)𝑑ty_{j-1}:=\int_{t_{j-1}}^{t_{j}}g(t)dt and zj:=g(tj)z_{j}:=g(t_{j}). Since |g(t)g(tj1)|ttj1|g(t)-g(t_{j-1})|\leq t-t_{j-1} for all t[tj1,tj]t\in[t_{j-1},t_{j}], we have

zj1(tjtj1)12(tjtj1)2yj1zj1(tjtj1)+12(tjtj1)2z_{j-1}(t_{j}-t_{j-1})-\frac{1}{2}(t_{j}-t_{j-1})^{2}\leq y_{j-1}\leq z_{j-1}(t_{j}-t_{j-1})+\frac{1}{2}(t_{j}-t_{j-1})^{2}

for j=2,,Nj=2,\cdots,N. Likewise, since |g(t)g(tj)|tjt|g(t)-g(t_{j})|\leq t_{j}-t for all t[tj1,tj]t\in[t_{j-1},t_{j}], we have

zj(tjtj1)12(tjtj1)2yj1zj(tjtj1)+12(tjtj1)2z_{j}(t_{j}-t_{j-1})-\frac{1}{2}(t_{j}-t_{j-1})^{2}\leq y_{j-1}\leq z_{j}(t_{j}-t_{j-1})+\frac{1}{2}(t_{j}-t_{j-1})^{2}

for j=2,,Nj=2,\cdots,N. To complete the proof, it suffices to show that conditions

|g(t)g(tj1)|ttj1and|g(t)g(tj)|tjt,t[tj1,tj]|g(t)-g(t_{j-1})|\leq t-t_{j-1}\;\;\text{and}\;\;|g(t)-g(t_{j})|\leq t_{j}-t,\forall t\in[t_{j-1},t_{j}] (3.27)

are adequate to cover the generic condition

|g(t)g(t′′)||tt′′|,t,t′′[a,b].|g(t^{\prime})-g(t^{\prime\prime})|\leq|t^{\prime}-t^{\prime\prime}|,\forall t^{\prime},t^{\prime\prime}\in[a,b]. (3.28)

We consider two cases.

Case 1. t,t′′[ti1,ti]t^{\prime},t^{\prime\prime}\in[t_{i-1},t_{i}] for some ii. In this case, the generic condition is adequately covered by |g(t)g(tj1)|ttj1|g(t)-g(t_{j-1})|\leq t-t_{j-1} for all t[tj1,tj]t\in[t_{j-1},t_{j}]. Because the objective depends only on tj1tjg(t)𝑑t\int_{t_{j-1}}^{t_{j}}g(t)dt.

Case 2. t,t′′t^{\prime},t^{\prime\prime} lie in two intervals, i.e., t[ti1,ti]t^{\prime}\in[t_{i-1},t_{i}] and t′′[tj1,tj]t^{\prime\prime}\in[t_{j-1},t_{j}], where i<ji<j. Then by (3.27),

|g(t)g(t′′)|\displaystyle|g(t^{\prime})-g(t^{\prime\prime})| \displaystyle\leq |g(t)g(ti)|+|g(ti)g(ti+1)|++|g(tj2)g(tj1)|+|g(tj1)g(t′′)|\displaystyle|g(t^{\prime})-g(t_{i})|+|g(t_{i})-g(t_{i+1})|+\cdots+|g(t_{j-2})-g(t_{j-1})|+|g(t_{j-1})-g(t^{\prime\prime})|
\displaystyle\leq tit+ti+1ti++tj1tj2+t′′tj1=t′′t.\displaystyle t_{i}-t^{\prime}+t_{i+1}-t_{i}+\cdots+t_{j-1}-t_{j-2}+t^{\prime\prime}-t_{j-1}=t^{\prime\prime}-t^{\prime}.

The proof is complete.  

3.2 Alternating iterative algorithm for computing RMOCE

We are now ready to discuss how to compute the RMOCE with the ambiguity set of piecewise linear utility functions constructed by the Kantorovich ball. Assume that the probability distribution of random variable ξ\xi is discrete with P(ξ=ξk)=pkP(\xi=\xi_{k})=p_{k} for k=1,,Kk=1,...,K and uN,uN0𝒰Nu_{N},u_{N}^{0}\in\mathscr{U}_{N}. Then we can rewrite the RMOCE problem (1.4) as

(RMOCEPLU)RN(ξ):=maxxIRminuN𝔹K(uN0,r)uN(x)+k=1KpkuN(ξkx).{\rm(RMOCE-PLU)}\quad\;\;R_{N}(\xi):=\displaystyle{\max_{x\in{\rm I\!R}}\min_{u_{N}\in\mathbb{B}_{K}(u^{0}_{N},r)}}\;\;u_{N}(x)+\sum_{k=1}^{K}p_{k}u_{N}(\xi_{k}-x). (3.29)

Recall that in 2.1, we show that the optimal solutions of MOCE are contained in interval [ξmin/2,ξmax/2][\xi_{\min}/2,\xi_{\max}/2] when utility function is strictly concave. Unfortunately, this result is not applicable to problem (3.34) because uNsu_{N}^{s} is piecewise linear. However, under some fairly moderate conditions, we are able to show that the optimal solutions are bounded. The next proposition states this.

Proposition 3.2

Consider MOCE problem (1.3). Let XX^{*} denote the set of optimal solutions. Assume: (a) uu is a piecewise linear concave function and (b) uu has at least two pieces in the interval [ξmin,ξmax][\xi_{\min},\xi_{\max}]. Then the following assertions hold.

  • (i)

    XX^{*} is a compact and convex set.

  • (ii)

    If 0[ξmin,ξmax]0\in[\xi_{\min},\xi_{\max}], then X[ξmin,ξmax]X^{*}\subset[\xi_{\min},\xi_{\max}].

  • (iii)

    If ξmin0\xi_{\min}\geq 0, then X[0,ξmax]X^{*}\subset[0,\xi_{\max}].

  • (iv)

    If ξmax0\xi_{\max}\leq 0, then X[ξmin,0]X^{*}\subset[\xi_{\min},0].

Proof. Part (i). Observe first that XX^{*} is a convex set since problem (1.3) is a convex optimization problem. Suppose for the sake of a contradiction that XX^{*} is unbounded. Then either XX^{*} is a right half line or a left half line. We consider the former. In that case, there exists xXx^{*}\in X^{*} sufficiently large such that

[ξmin,ξmax][ξminx,x].[\xi_{\min},\xi_{\max}]\subset[\xi_{\min}-x^{*},x^{*}]. (3.30)

By the first order optimality condition

0u(x)+𝔼P[u(ξx)]=u(x)𝔼P[u(ξx)].0\in\partial u(x^{*})+\partial{\mathbb{E}}_{P}[u(\xi-x^{*})]=\partial u(x^{*})-{\mathbb{E}}_{P}[\partial u(\xi-x^{*})]. (3.31)

The equality holds because of Clarke regularity, see [9, 11]. Since xξmaxxx^{*}\geq\xi_{\max}-x^{*}, then any subgradient in set u(x)\partial u(x^{*}) is greater or equal to the subgradient from u(ξx)\partial u(\xi-x^{*}) for all ξ[ξmin,ξmax]\xi\in[\xi_{\min},\xi_{\max}]. This means the optimality condition holds if and only if xx^{*} and ξminx\xi_{\min}-x^{*} are in the domain of the same linear piece. But this contradicts assumption (b). Using a similar argument, we can also show that XX^{*} cannot be a left half line.

Part (ii). Assume for a contradiction that x>ξmaxx^{*}>\xi_{\max}. Then inclusion (3.30) holds. Following a similar analysis to that in Part (i), we can show that in this case xx^{*} does not satisfy (3.31). If x<ξmin0x^{*}<\xi_{\min}\leq 0, then

[ξmin,ξmax][x,ξmaxx].[\xi_{\min},\xi_{\max}]\subset[x^{*},\xi_{\max}-x^{*}]. (3.32)

Consequently we can show that XX^{*} cannot satisfy the optimality condition (3.31).

Part (iii). In this case, we can show that xx^{*} cannot be larger that ξmax\xi_{\max} because otherwise we would have (3.30) and a contradiction to the optimality condition. Likewise if x<0x^{*}<0, then the inclusion (3.32) would be invoked.

Part (iv) is similar to Part (iii), we omit the details.  

Note that if we strengthen the condition on two linear pieces in the interval [ξmin,ξmax][\xi_{\min},\xi_{\max}] to a smaller interval [ξmin/2,ξmax/2][\xi_{\min}/2,\xi_{\max}/2], then we will be able to strengthen the conclusions in Parts (ii)-(iv) whereby XX^{*} is included in [ξmin/2,ξmax/2][\xi_{\min}/2,\xi_{\max}/2], we leave readers for an exercise.

Now we propose the alternating iterative algorithm for solving the maximin problem (3.29).

Algorithm 3.1

Step 0. Choose an initial point x0x^{0}.

Step 1. For s=1,…, solve

uNsargminuN𝔹K(uN0,r)uN(xs1)+k=1KpkuN(ξkxs1)\displaystyle{u^{s}_{N}\in\arg\min_{u_{N}\in\mathbb{B}_{K}(u^{0}_{N},r)}}u_{N}(x^{s-1})+\sum_{k=1}^{K}p_{k}u_{N}(\xi_{k}-x^{s-1}) (3.33)

and

xsargmaxxXuNs(x)+k=1KpkuNs(ξkx),\displaystyle{x^{s}\in\arg\max_{x\in X}u_{N}^{s}(x)+\sum_{k=1}^{K}p_{k}u_{N}^{s}(\xi_{k}-x),} (3.34)

where XX is a compact subset of IR{\rm I\!R}.

Step 2. Stop when xs+1=xsx^{s+1}=x^{s} and uNs+1=uNsu_{N}^{s+1}=u_{N}^{s}.

Note that in equation (3.34), we restrict xx to taking values in a convex and compact set XX since 3.2 guarantees that the optimal xx^{*} is contained in such a set. There is another important issue concerning the algorithm, that is, whether a sequence {xs}\{x^{s}\} generated by the algorithm converges to the optimal solution of (RMOCE-PLU). The next proposition addresses this.

Proposition 3.3

Algorithm 3.1 either terminates in a finite number of steps with a solution of the (RMOCE-PLU) model or generates a sequence {(xs,uNs)}\{(x^{s},u_{N}^{s})\} whose cluster points, if exist, are optimal solution of the (RMOCE-PLU) model.

Proof. Let (x,u)(x^{*},u^{*}) be a cluster point of the sequence generated by Algorithm 3.1. Then for all 𝔹K(uN0,r)\mathbb{B}_{K}(u^{0}_{N},r) and xXx\in X,

u(x)+𝔼P[u(ξx)]u(x)+𝔼P[u(ξx)]u(x)+𝔼P[u(ξx)].u^{*}(x)+{\mathbb{E}}_{P}[u^{*}(\xi-x)]\leq u^{*}(x^{*})+{\mathbb{E}}_{P}[u^{*}(\xi-x^{*})]\leq u(x^{*})+{\mathbb{E}}_{P}[u(\xi-x^{*})]. (3.35)

For s=1,2,s=1,2,...,

us+1(xs)+𝔼P[us+1(ξxs)]u(xs)+𝔼P[u(ξxs)]u^{s+1}(x^{s})+{\mathbb{E}}_{P}[u^{s+1}(\xi-x^{s})]\leq u(x^{s})+{\mathbb{E}}_{P}[u(\xi-x^{s})]

and

us(xs)+𝔼P[us(ξxs)]us(x)+𝔼P[us(ξx)].u^{s}(x^{s})+{\mathbb{E}}_{P}[u^{s}(\xi-x^{s})]\leq u^{s}(x)+{\mathbb{E}}_{P}[u^{s}(\xi-x)]. (3.36)

If Algorithm 3.1 terminates in finite steps, then xs+1=xsx^{s+1}=x^{s} and us+1=usu^{s+1}=u^{s} for some ss and (xs,us)(x^{s},u^{s}) satisfies (3.35). In what follows we consider the case that Algorithm 3.1 generates an infinite sequence {(xs,us)}\{(x^{s},u^{s})\}. Let (x^,u^)(\hat{x},\hat{u}) be a cluster point of {(xs,us)}\{(x^{s},u^{s})\}. For the simplicity of notation, we assume that (xs,us)(x^,u^)(x^{s},u^{s})\rightarrow(\hat{x},\hat{u}). If (x^,u^)(\hat{x},\hat{u}) is not a saddle point, then it violates one of the inequalities in (3.35). Without loss of generality, consider the case that the first inequality of (3.35) is violated, that is, there exists x0x_{0} such that

u^(x0)+𝔼P[u^(ξx0)]>u^(x^)+𝔼P[u^(ξx^)].\hat{u}(x_{0})+{\mathbb{E}}_{P}[\hat{u}(\xi-x_{0})]>\hat{u}(\hat{x})+{\mathbb{E}}_{P}[\hat{u}(\xi-\hat{x})].

Since u^\hat{u} is continuous, then for sufficiently large ss,

us(x0)+𝔼P[us(ξx0)]>us(xs)+𝔼P[us(ξxs)],u^{s}(x_{0})+{\mathbb{E}}_{P}[u^{s}(\xi-x_{0})]>u^{s}(x^{s})+{\mathbb{E}}_{P}[u^{s}(\xi-x^{s})],

which is a contradiction to (3.36). In the same manner, we can show that (x^,u^)(\hat{x},\hat{u}) satisfies the second inequality in (3.35). The proof is complete.  

Note that the cluster point is indeed a saddle point of the maximin problem (3.29) and existence of the latter is guaranteed by the fact that the objective function is linear in uu and concave in xx. Problem (3.33) is a convex problem because 𝔹K(uN0,r)\mathbb{B}_{K}(u^{0}_{N},r) is a compact and convex set. By writing each utility function uN𝒰Nu_{N}\in\mathscr{U}_{N} as

uN(t)=(a1t+b1)𝟙[t1,t2](t)+j=2N1(ajt+bj)𝟙(tj,tj+1](t)u_{N}(t)=(a_{1}t+b_{1})\mathds{1}_{[t_{1},t_{2}]}(t)+\sum_{j=2}^{N-1}(a_{j}t+b_{j})\mathds{1}_{(t_{j},t_{j+1}]}(t) (3.37)

for t[a,b]t\in[a,b] and writing down the Lagrange dual of problem (3.26),

minλji,i=1,2,3,4j=2,,N\displaystyle\displaystyle\min_{\begin{subarray}{c}\lambda^{i}_{j},i=1,2,3,4\\ j=2,\cdots,N\end{subarray}} 12j=2N(λj1+λj2+λj3+λj4)(tjtj1)2\displaystyle-\frac{1}{2}\sum_{j=2}^{N}(\lambda_{j}^{1}+\lambda_{j}^{2}+\lambda_{j}^{3}+\lambda_{j}^{4})(t_{j}-t_{j-1})^{2} (3.38a)
s.t.\displaystyle{\rm s.t.}~{}~{}~{}~{} (βj1βj10)+(λj1λj2+λj3λj4)=0,j=2,,N,\displaystyle(\beta_{j-1}-\beta_{j-1}^{0})+(\lambda_{j}^{1}-\lambda_{j}^{2}+\lambda_{j}^{3}-\lambda_{j}^{4})=0,j=2,\cdots,N, (3.38f)
(λj+12λj+11)(tj+1tj)+(λj4λj3)(tjtj1)=0,\displaystyle(\lambda_{j+1}^{2}-\lambda_{j+1}^{1})(t_{j+1}-t_{j})+(\lambda_{j}^{4}-\lambda_{j}^{3})(t_{j}-t_{j-1})=0,
j=2,,N1,\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad j=2,\cdots,N-1,
(λ22λ21)(t2t1)=0,\displaystyle(\lambda_{2}^{2}-\lambda_{2}^{1})(t_{2}-t_{1})=0,
(λN4λN3)(tNtN1)=0,\displaystyle(\lambda_{N}^{4}-\lambda_{N}^{3})(t_{N}-t_{N-1})=0,
λji0,j=2,,N,i=1,2,3,4.\displaystyle\lambda_{j}^{i}\leq 0,j=2,\cdots,N,i=1,2,3,4.

We can effectively reformulate problem (3.33) as a linear program:

(as,bs)argminaj,bj,j=1,,N1\displaystyle(a^{s},b^{s})\in\displaystyle\arg\min_{\begin{subarray}{c}a_{j},b_{j},\\ j=1,\cdots,N-1\end{subarray}} (a1xs1+b1)𝟙[t1,t2](xs1)+j=2N1(ajxs1+bj)𝟙(tj,tj+1](xs1)\displaystyle(a_{1}x^{s-1}+b_{1})\mathds{1}_{[t_{1},t_{2}]}(x^{s-1})+\sum_{j=2}^{N-1}(a_{j}x^{s-1}+b_{j})\mathds{1}_{(t_{j},t_{j+1}]}(x^{s-1})
+k=1Kpk{(a1(ξkxs1)+b1)𝟙[t1,t2](ξkxs1).\displaystyle+\sum_{k=1}^{K}p_{k}\Bigl{\{}(a_{1}(\xi^{k}-x^{s-1})+b_{1})\mathds{1}_{[t_{1},t_{2}]}(\xi^{k}-x^{s-1})\Bigr{.}
.+j=2N1(aj(ξkxs1)+bj)𝟙(tj,tj+1](ξkxs1)}\displaystyle\Bigl{.}+\sum_{j=2}^{N-1}(a_{j}(\xi^{k}-x^{s-1})+b_{j})\mathds{1}_{(t_{j},t_{j+1}]}(\xi^{k}-x^{s-1})\Bigr{\}}
s.t. aj1tj+bj1=ajtj+bj,j=2,,N1,\displaystyle a_{j-1}t_{j}+b_{j-1}=a_{j}t_{j}+b_{j},j=2,\cdots,N-1,
a1t1+b1=0,\displaystyle a_{1}t_{1}+b_{1}=0,
aN1tN+bN1=1,\displaystyle a_{N-1}t_{N}+b_{N-1}=1,
aj+1aj,j=1,,N2,\displaystyle a_{j+1}\leq a_{j},j=1,\cdots,N-2,
0ajL,j=1,,N1,\displaystyle 0\leq a_{j}\leq L,j=1,\cdots,N-1,
12j=2N(λj1+λj2+λj3+λj4)(tjtj1)2r,\displaystyle-\frac{1}{2}\sum_{j=2}^{N}(\lambda_{j}^{1}+\lambda_{j}^{2}+\lambda_{j}^{3}+\lambda_{j}^{4})(t_{j}-t_{j-1})^{2}\leq r,\qquad\qquad
(aj1aj10)+(λj1λj2+λj3λj4)=0,j=2,,N,\displaystyle(a_{j-1}-a_{j-1}^{0})+(\lambda_{j}^{1}-\lambda_{j}^{2}+\lambda_{j}^{3}-\lambda_{j}^{4})=0,j=2,\cdots,N,
(λj+12λj+11)(tj+1tj)+(λj4λj3)(tjtj1)=0,\displaystyle(\lambda_{j+1}^{2}-\lambda_{j+1}^{1})(t_{j+1}-t_{j})+(\lambda_{j}^{4}-\lambda_{j}^{3})(t_{j}-t_{j-1})=0,
j=2,,N1,\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad j=2,\cdots,N-1,
(λ22λ21)(t2t1)=0,\displaystyle(\lambda_{2}^{2}-\lambda_{2}^{1})(t_{2}-t_{1})=0,
(λN4λN3)(tNtN1)=0,\displaystyle(\lambda_{N}^{4}-\lambda_{N}^{3})(t_{N}-t_{N-1})=0,
λji0,j=2,,N,i=1,2,3,4,\displaystyle\lambda_{j}^{i}\leq 0,j=2,\cdots,N,i=1,2,3,4,

where aj10a_{j-1}^{0} denotes the slope of uN0u_{N}^{0} at interval [tj1,tj][t_{j-1},t_{j}]. Constraint (3.39) requires the piecewise linear function to be continuous at the kinks, constraint (3.39) and (3.39) represent the normalized conditions, (3.39) requires the concavity of utility function, (3.39) represents the Lipschitz condition, constraints (3.39)-(3.39) represent the bounded Kantorovich ball. Note that here we use the Lagrange dual problem (3.38) instead of the primal problem (3.26) because the latter would have bilinear terms (βj1βj10)yj1(\beta_{j-1}-\beta_{j-1}^{0})y_{j-1} otherwise.

4 RMOCE with non-piecewise linear utility functions

The computational schemes that we discussed in the previous section are applicable to the case when the ambiguity set is constructed by a Kantorovich ball of piecewise linear utility functions. In practice, the utility functions are not necessarily piecewise linear. This raises a question as to how much we may miss if we use (RMOCEPLU){\rm(RMOCE-PLU)} to compute (RMOCE) with the ambiguity set constructed by the Kantorovich ball of general utility functions. In this section, we address the issue which is essentially about error bound of modelling error. To maximize the scope of coverage, we consider ζ\zeta-ball instead of the Kantovich ball. Let 𝒰L\mathscr{U}_{L} be a class of continuous, non-decreasing, concave functions defined over [a,b][a,b] with Lipschitz condition with moludus LL and normalized conditions u(a)=0u(a)=0 and u(b)=1u(b)=1. For u0𝒰Lu^{0}\in\mathscr{U}_{L}, we define

𝔹(u0,r):={u𝒰L|𝖽𝗅𝒢(u,u0)r},\mathbb{B}(u^{0},r):=\left\{u\in\mathscr{U}_{L}\;|\;\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u,u^{0})\leq r\right\}, (4.40)

where

𝖽𝗅𝒢(u,v):=supg𝒢|g,ug,v|,\displaystyle{\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u,v):=\sup_{g\in\mathscr{G}}|\langle g,u\rangle-\langle g,v\rangle|}, (4.41)

𝒢\mathscr{G} is a set of measurable functions defined over IR{\rm I\!R} and g,u:=IRg(t)𝑑u(t)\langle g,u\rangle:=\int_{{\rm I\!R}}g(t)du(t). 𝖽𝗅𝒢\mathsf{d\kern-0.70007ptl}_{\mathscr{G}} is known as a pseudo metric. It can be observed that 𝖽𝗅𝒢(u,v)=0\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u,v)=0 if and only if g,u=g,v\langle g,u\rangle=\langle g,v\rangle for all g𝒢g\in\mathscr{G} but not necessarily u=vu=v unless 𝒢\mathscr{G} is sufficiently large. By specifying particular properties of functions in set 𝒢\mathscr{G}, we may obtain some specific metric such as Kantorovich metric 𝖽𝗅K\mathsf{d\kern-0.70007ptl}_{K} and the Kolmogorov metric with 𝒢=𝒢I\mathscr{G}=\displaystyle{\mathscr{G}_{I}}, where 𝒢I\mathscr{G}_{I} consists of all indicator functions defined as

𝟙(a,t](s):={1if s(a,t],0otherwise.\mathds{1}_{(a,t]}(s):=\begin{cases}1&\text{if }s\in(a,t],\\ 0&\text{otherwise}.\end{cases} (4.42)

With the definition of the ζ\zeta-ball and u0𝒰Lu^{0}\in\mathscr{U}_{L}, we may define the corresponding RMOCE as

(RMOCE(ζ))R(ξ):=maxxIRminu𝔹(u0,r)u(x)+𝔼P[u(ξx)]{\rm(RMOCE(\zeta))}\quad\;\;R(\xi):=\displaystyle{\max_{x\in{\rm I\!R}}\min_{u\in\mathbb{B}(u^{0},r)}}\;\;u(x)+{\mathbb{E}}_{P}[u(\xi-x)] (4.43)

and the one when the utility functions are restricted to be piecewise linear:

(RMOCE(ζ,N))RN(ξ):=maxxIRminuN𝔹N(uN0,r)uN(x)+𝔼P[uN(ξx)],{\rm(RMOCE(\zeta,N))}\quad\;\;R_{N}(\xi):=\displaystyle{\max_{x\in{\rm I\!R}}\min_{u_{N}\in\mathbb{B}_{N}(u_{N}^{0},r)}}\;\;u_{N}(x)+{\mathbb{E}}_{P}[u_{N}(\xi-x)], (4.44)

where

𝔹N(uN0,r):={uN𝒰N:𝖽𝗅𝒢(uN,uN0)r}.\mathbb{B}_{N}(u^{0}_{N},r):=\left\{u_{N}\in\mathscr{U}_{N}:\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u_{N},u^{0}_{N})\leq r\right\}. (4.45)

We investigate the difference between 𝔹N(uN0,r)\mathbb{B}_{N}(u_{N}^{0},r) and 𝔹(u0,r)\mathbb{B}(u^{0},r) and its propagation to the optimal values. Let 𝒰1\mathcal{U}_{1} and 𝒰2\mathcal{U}_{2} be two sets of utility function, 𝖽𝗅𝒢(u,𝒰1):=infu~𝒰1𝖽𝗅𝒢(u,u~)\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u,\mathcal{U}_{1}):=\inf_{\tilde{u}\in\mathcal{U}_{1}}\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u,\tilde{u}) between uu and 𝒰1\mathcal{U}_{1}, 𝔻𝒢(𝒰1,𝒰2):=supu𝒰1𝖽𝗅𝒢(u,𝒰2)\mathbb{D}_{\mathscr{G}}(\mathcal{U}_{1},\mathcal{U}_{2}):=\sup_{u\in\mathcal{U}_{1}}\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u,\mathcal{U}_{2}) be the deviation distance of 𝒰1\mathcal{U}_{1} from 𝒰2\mathcal{U}_{2}, and 𝒢(𝒰1,𝒰2):=max{𝔻𝒢(𝒰1,𝒰2),𝔻𝒢(𝒰2,𝒰1)}\mathbb{H}_{\mathscr{G}}(\mathcal{U}_{1},\mathcal{U}_{2}):=\max\{\mathbb{D}_{\mathscr{G}}(\mathcal{U}_{1},\mathcal{U}_{2}),\mathbb{D}_{\mathscr{G}}(\mathcal{U}_{2},\mathcal{U}_{1})\} be the Hausdorff distance between 𝒰1\mathcal{U}_{1} and 𝒰2\mathcal{U}_{2}.

4.1 Error bound on the ambiguity set

We start by quantifying the difference between the ambiguity sets. To this effect, we need a couple of technical results.

Proposition 4.1

([18, Proposition 4.1]) For each fixed u𝒰Lu\in\mathscr{U}_{L}, let uN𝒰Nu_{N}\in\mathscr{U}_{N} be such that uN(ti)=u(ti)u_{N}(t_{i})=u(t_{i}) for i=1,Ni=1,...N and

uN(t):=u(ti1)+u(ti)u(ti1)titi1(tti1)fort[ti1,ti],i=2,,N.u_{N}(t):=u(t_{i-1})+\frac{u(t_{i})-u(t_{i-1})}{t_{i}-t_{i-1}}(t-t_{i-1})\;{\rm for}\;t\in[t_{i-1},t_{i}],\;i=2,\cdots,N. (4.46)

Then

uNu:=supt[a,b]|uN(t)u(t)|LβN,\|u_{N}-u\|_{\infty}:=\sup_{t\in[a,b]}|u_{N}(t)-u(t)|\leq L\beta_{N}, (4.47)

where βN:=maxi=2,,N(titi1)\beta_{N}:=\max_{i=2,\cdots,N}(t_{i}-t_{i-1}). Moreover, in the case when 𝒢=𝒢K\mathscr{G}=\mathscr{G}_{K}, it holds that

𝖽𝗅𝒢(u,uN)2βN.\displaystyle{\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u,u_{N})\leq 2\beta_{N}}. (4.48)

In the case when 𝒢=𝒢I\mathscr{G}=\mathscr{G}_{I}, 𝖽𝗅𝒢(u,uN)LβN\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u,u_{N})\leq L\beta_{N}.

Here and later on, we call uNu_{N} defined in (4.48) as a projection of uu on 𝒰N\mathscr{U}_{N}. Next, we quantify the deviation distance and Hausdorff distance between ζ\zeta-balls in the 𝒰L\mathscr{U}_{L} and 𝒰N\mathscr{U}_{N}.

Lemma 4.1

Let uN𝒰Nu_{N}\in\mathscr{U}_{N}, u𝒰Lu\in\mathscr{U}_{L} and δ\delta, rr be any positive numbers. Then the following holds:

  • (i)

    𝔻𝒢(𝔹N(uN,r+δ),𝔹N(uN,r))δ\mathbb{D}_{\mathscr{G}}(\mathbb{B}_{N}(u_{N},r+\delta),\mathbb{B}_{N}(u_{N},r))\leq\delta, 𝔻𝒢(𝔹(u,r+δ),𝔹(u,r))δ\mathbb{D}_{\mathscr{G}}(\mathbb{B}(u,r+\delta),\mathbb{B}(u,r))\leq\delta,

  • (ii)

    If uNu_{N} is defined as in (4.46) and 𝒢=𝒢K𝒢I\mathscr{G}=\mathscr{G}_{K}\cup\mathscr{G}_{I}, then 𝒢(𝔹N(uN,r),𝔹(u,r))max{2,L}βN\mathbb{H}_{\mathscr{G}}(\mathbb{B}_{N}(u_{N},r),\mathbb{B}(u,r))\leq\max\{2,L\}\beta_{N} and 𝔻𝒢(𝔹(uN,r+δ),𝔹(uN,r))δ+2max{2,L}βN\mathbb{D}_{\mathscr{G}}(\mathbb{B}(u_{N},r+\delta),\mathbb{B}(u_{N},r))\leq\delta+2\max\{2,L\}\beta_{N}.

Proof. The proof is similar to that of [46], here we include a sketch for self-containedness.

Part (i). We only prove the first inequality, as the second one can be proved analogously. Let u~N𝔹N(uN,r+δ)𝔹N(uN,r)\tilde{u}_{N}\in\mathbb{B}_{N}(u_{N},r+\delta)\setminus\mathbb{B}_{N}(u_{N},r) and uNλ:=λu~N+(1λ)uN𝒰Nu_{N}^{\lambda}:=\lambda\tilde{u}_{N}+(1-\lambda)u_{N}\in\mathscr{U}_{N}, where λ:=r/𝖽𝗅𝒢(uN,u~N)(0,1)\lambda:=r/\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u_{N},\tilde{u}_{N})\in(0,1). By the definition of 𝖽𝗅𝒢\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}, we have 𝖽𝗅𝒢(uNλ,uN)=supg𝒢g,uNλuN=λ𝖽𝗅𝒢(uN,u~N)=r\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u_{N}^{\lambda},u_{N})=\sup_{g\in\mathscr{G}}\langle g,u_{N}^{\lambda}-u_{N}\rangle=\lambda\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u_{N},\tilde{u}_{N})=r, which implies uNλ𝔹N(uN,r)u_{N}^{\lambda}\in\mathbb{B}_{N}(u_{N},r). Thus

𝖽𝗅𝒢(u~N,𝔹N(uN,r))\displaystyle\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u}_{N},\mathbb{B}_{N}(u_{N},r)) \displaystyle\leq 𝖽𝗅𝒢(u~N,uNλ)=(1λ)𝖽𝗅𝒢(u~N,uN)\displaystyle\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u}_{N},u_{N}^{\lambda})=(1-\lambda)\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u}_{N},u_{N})
=\displaystyle= 𝖽𝗅𝒢(u~N,uN)rr+δr=δ.\displaystyle\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u}_{N},u_{N})-r\leq r+\delta-r=\delta.

Since 𝖽𝗅𝒢(u^N,𝔹N(uN,r))=0\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\hat{u}_{N},\mathbb{B}_{N}(u_{N},r))=0 for u^N𝔹N(uN,r)\hat{u}_{N}\in\mathbb{B}_{N}(u_{N},r), we have 𝖽𝗅𝒢(u^N,𝔹N(uN,r))δ\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\hat{u}_{N},\mathbb{B}_{N}(u_{N},r))\leq\delta for all u^N𝔹N(uN,r+δ)\hat{u}_{N}\in\mathbb{B}_{N}(u_{N},r+\delta). By the definition of 𝔻𝒢\mathbb{D}_{\mathscr{G}}, we have

𝔻𝒢(𝔹N(uN,r+δ),𝔹N(uN,r))=supu^N𝔹N(uN,r+δ)𝖽𝗅𝒢(u^N,𝔹(uN,r))\mathbb{D}_{\mathscr{G}}(\mathbb{B}_{N}(u_{N},r+\delta),\mathbb{B}_{N}(u_{N},r))=\sup_{\hat{u}_{N}\in\mathbb{B}_{N}(u_{N},r+\delta)}\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\hat{u}_{N},\mathbb{B}(u_{N},r))

and hence (i) holds.

Part (ii). Let u~N𝔹N(uN,r)\tilde{u}_{N}\in\mathbb{B}_{N}(u_{N},r). Under 4.1,

𝖽𝗅𝒢(u~N,u)𝖽𝗅𝒢(u~N,uN)+𝖽𝗅𝒢(uN,u)r+max{2,L}βN,\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u}_{N},u)\leq\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u}_{N},u_{N})+\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u_{N},u)\leq r+\max\{2,L\}\beta_{N},

which implies 𝔹N(uN,r)𝔹(u,r+max{2,L}βN)\mathbb{B}_{N}(u_{N},r)\subset\mathbb{B}(u,r+\max\{2,L\}\beta_{N}). By Part (i),

𝔻𝒢(𝔹N(uN,r),𝔹(u,r))𝔻𝒢(𝔹(u,r+2βN),𝔹(u,r))max{2,L}βN.\mathbb{D}_{\mathscr{G}}(\mathbb{B}_{N}(u_{N},r),\mathbb{B}(u,r))\leq\mathbb{D}_{\mathscr{G}}(\mathbb{B}(u,r+2\beta_{N}),\mathbb{B}(u,r))\leq\max\{2,L\}\beta_{N}.

Similarly, we have 𝔻𝒢(𝔹(u,r),𝔹N(uN,r))max{2,L}βN\mathbb{D}_{\mathscr{G}}(\mathbb{B}(u,r),\mathbb{B}_{N}(u_{N},r))\leq\max\{2,L\}\beta_{N}. The result holds due to the definition of Hausdorff distance under ζ\zeta-metric.

Now we turn to prove 𝔻𝒢(𝔹(uN,r+δ),𝔹(uN,r))δ+2max{2,L}βN\mathbb{D}_{\mathscr{G}}(\mathbb{B}(u_{N},r+\delta),\mathbb{B}(u_{N},r))\leq\delta+2\max\{2,L\}\beta_{N}. Since uN𝒰Nu_{N}\in\mathscr{U}_{N}, then we can find a u𝒰Lu\in\mathscr{U}_{L} such that uNu_{N} is the projection of uu. Hence for any u~𝔹(uN,r+δ)\tilde{u}\in\mathbb{B}(u_{N},r+\delta), we have 𝖽𝗅𝒢(u~,u)𝖽𝗅𝒢(u~,uN)+𝖽𝗅𝒢(uN,u)r+δ+𝖽𝗅𝒢(uN,u)\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u},u)\leq\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u},u_{N})+\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u_{N},u)\leq r+\delta+\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u_{N},u). Consequently, 𝔹(uN,r+δ)𝔹(u,r+δ+𝖽𝗅𝒢(uN,u))\mathbb{B}(u_{N},r+\delta)\subset\mathbb{B}(u,r+\delta+\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u_{N},u)). On the other hand, for any u~𝔹(u,r𝖽𝗅𝒢(uN,u))\tilde{u}\in\mathbb{B}(u,r-\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u_{N},u)), we have

𝖽𝗅𝒢(u~,uN)𝖽𝗅𝒢(u~,u)+𝖽𝗅𝒢(u,uN)r𝖽𝗅𝒢(u,uN)+𝖽𝗅𝒢(u,uN)=r,\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u},u_{N})\leq\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u},u)+\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u,u_{N})\leq r-\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u,u_{N})+\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u,u_{N})=r,

hence 𝔹(u,r𝖽𝗅𝒢(u,uN))𝔹(uN,r)\mathbb{B}(u,r-\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u,u_{N}))\subset\mathbb{B}(u_{N},r). Therefore, according to the definition of 𝔻𝒢\mathbb{D}_{\mathscr{G}} and Part (i),

𝔻𝒢(𝔹(uN,r+δ),𝔹(uN,r))\displaystyle\mathbb{D}_{\mathscr{G}}(\mathbb{B}(u_{N},r+\delta),\mathbb{B}(u_{N},r)) \displaystyle\leq 𝔻𝒢(𝔹(u,r+δ+𝖽𝗅𝒢(uN,u)),𝔹(u,r𝖽𝗅𝒢(u,uN)))\displaystyle\mathbb{D}_{\mathscr{G}}(\mathbb{B}(u,r+\delta+\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u_{N},u)),\mathbb{B}(u,r-\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u,u_{N})))
\displaystyle\leq δ+2𝖽𝗅𝒢(u,uN)δ+2max{2,L}βN.\displaystyle\delta+2\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u,u_{N})\leq\delta+2\max\{2,L\}\beta_{N}.

The proof is complete.  

With 4.1, we are ready to quantify the difference between 𝔹(u,r)\mathbb{B}(u,r) and 𝔹N(uN,r)\mathbb{B}_{N}(u_{N},r).

Theorem 4.1

Let u𝒰Lu\in\mathscr{U}_{L} and uNu_{N} is a projection of uu defined as in (4.46) and 𝒢=𝒢K𝒢I\mathscr{G}=\mathscr{G}_{K}\cup\mathscr{G}_{I}. Then

𝒢(𝔹(u,r),𝔹N(uN,r))5max{2,L}βN.\mathbb{H}_{\mathscr{G}}(\mathbb{B}(u,r),\mathbb{B}_{N}(u_{N},r))\leq 5\max\{2,L\}\beta_{N}. (4.49)

Proof. By the triangle inequality of the Hausdorff distance in the space of 𝒰L\mathscr{U}_{L}, we have

𝒢(𝔹(u,r),𝔹N(uN,r))𝒢(𝔹(u,r),𝔹(uN,r))+𝒢(𝔹(uN,r),𝔹N(uN,r)).\mathbb{H}_{\mathscr{G}}(\mathbb{B}(u,r),\mathbb{B}_{N}(u_{N},r))\leq\mathbb{H}_{\mathscr{G}}(\mathbb{B}(u,r),\mathbb{B}(u_{N},r))+\mathbb{H}_{\mathscr{G}}(\mathbb{B}(u_{N},r),\mathbb{B}_{N}(u_{N},r)).

From 4.1, 𝒢(𝔹(u,r),𝔹(uN,r))max{2,L}βN\mathbb{H}_{\mathscr{G}}(\mathbb{B}(u,r),\mathbb{B}(u_{N},r))\leq\max\{2,L\}\beta_{N}, so it suffices to show 𝒢(𝔹(uN,r),𝔹N(uN,r))4max{2,L}βN\mathbb{H}_{\mathscr{G}}(\mathbb{B}(u_{N},r),\mathbb{B}_{N}(u_{N},r))\leq 4\max\{2,L\}\beta_{N}. By the definition of 𝔻𝒢\mathbb{D}_{\mathscr{G}},

𝔻𝒢(𝔹(uN,r),𝔹N(uN,r))\displaystyle\mathbb{D}_{\mathscr{G}}(\mathbb{B}(u_{N},r),\mathbb{B}_{N}(u_{N},r)) =\displaystyle= supu~𝔹(uN,r)𝖽𝗅𝒢(u~,𝔹N(uN,r))\displaystyle\sup_{\tilde{u}\in\mathbb{B}(u_{N},r)}\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u},\mathbb{B}_{N}(u_{N},r))
\displaystyle\leq supu~𝔹(uN,r)[𝖽𝗅𝒢(u~,u~N)+𝖽𝗅𝒢(u~N,𝔹N(uN,r)]\displaystyle\sup_{\tilde{u}\in\mathbb{B}(u_{N},r)}[\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u},\tilde{u}_{N})+\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u}_{N},\mathbb{B}_{N}(u_{N},r)]
\displaystyle\leq supu~𝔹(uN,r)[max{2,L}βN+𝖽𝗅𝒢(u~N,𝔹N(uN,r)]\displaystyle\sup_{\tilde{u}\in\mathbb{B}(u_{N},r)}[\max\{2,L\}\beta_{N}+\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u}_{N},\mathbb{B}_{N}(u_{N},r)]
\displaystyle\leq 𝔻𝒢(𝔹N(uN,max{2,L}βN+r),𝔹N(uN,r))+max{2,L}βN\displaystyle\mathbb{D}_{\mathscr{G}}(\mathbb{B}_{N}(u_{N},\max\{2,L\}\beta_{N}+r),\mathbb{B}_{N}(u_{N},r))+\max\{2,L\}\beta_{N}
\displaystyle\leq 2max{2,L}βN,\displaystyle 2\max\{2,L\}\beta_{N},

where u~N\tilde{u}_{N} is the projection of u~\tilde{u}. The second inequality follows from (4.48), the third inequality is due to the fact that for any u~𝔹(uN,r)\tilde{u}\in\mathbb{B}(u_{N},r), its projection u~N\tilde{u}_{N} satisfies

𝖽𝗅𝒢(u~N,uN)𝖽𝗅𝒢(u~N,u~)+𝖽𝗅𝒢(u~,uN)max{2,L}βN+r,\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u}_{N},u_{N})\leq\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u}_{N},\tilde{u})+\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u},u_{N})\leq\max\{2,L\}\beta_{N}+r,

that is, u~N𝔹(uN,max{2,L}βN+r)\tilde{u}_{N}\in\mathbb{B}(u_{N},\max\{2,L\}\beta_{N}+r). The last inequality follows from part (i) of 4.1. Likewise, we have

𝔻𝒢(𝔹N(uN,r),𝔹(uN,r))\displaystyle\mathbb{D}_{\mathscr{G}}(\mathbb{B}_{N}(u_{N},r),\mathbb{B}(u_{N},r)) =\displaystyle= supu~N𝔹N(uN,r)𝖽𝗅𝒢(u~N,𝔹(uN,r))\displaystyle\sup_{\tilde{u}_{N}\in\mathbb{B}_{N}(u_{N},r)}\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u}_{N},\mathbb{B}(u_{N},r))
\displaystyle\leq supu~N𝔹N(uN,r)[𝖽𝗅𝒢(u~N,u~)+𝖽𝗅𝒢(u~,𝔹(uN,r))]\displaystyle\sup_{\tilde{u}_{N}\in\mathbb{B}_{N}(u_{N},r)}[\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u}_{N},\tilde{u})+\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u},\mathbb{B}(u_{N},r))]
\displaystyle\leq supu~N𝔹N(uN,r)max{2,L}βN+𝖽𝗅𝒢(u~,𝔹(uN,r))\displaystyle\sup_{\tilde{u}_{N}\in\mathbb{B}_{N}(u_{N},r)}\max\{2,L\}\beta_{N}+\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u},\mathbb{B}(u_{N},r))
\displaystyle\leq supu~N𝔹N(uN,r)max{2,L}βN+𝔻𝒢(𝔹(uN,max{2,L}βN+r),𝔹(uN,r))\displaystyle\sup_{\tilde{u}_{N}\in\mathbb{B}_{N}(u_{N},r)}\max\{2,L\}\beta_{N}+\mathbb{D}_{\mathscr{G}}(\mathbb{B}(u_{N},\max\{2,L\}\beta_{N}+r),\mathbb{B}(u_{N},r))
\displaystyle\leq 4max{2,L}βN,\displaystyle 4\max\{2,L\}\beta_{N},

where the third inequality is derived from the fact that for any u~N𝔹N(uN,r)\tilde{u}_{N}\in\mathbb{B}_{N}(u_{N},r), that is, 𝖽𝗅𝒢(u~N,uN)r\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u}_{N},u_{N})\leq r, we have 𝖽𝗅𝒢(u~,uN)𝖽𝗅𝒢(u~,u~N)+𝖽𝗅𝒢(u~N,uN)𝖽𝗅𝒢(u~,u~N)+rmax{2,L}βN+r\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u},u_{N})\leq\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u},\tilde{u}_{N})+\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u}_{N},u_{N})\leq\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u},\tilde{u}_{N})+r\leq\max\{2,L\}\beta_{N}+r, that is u~𝔹(uN,max{2,L}βN+r)\tilde{u}\in\mathbb{B}(u_{N},\max\{2,L\}\beta_{N}+r). The last inequality follows from part (ii) of 4.1. Finally, by the definition of Hausdorff distance under metric 𝖽𝗅𝒢\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}, the proof is complete.  

4.2 Error bound on the optimal value

Theorem 4.2

Let u0𝒰Lu^{0}\in\mathscr{U}_{L} and uN0𝒰Nu_{N}^{0}\in\mathscr{U}_{N} be defined as in (4.46). If 𝒢=𝒢K𝒢I\mathscr{G}=\mathscr{G}_{K}\cup\mathscr{G}_{I} in (4.45), then |R(ξ)RN(ξ)|10max{2,L}βN|R(\xi)-R_{N}(\xi)|\leq 10\max\{2,L\}\beta_{N}.

Proof. It is well known that

|RN(ξ)R(ξ)|\displaystyle|R_{N}(\xi)-R(\xi)| \displaystyle\leq maxx|infu𝔹(u0,r){u(x)+𝔼P[u(ξx)]}\displaystyle\max_{x\in\mathds{R}}\left|\inf_{u\in\mathbb{B}(u^{0},r)}\left\{u(x)+{\mathbb{E}}_{P}[u(\xi-x)]\right\}\right.
infuN𝔹N(uN0,r){uN(x)+𝔼P[uN(ξx)]}|.\displaystyle\left.\quad\quad\quad-\inf_{u_{N}\in\mathbb{B}_{N}(u^{0}_{N},r)}\left\{u_{N}(x)+{\mathbb{E}}_{P}[u_{N}(\xi-x)]\right\}\right|.

Let δ\delta be a small positive number. For any xx\in\mathds{R}, we can find ux𝔹(u0,r)u^{x}\in\mathbb{B}(u^{0},r) and uNx𝔹N(uN0,r)u^{x}_{N}\in\mathbb{B}_{N}(u^{0}_{N},r) depending on δ\delta such that

ux(x)+𝔼P[ux(ξx)]\displaystyle u^{x}(x)+{\mathbb{E}}_{P}[u^{x}(\xi-x)] \displaystyle\leq infu𝔹(u0,r){u(x)+𝔼P[u(ξx)]}+δ,\displaystyle\inf_{u\in{\mathbb{B}(u^{0},r)}}\{u(x)+{\mathbb{E}}_{P}[u(\xi-x)]\}+\delta,
uNx(x)+𝔼P[uNx(ξx)]\displaystyle u^{x}_{N}(x)+{\mathbb{E}}_{P}[u^{x}_{N}(\xi-x)] \displaystyle\geq infuN𝔹N(uN0,r){uN(x)+𝔼P[uN(ξx)]},\displaystyle\inf_{u_{N}\in{\mathbb{B}_{N}(u_{N}^{0},r)}}\{u_{N}(x)+{\mathbb{E}}_{P}[u_{N}(\xi-x)]\},
supt[a,b]|uNx(t)ux(t)|\displaystyle\sup_{t\in[a,b]}|u_{N}^{x}(t)-u^{x}(t)| \displaystyle\leq (𝔹N(uN0,r),𝔹(u0,r))+δ,\displaystyle\mathbb{H}(\mathbb{B}_{N}(u_{N}^{0},r),\mathbb{B}(u^{0},r))+\delta,

where \mathbb{H} denotes the Hausdorff distance in the space of continuous functions defined on [a,b][a,b] equipped with infinity norm \|\cdot\|_{\infty}. Combining the above inequalities

infuN𝔹N(uN0,r){uN(x)+𝔼P[uN(ξx)]}infu𝔹(u0,r){u(x)+𝔼P[u(ξx)]}\displaystyle\inf_{u_{N}\in{\mathbb{B}_{N}(u^{0}_{N},r)}}\{u_{N}(x)+{\mathbb{E}}_{P}[u_{N}(\xi-x)]\}-\inf_{u\in{\mathbb{B}(u^{0},r)}}\{u(x)+{\mathbb{E}}_{P}[u(\xi-x)]\}
𝔼P[uNx(ξx)ux(ξx)]+(uNx(x)ux(x))+δ\displaystyle\leq{\mathbb{E}}_{P}[u_{N}^{x}(\xi-x)-u^{x}(\xi-x)]+(u_{N}^{x}(x)-u^{x}(x))+\delta
2uNxux+δ\displaystyle\leq 2\|u_{N}^{x}-u^{x}\|_{\infty}+\delta
2(𝔹N(uN0,r),𝔹(u0,r))+3δ.\displaystyle\leq 2\mathbb{H}(\mathbb{B}_{N}(u^{0}_{N},r),\mathbb{B}(u^{0},r))+3\delta.

By exchanging the positions of 𝔹N(uN0,r)\mathbb{B}_{N}(u^{0}_{N},r) and 𝔹(u0,r)\mathbb{B}(u^{0},r), we have

infu𝔹(u0,r){u(x)+𝔼P[u(ξx)]}infuN𝔹N(uN0,r){uN(x)+𝔼P[uN(ξx)]}\displaystyle\inf_{u\in{\mathbb{B}(u^{0},r)}}\{u(x)+{\mathbb{E}}_{P}[u(\xi-x)]\}-\inf_{u_{N}\in{\mathbb{B}_{N}(u^{0}_{N},r)}}\{u_{N}(x)+{\mathbb{E}}_{P}[u_{N}(\xi-x)]\}
2(𝔹(u0,r),𝔹N(uN0,r))+3δ.\displaystyle\leq 2\mathbb{H}(\mathbb{B}(u^{0},r),\mathbb{B}_{N}(u^{0}_{N},r))+3\delta.

Since δ0\delta\geq 0 can be arbitrarily small, we obtain

|RN(ξ)R(ξ)|\displaystyle|R_{N}(\xi)-R(\xi)| \displaystyle\leq maxx|infu𝔹(u0,r){u(x)+𝔼P[u(ξx)]}infu𝔹N(uN0,r){u(x)+𝔼P[u(ξx)]}|\displaystyle\max_{x\in\mathds{R}}\left|\inf_{u\in{\mathbb{B}(u^{0},r)}}\{u(x)+{\mathbb{E}}_{P}[u(\xi-x)]\}-\inf_{u\in{\mathbb{B}_{N}(u_{N}^{0},r)}}\{u(x)+{\mathbb{E}}_{P}[u(\xi-x)]\}\right|
\displaystyle\leq 2(𝔹(u0,r),𝔹N(uN0,r)).\displaystyle 2\mathbb{H}(\mathbb{B}(u^{0},r),\mathbb{B}_{N}(u^{0}_{N},r)).

The main challenge here is that (𝔹(u0,r),𝔹N(uN0,r))\mathbb{H}(\mathbb{B}(u^{0},r),\mathbb{B}_{N}(u^{0}_{N},r)) differs from 𝒢(𝔹(u0,r),𝔹N(uN0,r))\mathbb{H}_{\mathscr{G}}(\mathbb{B}(u^{0},r),\mathbb{B}_{N}(u^{0}_{N},r)). In what follows, we show that

𝒢I(𝔹(u0,r),𝔹N(uN0,r))=(𝔹(u0,r),𝔹N(uN0,r)),\mathbb{H}_{\mathscr{G}_{I}}(\mathbb{B}(u^{0},r),\mathbb{B}_{N}(u^{0}_{N},r))=\mathbb{H}(\mathbb{B}(u^{0},r),\mathbb{B}_{N}(u^{0}_{N},r)), (4.50)

where 𝒢I={𝟙(a,t]():t[a,b]}\mathscr{G}_{I}=\{\mathds{1}_{(a,t]}(\cdot):t\in[a,b]\}. Let u~𝔹(u0,r)\tilde{u}\in\mathbb{B}(u^{0},r) and u~N𝔹N(uN0,r)\tilde{u}_{N}\in\mathbb{B}_{N}(u^{0}_{N},r),

𝖽𝗅𝒢I(u~,u~N)\displaystyle\mathsf{d\kern-0.70007ptl}_{\mathscr{G}_{I}}(\tilde{u},\tilde{u}_{N}) =\displaystyle= supg𝒢I|abg(t)𝑑u~(t)abg(t)𝑑u~N(t)|\displaystyle\sup_{g\in\mathscr{G}_{I}}\left|\int_{a}^{b}g(t)d\tilde{u}(t)-\int_{a}^{b}g(t)d\tilde{u}_{N}(t)\right| (4.51)
=\displaystyle= supt[a,b]|at1𝑑u~(s)at1𝑑u~N(s)|\displaystyle\sup_{t\in[a,b]}\left|\int_{a}^{t}1d\tilde{u}(s)-\int_{a}^{t}1d\tilde{u}_{N}(s)\right|
=\displaystyle= supt[a,b]|u~(t)u~N(t)u~(a)+u~N(a)|\displaystyle\sup_{t\in[a,b]}\left|\tilde{u}(t)-\tilde{u}_{N}(t)-\tilde{u}(a)+\tilde{u}_{N}(a)\right|
=\displaystyle= supt[a,b]|u~(t)u~N(t)|.\displaystyle\sup_{t\in[a,b]}|\tilde{u}(t)-\tilde{u}_{N}(t)|.

The last equality is due to the fact that u(a)=uN(a)=0u(a)=u_{N}(a)=0. By taking infimum w.r.t. u~N𝔹N(uN0,r)\tilde{u}_{N}\in\mathbb{B}_{N}(u^{0}_{N},r) and taking superemum w.r.t. u~𝔹(u0,r)\tilde{u}\in\mathbb{B}(u^{0},r) on both sides of the equality above, we obtain 𝔻𝒢I(𝔹(u0,r),𝔹N(uN0,r))=𝔻(𝔹(u0,r),𝔹N(uN0,r)).\mathbb{D}_{\mathscr{G}_{I}}(\mathbb{B}(u^{0},r),\mathbb{B}_{N}(u_{N}^{0},r))=\mathbb{D}(\mathbb{B}(u^{0},r),\mathbb{B}_{N}(u_{N}^{0},r)). Swapping the positions between 𝔹(u0,r)\mathbb{B}(u^{0},r) and 𝔹N(uN0,r)\mathbb{B}_{N}(u_{N}^{0},r), we obtain (4.50). Combining with Theorem 4.1, we obtain the conclusion.  

5 Extensions

In this section, we discuss potential extensions of the MOCE and RMOCE models by considering utility functions with unbounded domain and multivariate utility functions.

5.1 Utility function with unbounded domain

In some important applications such as finance and economics, the underlying random variables which represent market demand, stock price and rate of return often have unbounded support. This raises a question as to whether our proposed model and computational schemes in the previous sections can be effectively applied to these situations. Here we discuss this issue.

We start by defining a set of nonconstant increasing function defined over IR{\rm I\!R} denoted by 𝒰\mathscr{U}_{\infty}. We no longer restrict the domain of uu to a bounded interval [a,b][a,b]. Let u0𝒰u^{0}\in\mathscr{U}_{\infty}, the ζ\zeta-ball in 𝒰\mathscr{U}_{\infty} is defined as

𝔹(u0,r):={u𝒰|𝖽𝗅𝒢(u,u0)r},\mathbb{B}_{\infty}(u^{0},r):=\{u\in\mathscr{U}_{\infty}\;|\;\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u,u^{0})\leq r\},

where 𝖽𝗅𝒢\mathsf{d\kern-0.70007ptl}_{\mathscr{G}} is the pseudo metric defined in (4.41) and 𝒢\mathscr{G} is a set of measurable function throughout this section. The robust modified optimized certainty equivalent model based on 𝔹(u0,r)\mathbb{B}_{\infty}(u^{0},r) is defined as

(RMOCE)R(ξ):=maxxXminu𝔹(u0,r)u(x)+𝔼P[u(ξx)],{\rm(RMOCE)_{\infty}}\quad\;\;R_{\infty}(\xi):=\displaystyle{\max_{x\in X}\min_{u\in\mathbb{B}_{\infty}(u^{0},r)}}\;\;u(x)+{\mathbb{E}}_{P}[u(\xi-x)], (5.52)

where XX is a compact implementable decisions over XIRX\subset{\rm I\!R}. Our aim is to solve (RMOCE)\rm(RMOCE)_{\infty} and our concern is that the numerical schemes proposed in Section 3 cannot be applied to this problem directly. Let utruc0u^{0}_{\rm truc} be the truncation of u0u^{0} over [a,b][a,b], define

𝔹[a,b](utruc0,r):={u𝒰[a,b]|𝖽𝗅𝒢(u,utruc0)r},\mathbb{B}_{[a,b]}(u^{0}_{\rm truc},r):=\{u\in\mathscr{U}_{[a,b]}\;|\;\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u,u^{0}_{\rm truc})\leq r\}, (5.53)

where 𝒰[a,b]\mathscr{U}_{[a,b]} denotes the set of nonconstant nondecreasing functions defined over [a,b][a,b]. We rewrite (4.43) as

(RMOCE)[a,b]R[a,b](ξ):=maxxXminu𝔹[a,b](utruc0,r)u(x)+𝔼P[u(ξx)].{\rm(RMOCE)_{[a,b]}}\quad\;\;R_{[a,b]}(\xi):=\displaystyle{\max_{x\in X}\min_{u\in\mathbb{B}_{[a,b]}(u^{0}_{\rm truc},r)}}\;\;u(x)+{\mathbb{E}}_{P}[u(\xi-x)]. (5.54)

What we are interested here is the difference between (RMOCE)\rm(RMOCE)_{\infty} and (RMOCE)[a,b]\rm(RMOCE)_{[a,b]} in terms of the optimal value. We will show that the difference between R(ξ)R_{\infty}(\xi) and R[a,b](ξ)R_{[a,b]}(\xi) is only related with the radius of the ζ\zeta-ball under some moderate conditions and therefore we may solve (RMOCE)\rm(RMOCE)_{\infty} approximately by solving (RMOCE)[a,b]\rm(RMOCE)_{[a,b]}. The latter can be solved by the piecewise linear approximation scheme detailed in Section 3.

To build the bridge between 𝔹[a,b](utruc0,r)\mathbb{B}_{[a,b]}(u^{0}_{\rm truc},r) and 𝔹(u0,r)\mathbb{B}_{\infty}(u^{0},r), we define the following set

𝔹~[a,b](u0,r):={u𝒰:𝖽𝗅𝒢(u,u0)r,u(t)=u(a) for t<a,u~(t)=u(b) for t>b}.\tilde{\mathbb{B}}_{[a,b]}(u^{0},r):=\{u\in{\mathscr{U}}_{\infty}:\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u,u^{0})\leq r,u(t)=u(a)\text{~{}for~{}}t<a,\tilde{u}(t)=u(b)\text{~{}for~{}}t>b\}. (5.55)

Notice that 𝔹~[a,b](u0,r)\tilde{\mathbb{B}}_{[a,b]}(u^{0},r) is not a ball which is defined under the pseudo metric. Then we can establish the connection between 𝔹~[a,b]\tilde{\mathbb{B}}_{[a,b]} and 𝔹\mathbb{B}_{\infty} in the following proposition.

Proposition 5.1

Let u0𝒰u^{0}\in\mathscr{U}_{\infty} and assume that there exists a position number θ\theta such that

supg𝒢,u𝔹(u0,r)IR|g(t)|𝑑u(t)θ.\sup_{g\in\mathscr{G},u\in\mathbb{B}_{\infty}(u^{0},r)}\int_{{\rm I\!R}}|g(t)|du(t)\leq\theta. (5.56)

Then for any ϵ>0\epsilon>0 there exist constants a<0a<0 and b>0b>0 such that

𝒢(𝔹~[a,b](u0,r+ϵ),𝔹(u0,r))supg𝒢,u𝔹(u0,r)|IR[a,b]g(t)𝑑u(t)|ϵ.\mathbb{H}_{\mathscr{G}}(\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon),\mathbb{B}_{\infty}(u^{0},r))\leq\sup_{g\in\mathscr{G},u\in\mathbb{B}_{\infty}(u^{0},r)}\left|\int_{{\rm I\!R}\setminus[a,b]}g(t)du(t)\right|\leq\epsilon. (5.57)

Proof. From the condition (5.56), for any ϵ>0\epsilon>0 there exist constants a<0a<0 and b>0b>0 such that

supg𝒢,u𝔹(u0,r)IR[a,b]|g(t)|𝑑u(t)ϵ.\sup_{g\in\mathscr{G},u\in\mathbb{B}_{\infty}(u^{0},r)}\int_{{\rm I\!R}\setminus[a,b]}|g(t)|du(t)\leq\epsilon. (5.58)

For any fixed u𝔹(u0,r)u\in\mathbb{B}_{\infty}(u^{0},r) , let u~=u(t)\tilde{u}=u(t) for t[a,b]t\in[a,b] and u~(t)=u(a)\tilde{u}(t)=u(a) for t<at<a and u~(t)=u(b)\tilde{u}(t)=u(b) for t>bt>b. Then we can obtain

𝖽𝗅𝒢(u~,u0)\displaystyle\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(\tilde{u},u^{0}) =\displaystyle= supg𝒢|IRg(t)𝑑u~(t)IRg(t)𝑑u0(t)|\displaystyle\sup_{g\in\mathscr{G}}\left|\int_{{\rm I\!R}}g(t)d\tilde{u}(t)-\int_{{\rm I\!R}}g(t)du^{0}(t)\right|
=\displaystyle= supg𝒢|abg(t)d(u~(t)u0(t))IR[a,b]g(t)𝑑u0(t)|\displaystyle\sup_{g\in\mathscr{G}}\left|\int_{a}^{b}g(t)d(\tilde{u}(t)-u^{0}(t))-\int_{{\rm I\!R}\setminus[a,b]}g(t)du^{0}(t)\right|
=\displaystyle= supg𝒢|abg(t)d(u~(t)u0(t))+IR[a,b]g(t)d(u(t)u0(t))IR[a,b]g(t)𝑑u(t)|\displaystyle\sup_{g\in\mathscr{G}}\left|\int_{a}^{b}g(t)d(\tilde{u}(t)-u^{0}(t))+\int_{{\rm I\!R}\setminus[a,b]}g(t)d(u(t)-u^{0}(t))-\int_{{\rm I\!R}\setminus[a,b]}g(t)du(t)\right|
\displaystyle\leq r+supg𝒢|IR[a,b]g(t)𝑑u(t)|r+ϵ.\displaystyle r+\sup_{g\in\mathscr{G}}\left|\int_{{\rm I\!R}\setminus[a,b]}g(t)du(t)\right|\leq r+\epsilon.

Hence u~𝔹~[a,b](u0,r+ϵ)\tilde{u}\in\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon) and

𝔻𝒢(u,𝔹~[a,b](u0,r+ϵ))𝖽𝗅𝒢(u,u~)=supg𝒢|IR[a,b]g(t)𝑑u(t)|ϵ.\mathbb{D}_{\mathscr{G}}(u,\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon))\leq\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(u,\tilde{u})=\sup_{g\in\mathscr{G}}\left|\int_{{\rm I\!R}\setminus[a,b]}g(t)du(t)\right|\leq\epsilon. (5.59)

By taking supremum w.r.t. uu over 𝔹(u0,r)\mathbb{B}_{\infty}(u^{0},r) on both sides of (5.59), we obtain

𝔻𝒢(𝔹(u0,r),𝔹~[a,b](u0,r+ϵ))ϵ.\mathbb{D}_{\mathscr{G}}(\mathbb{B}_{\infty}(u^{0},r),\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon))\leq\epsilon.

Note that 𝔹~[a,b](u0,r+ϵ)𝔹(u0,r+ϵ)\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon)\subset\mathbb{B}_{\infty}(u^{0},r+\epsilon), then we have

𝔻𝒢(𝔹~[a,b](u0,r+ϵ),𝔹(u0,r))𝔻𝒢(𝔹(u0,r+ϵ),𝔹(u0,r))ϵ,\mathbb{D}_{\mathscr{G}}(\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon),\mathbb{B}_{\infty}(u^{0},r))\leq\mathbb{D}_{\mathscr{G}}(\mathbb{B}_{\infty}(u^{0},r+\epsilon),\mathbb{B}_{\infty}(u^{0},r))\leq\epsilon,

where the last inequality is from part (i) of 4.1. Consequently (5.57) follows.  

From 5.1, we can see that when the interval [a,b][a,b] is large enough, the difference between 𝔹~[a,b](u0,r+ϵ)\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon) and 𝔹(u0,r)\mathbb{B}_{\infty}(u^{0},r) will not be significant. We now turn to compare 𝔹~[a,b](u0,r+ϵ)\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon) with 𝔹[a,b](utruc0,r)\mathbb{B}_{[a,b]}(u^{0}_{\rm truc},r), and we could get similar conclusion in [18, Section 6.2] that the extended function u~\tilde{u} of u𝔹[a,b](utruc0,r)u\in\mathbb{B}_{[a,b]}(u^{0}_{\rm truc},r) is in 𝔹~[a,b](u0,r+ϵ)\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon), where ϵsupg𝒢|IR[a,b]g(t)𝑑u0(t)|\epsilon\geq\sup_{g\in\mathscr{G}}\left|\int_{{\rm I\!R}\setminus[a,b]}g(t)du^{0}(t)\right|, because u0𝔹(u0,r)u^{0}\in\mathbb{B}_{\infty}(u^{0},r). By exploiting the relationship, we can quantify the difference between R(ξ)R_{\infty}(\xi) and R[a,b](ξ)R_{[a,b]}(\xi) in the following theorem.

Theorem 5.1

Assume there exists a constant δ>0\delta>0 such that

supu𝔹(u0,r+δ),xXIR|u(ξx)|P(dξ)<\sup_{u\in\mathbb{B}_{\infty}(u^{0},r+\delta),x\in X}\int_{{\rm I\!R}}|u(\xi-x)|P(d\xi)<\infty (5.60)

and the condition in 5.1 is fulfilled. Then for any ϵ>0\epsilon>0, there exist constants a<0a<0 and b>0b>0 such that

|R(ξ)R[a,b](ξ)|3ϵ.|R_{\infty}(\xi)-R_{[a,b]}(\xi)|\leq 3\epsilon. (5.61)

Proof. It follows from conditions (5.60) and (5.56) that for any 0<ϵ<δ0<\epsilon<\delta there exist constants a<0<ba<0<b such that

supu𝔹(u0,r+δ),xXξxIR[a,b]|u(ξx)|P(dξ)ϵ/3\sup_{u\in\mathbb{B}_{\infty}(u^{0},r+\delta),x\in X}\int_{\xi-x\in{\rm I\!R}\setminus[a,b]}|u(\xi-x)|P(d\xi)\leq\epsilon/3 (5.62)

and (5.58) holds. Since 𝔹~[a,b](u0,r+ϵ)𝔹(u0,r+δ)\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon)\subset\mathbb{B}_{\infty}(u^{0},r+\delta), the above inequality implies

supu𝔹~[a,b](u0,r+ϵ)(|u(a)|P((,a))+|u(b)|P((b,+)))ϵ/3.\sup_{u\in\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon)}\bigl{(}|u(a)|P((-\infty,a))+|u(b)|P((b,+\infty))\bigr{)}\leq\epsilon/3. (5.63)

By definitions of R(ξ)R_{\infty}(\xi) and R[a,b](ξ)R_{[a,b]}(\xi),

|R(ξ)R[a,b](ξ)|\displaystyle|R_{\infty}(\xi)-R_{[a,b]}(\xi)|
\displaystyle\leq supxX|infu𝔹(u0,r)[u(x)+IRu(ξx)P(dξ)]infu^𝔹[a,b](utruc0,r)[u^(x)+ξx[a,b]u^(ξx)P(dξ)]|\displaystyle\sup_{x\in X}\left|\inf_{u\in\mathbb{B}_{\infty}(u^{0},r)}\left[u(x)+\int_{{\rm I\!R}}u(\xi-x)P(d\xi)\right]-\inf_{\hat{u}\in\mathbb{B}_{[a,b]}(u^{0}_{\rm truc},r)}\left[\hat{u}(x)+\int_{\xi-x\in[a,b]}\hat{u}(\xi-x)P(d\xi)\right]\right|
\displaystyle\leq supxX|infu𝔹(u0,r)[u(x)+IRu(ξx)P(dξ)]infu~𝔹~[a,b](u0,r+ϵ)[u~(x)+IRu~(ξx)P(dξ)]|\displaystyle\sup_{x\in X}\left|\inf_{u\in\mathbb{B}_{\infty}(u^{0},r)}\left[u(x)+\int_{{\rm I\!R}}u(\xi-x)P(d\xi)\right]-\inf_{\tilde{u}\in\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon)}\left[\tilde{u}(x)+\int_{{\rm I\!R}}\tilde{u}(\xi-x)P(d\xi)\right]\right|
+\displaystyle+ supxX|infu~𝔹~[a,b](u0,r+ϵ)[u~(x)+IRu~(ξx)P(dξ)]infu^𝔹[a,b](utruc0,r)[u^(x)+ξx[a,b]u^(ξx)P(dξ)]|.\displaystyle\sup_{x\in X}\left|\inf_{\tilde{u}\in\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon)}\left[\tilde{u}(x)+\int_{{\rm I\!R}}\tilde{u}(\xi-x)P(d\xi)\right]-\inf_{\hat{u}\in\mathbb{B}_{[a,b]}(u^{0}_{\rm truc},r)}\left[\hat{u}(x)+\int_{\xi-x\in[a,b]}\hat{u}(\xi-x)P(d\xi)\right]\right|.

Let us estimate the first term at the right side of the last inequality above. Observe that

infu𝔹(u0,r)[u(x)+IRu(ξx)P(dξ)]infu~𝔹~[a,b](u0,r+ϵ)[u~(x)+IRu~(ξx)P(dξ)]\displaystyle\inf_{u\in\mathbb{B}_{\infty}(u^{0},r)}\left[u(x)+\int_{{\rm I\!R}}u(\xi-x)P(d\xi)\right]-\inf_{\tilde{u}\in\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon)}\left[\tilde{u}(x)+\int_{{\rm I\!R}}\tilde{u}(\xi-x)P(d\xi)\right]
infu𝔹(u0,r)supu~𝔹~[a,b](u0,r+ϵ)[|IRu(ξx)P(dξ)IRu~(ξx)P(dξ)|+|u(x)u~(x)|]\displaystyle\leq\inf_{u\in\mathbb{B}_{\infty}(u^{0},r)}\sup_{\tilde{u}\in\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon)}\left[\left|\int_{{\rm I\!R}}u(\xi-x)P(d\xi)-\int_{{\rm I\!R}}\tilde{u}(\xi-x)P(d\xi)\right|+|u(x)-\tilde{u}(x)|\right]
infu𝔹(u0,r)supu~𝔹~[a,b](u0,r+ϵ)[|ξx[a,b]u(ξx)P(dξ)ξx[a,b]u~(ξx)P(dξ)|\displaystyle\leq\inf_{u\in\mathbb{B}_{\infty}(u^{0},r)}\sup_{\tilde{u}\in\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon)}\Bigg{[}\left|\int_{\xi-x\in[a,b]}u(\xi-x)P(d\xi)-\int_{\xi-x\in[a,b]}\tilde{u}(\xi-x)P(d\xi)\right|
+|u(x)u~(x)|+2ϵ3]\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad+|u(x)-\tilde{u}(x)|+\frac{2\epsilon}{3}\Bigg{]}
infu𝔹(u0,r)supu~𝔹~[a,b](u0,r+ϵ)[supt[a,b]|u(t)u~(t)|+|u(x)u~(x)|+2ϵ3]\displaystyle\leq\inf_{u\in\mathbb{B}_{\infty}(u^{0},r)}\sup_{\tilde{u}\in\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon)}\left[\sup_{t\in[a,b]}|u(t)-\tilde{u}(t)|+|u(x)-\tilde{u}(x)|+\frac{2\epsilon}{3}\right]
infu𝔹(u0,r)supu~𝔹~[a,b](u0,r+ϵ)[𝖽𝗅𝒢I(u,u~)+|u(x)u~(x)|+2ϵ3].\displaystyle\leq\inf_{u\in\mathbb{B}_{\infty}(u^{0},r)}\sup_{\tilde{u}\in\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon)}\left[\mathsf{d\kern-0.70007ptl}_{\mathscr{G}_{I}}(u,\tilde{u})+|u(x)-\tilde{u}(x)|+\frac{2\epsilon}{3}\right].

Thus

supxX|infu𝔹(u0,r)[u(x)+IRu(ξx)P(dξ)]infu~𝔹~[a,b](u0,r+ϵ)[u~(x)+IRu~(ξx)P(dξ)]|\displaystyle\sup_{x\in X}\left|\inf_{u\in\mathbb{B}_{\infty}(u^{0},r)}\left[u(x)+\int_{{\rm I\!R}}u(\xi-x)P(d\xi)\right]-\inf_{\tilde{u}\in\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon)}\left[\tilde{u}(x)+\int_{{\rm I\!R}}\tilde{u}(\xi-x)P(d\xi)\right]\right|
2𝒢I(𝔹(u0,r),𝔹~[a,b](u0,r+ϵ))+2ϵ3.\displaystyle\leq 2\mathbb{H}_{\mathscr{G}_{I}}(\mathbb{B}_{\infty}(u^{0},r),\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon))+\frac{2\epsilon}{3}. (5.64)

The last inequality holds due to X[a,b]X\subset[a,b]. Now let us turn to the second term. For any xXx\in X and a fixed positive number ε\varepsilon, we can find u^ξ𝔹[a,b](utruc0,r)\hat{u}_{\xi}\in\mathbb{B}_{[a,b]}(u^{0}_{\rm truc},r) and its extended function u~ξ𝔹~[a,b](u0,r+ϵ)\tilde{u}_{\xi}\in\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon) such that

u^ξ(x)+ξx[a,b]u^ξ(ξx)P(dξ)\displaystyle\hat{u}_{\xi}(x)+\int_{\xi-x\in[a,b]}\hat{u}_{\xi}(\xi-x)P(d\xi) \displaystyle\leq infu^𝔹[a,b](utruc0,r)[u^(x)+ξx[a,b]u^(ξx)P(dξ)]+ε,\displaystyle\inf_{\hat{u}\in\mathbb{B}_{[a,b]}(u^{0}_{\rm truc},r)}\left[\hat{u}(x)+\int_{\xi-x\in[a,b]}\hat{u}(\xi-x)P(d\xi)\right]+\varepsilon,
u~ξ(x)+IRu~ξ(ξx)P(dξ)\displaystyle\tilde{u}_{\xi}(x)+\int_{{\rm I\!R}}\tilde{u}_{\xi}(\xi-x)P(d\xi) \displaystyle\geq infu~𝔹~[a,b](u0,r+ϵ)[u~(x)+IRu~(ξx)P(dξ)].\displaystyle\inf_{\tilde{u}\in\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon)}\left[\tilde{u}(x)+\int_{{\rm I\!R}}\tilde{u}(\xi-x)P(d\xi)\right].

Consequently we have

infu~𝔹~[a,b](u0,r+ϵ)[u~(x)+IRu~(ξx)P(dξ)]infu^𝔹[a,b](utruc0,r)[u^(x)+ξx[a,b]u^(ξx)P(dξ)]\displaystyle\inf_{\tilde{u}\in\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon)}\left[\tilde{u}(x)+\int_{{\rm I\!R}}\tilde{u}(\xi-x)P(d\xi)\right]-\inf_{\hat{u}\in\mathbb{B}_{[a,b]}(u^{0}_{\rm truc},r)}\left[\hat{u}(x)+\int_{\xi-x\in[a,b]}\hat{u}(\xi-x)P(d\xi)\right]
u~ξ(x)+IRu~ξ(ξx)P(dξ)u^ξ(x)ξx[a,b]u^ξ(ξx)P(dξ)+ε\displaystyle\leq\tilde{u}_{\xi}(x)+\int_{{\rm I\!R}}\tilde{u}_{\xi}(\xi-x)P(d\xi)-\hat{u}_{\xi}(x)-\int_{\xi-x\in[a,b]}\hat{u}_{\xi}(\xi-x)P(d\xi)+\varepsilon
=ξxIR[a,b]u~ξ(ξx)P(dξ)+ε\displaystyle=\int_{\xi-x\in{\rm I\!R}\setminus[a,b]}\tilde{u}_{\xi}(\xi-x)P(d\xi)+\varepsilon
supu~𝔹~[a,b](u0,r+ϵ)(|u~(a)|P((,a))+|u~(b)|P((b,+)))+ε.\displaystyle\leq\sup_{\tilde{u}\in\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon)}\Bigl{(}|\tilde{u}(a)|P((-\infty,a))+|\tilde{u}(b)|P((b,+\infty))\Bigr{)}+\varepsilon. (5.65)

The second equality is satisfied because u~ξ\tilde{u}_{\xi} is the extended function of u^ξ\hat{u}_{\xi} and x[a,b]x\in[a,b]. By exchanging the positions of 𝔹~[a,b](u0,r+ϵ)\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon) and 𝔹[a,b](utruc0,r)\mathbb{B}_{[a,b]}(u^{0}_{\rm truc},r), we have

infu^𝔹[a,b](utruc0,r)[u^(x)+ξx[a,b]u^(ξx)P(dξ)]infu~𝔹~[a,b](u0,r+ϵ)[u~(x)+IRu~(ξx)P(dξ)]\displaystyle\inf_{\hat{u}\in\mathbb{B}_{[a,b]}(u^{0}_{\rm truc},r)}\left[\hat{u}(x)+\int_{\xi-x\in[a,b]}\hat{u}(\xi-x)P(d\xi)\right]-\inf_{\tilde{u}\in\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon)}\left[\tilde{u}(x)+\int_{{\rm I\!R}}\tilde{u}(\xi-x)P(d\xi)\right]
supu~𝔹~[a,b](u0,r+ϵ)(|u~(a)|P((,a))+|u~(b)|P((b,+)))+ε.\displaystyle\leq\sup_{\tilde{u}\in\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon)}\Bigl{(}|\tilde{u}(a)|P((-\infty,a))+|\tilde{u}(b)|P((b,+\infty))\Bigr{)}+\varepsilon. (5.66)

Since ε\varepsilon can be arbitrarily small, we obtain

supxX|infu^𝔹[a,b](utruc0,r)[u^(x)+ξx[a,b]u^(ξx)P(dξ)]\displaystyle\sup_{x\in X}\Bigg{|}\inf_{\hat{u}\in\mathbb{B}_{[a,b]}(u^{0}_{\rm truc},r)}\left[\hat{u}(x)+\int_{\xi-x\in[a,b]}\hat{u}(\xi-x)P(d\xi)\right]
infu~2𝔹~[a,b](u0,r+ϵ)[u~(x)+IRu~(ξx)P(dξ)]|\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad-\inf_{\tilde{u}_{2}\in\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon)}\left[\tilde{u}(x)+\int_{{\rm I\!R}}\tilde{u}(\xi-x)P(d\xi)\right]\Bigg{|}
supu~𝔹~[a,b](u0,r+ϵ)(|u~(a)|P((,a))+|u~(b)|P((b,+)))ϵ/3.\displaystyle\leq\sup_{\tilde{u}\in\tilde{\mathbb{B}}_{[a,b]}(u^{0},r+\epsilon)}(|\tilde{u}(a)|P((-\infty,a))+|\tilde{u}(b)|P((b,+\infty)))\leq\epsilon/3. (5.67)

Combining (5.64) -(5.67), we obtain (5.61) from (5.57).  

5.2 Multiattribute utility case

The OCE models that we discussed so far are for single attribute decision making. It might be interesting to ask whether the models can be extended to multi-attribute decision making. The answer is yes. Here we present two potential extended models. One is to consider the case that the utility function has an additive structure, that is, the multivariate utility function is the sum of the marginal utility functions of each attribute. Such utility functions are widely used in the literature, see e.g. [28, 1, 2]. In that case, given ξ:ΩIRm\xi:\Omega\rightarrow{\rm I\!R}^{m} and U:IRmIRU:{\rm I\!R}^{m}\rightarrow{\rm I\!R}, we may define the MOCE as

(MMOCEA)Mu(ξ):=supxIRmU(x)+𝔼P[U(ξx)],{\rm(MMOCE-A)}\quad\quad\displaystyle{M_{u}(\xi):=\sup_{x\in{\rm I\!R}^{m}}\;\;U(x)+{\mathbb{E}}_{P}[U(\xi-x)]}, (5.68)

where the multiattribute utility function U(x)=i=1mui(xi)U(x)=\sum_{i=1}^{m}u_{i}(x_{i}) and ui:IRIRu_{i}:{\rm I\!R}\rightarrow{\rm I\!R} is the marginal utility function with respect to the iith attribute. The formulation can be simplified when the probability distribution of ξ\xi is the product of its marginal distributions:

Mu(ξ)=i=1msupxiIR{ui(xi)+𝔼Pi[ui(ξixi)]}.\displaystyle{M_{u}(\xi)=\sum_{i=1}^{m}\sup_{x_{i}\in{\rm I\!R}}\;\;\{u_{i}(x_{i})+{\mathbb{E}}_{P_{i}}[u_{i}(\xi_{i}-x_{i})]\}}. (5.69)

The economic interpretation of the model is that the decision maker might have a portfolio of random assets xix_{i}, i=1,mi=1,\cdots m and the DM would like to cash out xix_{i} from asset ii. The marginal utilities may be the same or different. Problem (5.69) is decomposable as it stands, thus it retains the properties outlined in Section 2 and can be calculated by calculating mm single attribute MOCE simultaneously.

When the utility function is non-additive, we may consider the following model:

(MMOCEB)Mu(ξ):=suptIR+{u(td)+𝔼P[u(ξtd)]},{\rm(MMOCE-B)}\quad\quad\displaystyle{M_{u}(\xi):=\sup_{t\in{\rm I\!R}_{+}}\;\;\{u(td)+{\mathbb{E}}_{P}[u(\xi-td)]\}}, (5.70)

where dd is a fixed vector of weights. In this model, cash to be taken out from the assets is in a prefixed proportion. (MMOCE-B) is essentially a single variate MOCE model. Note that it is possible to further extend model (MMOCE-A) by replacing deterministic vector xx with a random vector XX:

(MMOCEA)Mu(ξ):=supX𝔼[U(X)]+𝔼[U(ξX)].{\rm(MMOCE-A^{\prime})}\quad\quad\displaystyle{M_{u}(\xi):=\sup_{X}\;\;{\mathbb{E}}[U(X)]+{\mathbb{E}}[U(\xi-X)]}. (5.71)

This kind of model has potential applications in finance where a firm detaches risk assets from non-risky assets in order to reduce the systemic risk [47]. In that context, problem (MMOCE-A’) is to find optimal separation XX from the existing overall portfolio of assets ξ\xi. The problem is intrinsically two-stage, one may use linear/polynomial decision rule [5] or K-adapativity method [8] to obtain a (MMOCE-A)-version of approximation. Note also that model (MMOCE-A’) is related to the IDR-based CDE model recently studied by Qi et al. [39] who use OCE for optimizing individualised medical treatment. Since all of the extended models outlined above require much more detailed analysis, we leave them for future research.

6 Quantitative statistical robustness

6.1 Motivation

In Section 3, we discuss in detail how to obtain an approximate solution of (RMOCE) (to ease reading, we repeat the model here):

(RMOCEP)R(P):=maxxXinfu𝒰𝔼P[u(x)+u(ξx)],{\rm(RMOCE-P)}\quad\quad\displaystyle{R(P):=\max_{x\in X}\inf_{u\in\mathcal{U}}{\mathbb{E}}_{P}[u(x)+u(\xi-x)]}, (6.72)

where ξ\xi follows probability distribution PP. A key assumption is that the true probability distribution PP is known and discretely distributed. This assumption may not be satisfied in data-driven problems where the true PP is unknown, and one often uses empirical data to construct an approximation of PP. Even worse is that such data may be contaminated.

Let ξ~1,,ξ~N\tilde{\xi}^{1},...,\tilde{\xi}^{N} denote the contaminated empirical data (we call them perceived data and we use NN to denote the size of samples rather than number of breakpoints without causing confusion henceforth). Let QN:=1Ni=1Nδξ~iQ_{N}:=\frac{1}{N}\sum_{i=1}^{N}\delta_{\tilde{\xi}_{i}} be the empirical distribution constructed with the perceived data, where δξ~i\delta_{\tilde{\xi}_{i}} is the Dirac measure at ξ~i\tilde{\xi}_{i}. We use the perceived data to solve the RMOCE model (assume that the model is solved precisely without computational error):

(RMOCEQN)R(QN):=maxxXinfu𝒰𝔼QN[u(x)+u(ξx)].{\rm(RMOCE-Q_{N})}\quad\quad\displaystyle{R(Q_{N}):=\max_{x\in X}\inf_{u\in\mathcal{U}}{\mathbb{E}}_{Q_{N}}[u(x)+u(\xi-x)]}. (6.73)

We then ask ourselves as to whether R(QN)R(Q_{N}) is a good estimation of R(P)R(P) from statistical point of view. This question is concerned with data perturbation rather than modelling/computational errors as discussed in Section 3.

To proceed the analysis, we introduce another empirical distribution, denoted by PN:=1Ni=1NδξiP_{N}:=\frac{1}{N}\sum_{i=1}^{N}\delta_{\xi_{i}}, which is constructed by the purified perceived data ξ1,,ξN\xi^{1},...,\xi^{N} (the noise in the perceived data is detached, we call them real data henceforth). In practice, it is impossible to detach the noise, we introduce the notion purely for the convenience of statistical analysis. Let R(PN)R(P_{N}) be the optimal value of (RMOCE-P) by replacing PP with PNP_{N}. By the classical law of large numbers, we know that PNPP_{N}\to P and R(PN)R(P)R(P_{N})\to R(P) under moderate conditions. Thus in the literature of stochastic programming, R(PN)R(P_{N}) is called a statistical estimator of R(P)R(P) and here we emphasize that this estimator is based on real data.

Our question is then whether R(QN)R(Q_{N}) is close to R(PN)R(P_{N}) because the former is the only quantity that we are able to obtain. To address this question, we assume the perceived data are iid which means QNQQ_{N}\to Q for some QQ as NN\to\infty. In other words, the perceived data may be viewed as if they are generated by the invisible distribution QQ. Let R(Q)R(Q) denote the optimal value of (RMOCE) with PP being replaced by QQ. We then have

R(QN)R(PN)=R(QN)R(Q)+R(Q)R(P)+R(P)R(PN).R(Q_{N})-R(P_{N})=R(Q_{N})-R(Q)+R(Q)-R(P)+R(P)-R(P_{N}).

Thus if R(QN)R(Q)R(Q_{N})\to R(Q) as NN\to\infty uniformly for all QQ close to PP and R(Q)R(P)R(Q)\to R(P) as QPQ\to P, then R(QN)R(Q_{N}) is close R(PN)R(P_{N}). This explains roughly the motivation of this section. The formal quantitative statistical robust analysis is a bit more complex as we will examine the difference between the probability distributions of R(QN)R(Q_{N}) and R(PN)R(P_{N}) under some metric rather than estimating R(QN)R(PN)R(Q_{N})-R(P_{N}) for each given set of perceived data.

6.2 Statistical analysis

For any two probability measures P,Q𝒫(IR)P,Q\in\mathscr{P}({\rm I\!R}), define the pseudo-metric between PP and QQ by

𝖽𝗅𝒢(P,Q):=supg𝒢|𝔼P[g(ξ)]𝔼Q[g(ξ)]|.\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(P,Q):=\sup_{g\in\mathscr{G}}\bigl{|}{\mathbb{E}}_{P}[g(\xi)]-{\mathbb{E}}_{Q}[g(\xi)]\bigr{|}. (6.74)

It can be seen that 𝖽𝗅𝒢(P,Q)\mathsf{d\kern-0.70007ptl}_{\mathscr{G}}(P,Q) is the maximal difference between the expected values of the class of measurable functions 𝒢\mathscr{G} with respect to PP and QQ. The specific pseudo metrics that we consider in this paper are the Fortet-Mourier metric and the Kantorovich metric. Recall that the pp-th order Fortet-Mourier metric with p1p\geq 1 for P,Q𝒫(IR)P,Q\in\mathscr{P}({\rm I\!R}):

𝖽𝗅FM,p(P,Q):=supg𝒢p(IR)|IRg(ξ)P(dξ)IRg(ξ)Q(dξ)|,\mathsf{d\kern-0.70007ptl}_{FM,p}(P,Q):=\sup_{g\in\mathscr{G}_{p}({\rm I\!R})}\left|\int_{{\rm I\!R}}g(\xi)P(d\xi)-\int_{{\rm I\!R}}g(\xi)Q(d\xi)\right|, (6.75)

where

𝒢p(IR):={g:IRIR|g(ξ)g(ξ~)|cp(ξ,ξ~)ξξ~,ξ,ξ~IR}\mathscr{G}_{p}({\rm I\!R}):=\{g:{\rm I\!R}\rightarrow{\rm I\!R}\mid|g(\xi)-g(\tilde{\xi})|\leq c_{p}(\xi,\tilde{\xi})\|\xi-\tilde{\xi}\|,\forall\xi,\tilde{\xi}\in{\rm I\!R}\}

and

cp(ξ,ξ~):=max{1,ξ,ξ~}p1,ξ,ξ~IR.c_{p}(\xi,\tilde{\xi}):=\max\{1,\|\xi\|,\|\tilde{\xi}\|\}^{p-1},\quad\forall\xi,\tilde{\xi}\in{\rm I\!R}.

When p=1p=1, the functions in 𝒢p(IR)\mathscr{G}_{p}({\rm I\!R}) are globally Lipschitz continuous with modulus 11 and 𝒢p(IR)\mathscr{G}_{p}({\rm I\!R}) coincides with 𝒢K\mathscr{G}_{K} in (3.25). Thus 𝖽𝗅FM,1(P,Q)=𝖽𝗅K(P,Q)\mathsf{d\kern-0.70007ptl}_{FM,1}(P,Q)=\mathsf{d\kern-0.70007ptl}_{K}(P,Q). For more details, see [16, 42, 48].

To get the statistical robustness result, let IRN{\rm I\!R}^{\otimes N} and (IR)N\mathcal{B}({\rm I\!R})^{\otimes N} denote the Cartesian product IR××IR{\rm I\!R}\times\cdot\cdot\cdot\times{\rm I\!R} and its Borel sigma algebra. Let PNP^{\otimes N} denote the probability measure on the measurable space (IRN,(IR)N)({\rm I\!R}^{\otimes N},\mathcal{B}({\rm I\!R})^{\otimes N}) with marginal PP on each (IR,(IR))({\rm I\!R},\mathcal{B}({\rm I\!R})) and QNQ^{\otimes N} with marginal QQ. Now we can state the definition of statistical robustness of a statistic estimator, which is proposed in [18, 48].

Definition 6.1 (Quantitative statistical robustness)

Let 𝒫(IR){\cal M}\subset\mathscr{P}({\rm I\!R}) be a set of probability measures. A sequence of statistical estimators T^N\hat{T}_{N} is said to be quantitatively statistically robust on \mathcal{M} w.r.t. (𝖽𝗅K,𝖽𝗅FM,p)(\mathsf{d\kern-0.70007ptl}_{K},\mathsf{d\kern-0.70007ptl}_{FM,p}) if there exists a positive constant CC such that for all NN

𝖽𝗅K(PNT^N1,QNT^N1)C𝖽𝗅FM,p(P,Q)<+,P,Q,\displaystyle\mathsf{d\kern-0.70007ptl}_{K}(P^{\otimes N}\circ\hat{T}_{N}^{-1},Q^{\otimes N}\circ\hat{T}_{N}^{-1})\leq C\mathsf{d\kern-0.70007ptl}_{FM,p}(P,Q)<+\infty,\;\forall P,Q\in\mathcal{M}, (6.76)

where 𝖽𝗅K\mathsf{d\kern-0.70007ptl}_{K} is the Kantorovich metric on 𝒫(IR)\mathscr{P}({\rm I\!R}) and 𝖽𝗅FM,p\mathsf{d\kern-0.70007ptl}_{FM,p} is the Fortet-Mourier metric on 𝒫(IR)\mathscr{P}({\rm I\!R}).

Here PNT^N1P^{\otimes N}\circ\hat{T}_{N}^{-1} and QNT^N1Q^{\otimes N}\circ\hat{T}_{N}^{-1} are probability measures/distributions on IR{\rm I\!R}. The next theorem states quantitative statistical robustness of R^N:=R(QN)\hat{R}_{N}:=R(Q_{N}).

Theorem 6.1

Assume: (a) There exists a positive constant L>0L>0 such that for all xXx\in X and u𝒰u\in\mathcal{U},

|u(ξx)u(ξx)|Lmax{1,|ξ|,|ξ|}p1|ξξ|,|u(\xi-x)-u(\xi^{\prime}-x)|\leq L\max\left\{1,|\xi|,|\xi^{\prime}|\right\}^{p-1}|\xi-\xi^{\prime}|,

(b) set 𝒰\mathcal{U} is chosen such that ψ(t):=supu𝒰|u(t)|\psi(t):=\sup_{u\in\mathcal{U}}|u(t)| is a gauge function, that is, ψ:IR[0,)\psi:{\rm I\!R}\rightarrow[0,\infty) is continuous and ψ1\psi\geq 1 holds outside a compact set. Then for any NN\in\mathbb{N},

𝖽𝗅K(PNR^N1,QNR^N1)L𝖽𝗅FM,p(P,Q),P,Qϕ,\displaystyle{\mathsf{d\kern-0.70007ptl}_{K}(P^{\otimes N}\circ\hat{R}_{N}^{-1},Q^{\otimes N}\circ\hat{R}_{N}^{-1})\leq L\mathsf{d\kern-0.70007ptl}_{FM,p}(P,Q),\forall P,Q\in\mathcal{M}^{\phi}}, (6.77)

where ϕ(ξ):=C0+L(|ξ|+|ξ|p)\phi(\xi):=C_{0}+L(|\xi|+|\xi|^{p}) for some constant C0>1C_{0}>1.

Proof. By definition

𝖽𝗅K(PNR^N1,QNR^N1)\displaystyle\mathsf{d\kern-0.70007ptl}_{K}(P^{\otimes N}\circ\hat{R}_{N}^{-1},Q^{\otimes N}\circ\hat{R}_{N}^{-1})
=supg𝒢K|IRg(t)PNR^N1(dt)IRg(t)QNR^N1(dt)|\displaystyle=\sup_{g\in\mathscr{G}_{K}}\left|\int_{{\rm I\!R}}g(t)P^{\otimes N}\circ\hat{R}_{N}^{-1}(dt)-\int_{{\rm I\!R}}g(t)Q^{\otimes N}\circ\hat{R}_{N}^{-1}(dt)\right|
=supg𝒢K|IRNg(R^(𝝃N))PN(d𝝃N)IRNg(R^(𝝃N))QN(d𝝃N)|,\displaystyle=\sup_{g\in\mathscr{G}_{K}}\left|\int_{{\rm I\!R}^{\otimes N}}g(\hat{R}(\boldsymbol{\xi}^{N}))P^{\otimes N}(d\boldsymbol{\xi}^{N})-\int_{{\rm I\!R}^{\otimes N}}g(\hat{R}(\boldsymbol{\xi}^{N}))Q^{\otimes N}(d\boldsymbol{\xi}^{N})\right|, (6.78)

where 𝝃N=(ξ1,,ξN)\boldsymbol{\xi}^{N}=(\xi^{1},...,\xi^{N}) and we write R^(𝝃N)\hat{R}(\boldsymbol{\xi}^{N}) for R^N\hat{R}_{N}. To see the well-definedness of the pseudo-metric, notice that for every g𝒢Kg\in\mathscr{G}_{K} and a fixed 𝝃0NIRN\boldsymbol{\xi}_{0}^{N}\in{\rm I\!R}^{\otimes N}

|g(R^(𝝃N))||g(R^(𝝃0N))|+|R^(𝝃N)R^(𝝃0N)|,|g(\hat{R}(\boldsymbol{\xi}^{N}))|\leq|g(\hat{R}(\boldsymbol{\xi}_{0}^{N}))|+|\hat{R}(\boldsymbol{\xi}^{N})-\hat{R}(\boldsymbol{\xi}_{0}^{N})|, (6.79)

where 𝝃0NIRN\boldsymbol{\xi}_{0}^{N}\in{\rm I\!R}^{\otimes N} is fixed. From condition (b) and nondecreasing property of uu, there exists a positive number C0C_{0} such that

supu𝒰,xX|u(x)+u(ξx)|C0+L(|ξ|+|ξ|p),ξIR.\sup_{u\in\mathcal{U},x\in X}|u(x)+u(\xi-x)|\leq C_{0}+L(|\xi|+|\xi|^{p}),\forall\xi\in{\rm I\!R}. (6.80)

By the definition of R^(𝝃N)\hat{R}(\boldsymbol{\xi}^{N}), it follows that

|R^(𝝃N)|=|maxxXinfu𝒰1Nk=1N[u(x)+u(ξkx)]|1Nk=1Nϕ(ξk).|\hat{R}(\boldsymbol{\xi}^{N})|=\left|\max_{x\in X}\inf_{u\in\mathcal{U}}\frac{1}{N}\sum_{k=1}^{N}[u(x)+u(\xi^{k}-x)]\right|\leq\frac{1}{N}\sum_{k=1}^{N}\phi(\xi^{k}).

Moreover,

IRN|R^(𝝃N)|PN(d𝝃N)\displaystyle\int_{{\rm I\!R}^{\otimes N}}|\hat{R}(\boldsymbol{\xi}^{N})|P^{\otimes N}(d\boldsymbol{\xi}^{N}) \displaystyle\leq IRN1Nk=1Nϕ(ξk)PN(d𝝃N)\displaystyle\int_{{\rm I\!R}^{\otimes N}}\frac{1}{N}\sum_{k=1}^{N}\phi(\xi^{k})P^{\otimes N}(d\boldsymbol{\xi}^{N}) (6.81)
=\displaystyle= IRϕ(ξ)P(dξ)<,Pϕ,\displaystyle\int_{{\rm I\!R}}\phi(\xi)P(d\xi)<\infty,\forall P\in\mathcal{M}^{\phi},

where the equality holds due to the fact that ξ1,,ξN\xi^{1},...,\xi^{N} are independent and identically distributed. Combining (6.79) and (6.81) we can obtain

IRNg(R^(𝝃N))PN(d𝝃N)<,Pϕ.\int_{{\rm I\!R}^{\otimes N}}g(\hat{R}(\boldsymbol{\xi}^{N}))P^{\otimes N}(d\boldsymbol{\xi}^{N})<\infty,\forall P\in\mathcal{M}^{\phi}.

Similar argument can be made on IRNg(R^(𝝃N))QN(d𝝃N)\int_{{\rm I\!R}^{\otimes N}}g(\hat{R}(\boldsymbol{\xi}^{N}))Q^{\otimes N}(d\boldsymbol{\xi}^{N}) for any QϕQ\in\mathcal{M}^{\phi}. Next, for any P,QϕP,Q\in\mathcal{M}^{\phi},

|R(P)R(Q)|\displaystyle|R(P)-R(Q)| =\displaystyle= |supxXinfu𝒰{u(x)+𝔼P[u(ξx)]}supxXinfu𝒰{u(x)+𝔼Q[u(ξx)]}|\displaystyle\left|\sup_{x\in X}\inf_{u\in\mathcal{U}}\{u(x)+{\mathbb{E}}_{P}[u(\xi-x)]\}-\sup_{x\in X}\inf_{u\in\mathcal{U}}\{u(x)+{\mathbb{E}}_{Q}[u(\xi-x)]\}\right|
\displaystyle\leq supxXsupu𝒰|𝔼P[u(ξx)]𝔼Q[u(ξx)]|\displaystyle\sup_{x\in X}\sup_{u\in\mathcal{U}}|{\mathbb{E}}_{P}[u(\xi-x)]-{\mathbb{E}}_{Q}[u(\xi-x)]|
\displaystyle\leq 𝖽𝗅FM,p(P,Q),\displaystyle\mathsf{d\kern-0.70007ptl}_{FM,p}(P,Q),

where the last inequality follows from condition (a). Then we can obtain

|g(R^(ξ~1,,ξ~N)g(R^(ξ^1,,ξ^N))||R^(ξ~1,,ξ~N)R~(ξ^1,,ξ^N)|1Nk=1NsupxX,u𝒰|u(ξ~kx)u(ξ^kx)|LNk=1Nmax{1+|ξ~k|+|ξ^k|}p1|ξ~kξ^k|.|g(\hat{R}(\tilde{\xi}^{1},...,\tilde{\xi}^{N})-g(\hat{R}(\hat{\xi}^{1},...,\hat{\xi}^{N}))|\leq|\hat{R}(\tilde{\xi}^{1},...,\tilde{\xi}^{N})-\tilde{R}(\hat{\xi}^{1},...,\hat{\xi}^{N})|\\ \leq\frac{1}{N}\sum_{k=1}^{N}\sup_{x\in X,u\in\mathcal{U}}|u(\tilde{\xi}^{k}-x)-u(\hat{\xi}^{k}-x)|\leq\frac{L}{N}\sum_{k=1}^{N}\max\{1+|\tilde{\xi}^{k}|+|\hat{\xi}^{k}|\}^{p-1}|\tilde{\xi}^{k}-\hat{\xi}^{k}|. (6.82)

It follows by [48, Lemma 4.4] that

(6.78)𝖽𝗅FM,p(PN,QN)L𝖽𝗅FM,p(P,Q)\displaystyle(\ref{eq:Kantorovich-distance})\leq\mathsf{d\kern-0.70007ptl}_{FM,p}(P^{\otimes N},Q^{\otimes N})\leq L\mathsf{d\kern-0.70007ptl}_{FM,p}(P,Q) (6.83)

and hence inequality (6.77).  

7 Numerical tests

We have carried out some tests on the numerical schemes for computing RMOCE. In this section, we report the preliminary numerical results.

The first set of tests are about the comparison between the MOCE model (1.3) and OCE model (1.1) in terms of the optimal values and the optimal solutions. We do so by considering ξ\xi following some specific distributions including uniform, Gamma, lognormal and normalized Pareto distribution. The second set of tests are on the RMOCE model and numerical schemes proposed in Section 3. We investigate how the optimal value and the worst case utility function in the RMOCE model change as the radius of ambiguity set and the number of breakpoints vary. We use the parallel particle swarm optimization method [29, 34] to solve problem (3.29) and CVX solver to solve inner minimization problem (3.33). All the tests are carried out in Matlab R2021a installed on a PC (16GB RAM, CPU 2.3 GHz) with Intel Core i7 processor.

Throughout the section we restrict 𝒰\mathscr{U} to a set of all increasing concave utility functions mapping from a compact interval [a,b][a,b]. We take [a,b][a,b] as the domain of uu which is the union of ranges of xx and ξx\xi-x for x[ξmin/2,ξmax/2]x\in[\xi_{\min}/2,\xi_{\max}/2] by 3.2 because the number NN of breakpoints can guarantee that βNξmaxξmin\beta_{N}\leq\xi_{\max}-\xi_{\min}. We generate iid samples ξ1,,ξK\xi^{1},...,\xi^{K} for random variable ξ\xi with equal probabilities pk=1/Kp_{k}=1/K for k=1,,Kk=1,...,K.

In the first set of tests of OCE and MOCE, we set the nomial utility function as u0(t)=(1e2t)/2u^{0}(t)=(1-e^{-2t})/2. Table 1 displays the optimal values and the optimal solutions as well as the CPU times. The 33th and 44th columns present the optimal values of OCE and MOCE model, and the 55th and 66th columns present the optimal solutions of OCE and MOCE model, respectively. As we can see, the OCE values are consistently larger than the MOCE values, this is because u(t)<tu(t)<t. Moreover, we find the optimal solutions of MOCE problem (under xx^{*}) fall within [ξmin/2,ξmax/2][\xi_{\min}/2,\xi_{\max}/2] although we have not displayed the intervals due to the limitation of space. This complies with 2.1.

Table 1: Numerical results of MOCE
Distribution K Mu(ξ)M_{u}(\xi) Su(ξ)S_{u}(\xi) xx^{*} η\eta^{*} CPU time
Uniform (-1,1) 10 -0.5590 -0.4440 -0.2220 -0.4441 0.8700
100 -0.1950 -0.1782 -0.0891 -0.1782 0.9914
1000 -0.3508 -0.3008 -0.1504 -0.3008 3.8415
Lognormal (0,1) 10 0.4929 0.6792 0.3396 0.6792 0.4279
100 0.5182 0.7303 0.3651 0.7303 0.8139
1000 0.5313 0.7578 0.3789 0.7578 3.7001
Pareto (1,1.5) 10 0.8692 2.0337 1.0169 2.0337 0.4484
100 0.8990 2.2926 1.1463 2.2925 0.7263
1000 0.8942 2.2461 1.1231 2.2461 3.6693
Gamma (0.53,3) 10 0.3392 0.4143 0.2072 0.4143 0.5002
100 0.4415 0.5824 0.2912 0.5825 0.7094
1000 0.4088 0.5255 0.2628 0.5255 3.7729

In the second set of tests about RMOCE, we set the nominal utility as u0(t)=(1eα)/2u^{0}(t)=(1-e^{-\alpha})/2 where αIR+\alpha\in{\rm I\!R}_{+} is a parameter which determines the degree of concavity of the utility function. The number of random samples is fixed at K=100K=100 for the uniform distribution and K=10K=10 for Gamma, lognormal and normalized Pareto distribution. The parameters of the tests are listed in Table 3, the 4th column represents the Lipschitz modulus of utility functions. For the cases where the random samples are generated by uniform distribution, Figures 3 and 3 visualize the worst case utility functions and the optimal values as the radius decreases. Figure 3 visualizes the change of optimal values as the number of breakpoints increases. It can be seen that the number of breakpoints has little effect on the optimal value. For the cases when ξ\xi follows Gamma, lognormal and normalized Pareto distribution, Figures 4 and 5 visualize the changes of the worst case utility functions and the optimal values as the radius decreases. We can see that the worst utility function moves closer to the nominal utility function as the radius of the ambiguity set decreases to zero, the optimal value increases as the radius decreases. This is because the Kantorovich ball becomes smaller when the radius decreases. In the case that r=0r=0, the worst case utility function is the piecewise linear approximation of the nominal utility function. The error bound of the optimal value is also depicted in Figures 3 and 5, note that the error bound is getting smaller when the number of breakpoints increases in Figure 3. Table 3 provides the optimal values and running time for different number of breakpoints.

Table 2: Parameters of RMOCE tests
Distribution α\alpha N L
Uniform (-1,1) 2 10 30
Lognormal (0,1) 1/2 300 10
Pareto (1,1.5) 1/3 300 10
Gamma (0.53,3) 1/2 300 10
Table 3: Running time for different number of breakpoints
N Optimal value CPU time
20 -108.4846 20.7796
40 -108.6524 24.1097
60 -108.5553 31.4506
80 -108.5657 36.7656
100 -108.5648 41.9548
\setcaptionwidth

1.5in [Uncaptioned image] Figure 1: Worst case utility functions with different r [Uncaptioned image] Figure 2: Optimal values with different r [Uncaptioned image] Figure 3: Optimal values with different N

lognormal (0,1)
Refer to caption
Pareto (1,1.5)
Refer to caption
Gamma (0.53,3)
Refer to caption
Figure 4: Worst case utility functions with heavy-tailed distributions
lognormal (0,1)
Refer to caption
Pareto (1,1.5)
Refer to caption
Gamma (0.53,3)
Refer to caption
Figure 5: Optimal values with heavy-tailed distributions

8 Conclusion

In this paper we explore variations of the concept of optimized certainty equivalent with a number of new inputs. First, we propose a modified optimized certainty equivalent (MOCE) model by considering the utility of present consumption. The optimal strategy (which balances the present and future consumption) is uniquely determined by the decision maker’s risk preference rather than by his/her utility representations (which is not unique). The resulting MOCE value is positive homogeneous in uu. The MOCE is also in alignment with the consumption models in economics. Second, there is a distinction between OCE and MOCE in terms of the utility functions to be used in the model. In the classical OCE model, it requires the utility function to satisfy u(0)=0u(0)=0 and 1u(0)1\in\partial u(0). The new MOCE model does not require these conditions. Third, we propose a preference robust version of the new MOCE model for the case that the decision maker’s true utility function is ambiguous. Ambiguity does exist in practice and this paper provides a comprehensive treatment of the preference robust MOCE model from modelling to computational scheme and underlying theory. Fourth, in the case that the proposed RMOCE model is applied to data-driven problems where the underlying exogenous data (samples of ξ\xi) are potentially contaminated, we derive sufficient conditions under which the RMOCE calculated with the data is statistically robust. Fifth, we outline potential extensions of the MOCE model from single decision making to multi-attribute decision making and point out potential applications in asset re-organization. In summary, this paper provides a new outlook of OCE in both modelling and analysis, which complement the existing research in the literature.

References

  • [1] A. E. Abbas. Multiattribute utility copulas. Operations Research, 57(6):1367–1383, 2009.
  • [2] A. E. Abbas and Z. Sun. Multiattribute utility functions satisfying mutual preferential independence. Operations Research, 63(2):378–393, 2015.
  • [3] B. Armbruster and E. Delage. Decision making under uncertainty when preference information is incomplete. Management science, 61(1):111–128, 2015.
  • [4] R. J. Aumann. Integrals of set-valued functions. Journal of mathematical analysis and applications, 12(1):1–12, 1965.
  • [5] D. Bampou and D. Kuhn. Scenario-free stochastic programming with polynomial decision rules. In 2011 50th IEEE Conference on Decision and Control and European Control Conference, pages 7806–7812. IEEE, 2011.
  • [6] A. Ben-Tal and M. Teboulle. Expected utility, penalty functions, and duality in stochastic nonlinear programming. Management Science, 32(11):1445–1466, 1986.
  • [7] A. Ben‐Tal and M. Teboulle. An old‐new concept of convex risk measures: The optimized certainty equivalent. Mathematical Finance, 17(3):449–476, 2007.
  • [8] D. Bertsimas and C. Caramanis. Adaptability via sampling. In 2007 46th IEEE Conference on Decision and Control, pages 4717–4722. IEEE, 2007.
  • [9] J. V. Burke, X. Chen, and H. Sun. The subdifferential of measurable composite max integrands and smoothing approximation. Mathematical Programming, 181(2):229–264, 2020.
  • [10] S. Cerreia-Vioglio, D. Dillenberger, and P. Ortoleva. Cautious expected utility and the certainty effect. Econometrica, 83(2):693–728, 2015.
  • [11] F. H. Clarke. Optimization and Nonsmooth Analysis. SIAM, 1990.
  • [12] J. H. Cochrane. Asset pricing: Revised edition. Princeton university press, 2009.
  • [13] R. Cont, R. Deguest, and G. Scandolo. Robustness and sensitivity analysis of risk measurement procedures. Quantitative finance, 10(6):593–606, 2010.
  • [14] E. Delage, S. Guo, and H. Xu. Shortfall risk models when information on loss function is incomplete. Operations Research, 2022.
  • [15] P. H. Farquhar. State of the art-utility assessment methods. Management science, 30(11):1283–1300, 1984.
  • [16] A. L. Gibbs and F. E. Su. On choosing and bounding probability metrics. International statistical review, 70(3):419–435, 2002.
  • [17] S. Guo and H. Xu. Statistical robustness in utility preference robust optimization models. Mathematical Programming, pages 1–42, 2021.
  • [18] S. Guo and H. Xu. Utility preference robust optimization with moment-type information structure. Optimization online, 2021.
  • [19] N. H. Hakansson. Optimal investment and consumption strategies under risk for a class of utility functions, pages 525–545. Elsevier, 1975.
  • [20] F. R. Hampel. A general qualitative definition of robustness. The annals of mathematical statistics, 42(6):1887–1896, 1971.
  • [21] W. Haskell, H. Xu, and W. Huang. Preference robust optimization for choice functions on the space of cdfs. SIAM Journal on Optimization, 2022.
  • [22] W. B. Haskell, W. Huang, and H. Xu. Preference elicitation and robust optimization with multi-attribute quasi-concave choice functions. arXiv preprint arXiv:1805.06632, 2018.
  • [23] J. Hu, M. Bansal, and S. Mehrotra. Robust decision making using a general utility set. European Journal of Operational Research, 269(2):699–714, 2018.
  • [24] J. Hu and S. Mehrotra. Robust decision making over a set of random targets or risk-averse utilities with an application to portfolio optimization. IIE Transactions, 47(4):358–372, 2015.
  • [25] J. Hu and G. Stepanyan. Optimization with reference-based robust preference constraints. SIAM Journal on Optimization, 27(4):2230–2257, 2017.
  • [26] P. J. Huber and E.M. Ronchetti. Robust Statistics. Wiley, Hoboken, 2nd edition edition, 2009.
  • [27] U. S. Karmarkar. Subjectively weighted utility: A descriptive extension of the expected utility model. Organizational behavior and human performance, 21(1):61–72, 1978.
  • [28] R. L. Keeney, H. Raiffa, and R. F. Meyer. Decisions with multiple objectives: preferences and value trade-offs. Cambridge university press, 1993.
  • [29] J. Kennedy and R. Eberhart. Particle swarm optimization. In Proceedings of ICNN’95-international conference on neural networks, volume 4, pages 1942–1948. IEEE, 1995.
  • [30] V. Krätschmer, A. Schied, and H. Zähle. Qualitative and infinitesimal robustness of tail-dependent statistical functionals. Journal of Multivariate Analysis, 103(1):35–47, 2012.
  • [31] J. Y. Li. Inverse optimization of convex risk functions. Management Science, 67(11):7113–7141, 2021.
  • [32] J. Liu, Z. Chen, and H. Xu. Multistage utility preference robust optimization. arXiv preprint arXiv:2109.04789, 2021.
  • [33] F. Maccheroni. Maxmin under risk. Economic Theory, 19(4):823–831, 2002.
  • [34] E. Mezura-Montes and C. A. C. Coello. Constraint-handling in nature-inspired numerical optimization: past, present and future. Swarm and Evolutionary Computation, 1(4):173–194, 2011.
  • [35] O. Morgenstern and J. Von Neumann. Theory of games and economic behavior. Princeton university press, 1953.
  • [36] M. Nouiehed, J. Pang, and M. Razaviyayn. On the pervasiveness of difference-convexity in optimization and statistics. Mathematical Programming, 174(1):195–222, 2019.
  • [37] W. Ogryczak and A. Ruszczynski. Dual stochastic dominance and related mean-risk models. SIAM Journal on Optimization, 13(1):60–78, 2002.
  • [38] G. C. Pflug and W. Römisch. Modeling, measuring and managing risk. World Scientific, 2007.
  • [39] Z. Qi, Y. Cui, Y. Liu, and J. Pang. Estimation of individualized decision rules based on an optimized covariate-dependent equivalent of random outcomes. SIAM Journal on Optimization, 29(3):2337–2362, 2019.
  • [40] R. T. Rockafellar. Convex analysis, volume 36. Princeton university press, 1970.
  • [41] R. T. Rockafellar and S. Uryasev. Optimization of conditional value-at-risk. Journal of risk, 2:21–42, 2000.
  • [42] W. Römisch. Stability of stochastic programming problems. Handbooks in operations research and management science, 10:483–554, 2003.
  • [43] D. Sauré and J. P. Vielma. Ellipsoidal methods for adaptive choice-based conjoint analysis. Operations Research, 67(2):315–338, 2019.
  • [44] L. L. Thurstone. A law of comparative judgment. Psychological review, 34(4):273, 1927.
  • [45] K. E. Train. Discrete choice methods with simulation. Cambridge university press, 2009.
  • [46] W. Wang and H. Xu. Robust spectral risk optimization when information on risk spectrum is incomplete. SIAM Journal on Optimization, 30(4):3198–3229, 2020.
  • [47] W. Wang, H. Xu, and T. Ma. Optimal scenario-dependent multivariate shortfall risk measure and its application in capital allocation. Available at SSRN 3849125, 2021.
  • [48] W. Wang, H. Xu, and T. Ma. Quantitative statistical robustness for tail-dependent law invariant risk measures. Quantitative Finance, pages 1–17, 2021.
  • [49] M. Weber. Decision making with incomplete information. European journal of operational research, 28(1):44–57, 1987.
  • [50] W. Wiesemann, D. Kuhn, and M. Sim. Distributionally robust convex optimization. Operations Research, 62(6):1358–1376, 2014.
  • [51] H. Xu and S. Zhang. Quantitative statistical robustness in distributionally robust optimization models. Pacific Journal of Optimization Special Issue, 2021.
  • [52] S. Zhang and H. Xu. Preference robust generalized shortfall risk measure based on the cumulative prospect theory when the value function and weighting functions are ambiguous. arXiv preprint arXiv:2112.10142, 2021.