multistage optimization of option portfolio using higher order coherent risk measures

9
Innovative Applications of O.R. Multistage optimization of option portfolio using higher order coherent risk measures Yassine Matmoura a , Spiridon Penev b,a Ecole Centrale, Paris, France b The University of New South Wales, Sydney, Australia article info Article history: Received 6 January 2012 Accepted 11 December 2012 Available online 21 December 2012 Keywords: Coherent risk measures Duality Average value-at-risk Monte Carlo simulation Kusuoka measure Stochastic programming abstract Choosing a suitable risk measure to optimize an option portfolio’s performance represents a significant challenge. This paper is concerned with illustrating the advantages of Higher order coherent risk mea- sures to evaluate option risk’s evolution. It discusses the detailed implementation of the resulting dynamic risk optimization problem using stochastic programming. We propose an algorithmic procedure to optimize an option portfolio based on minimization of conditional higher order coherent risk mea- sures. Illustrative examples demonstrate some advantages in the performance of the portfolio’s levels when higher order coherent risk measures are used in the risk optimization criterion. Crown Copyright Ó 2012 Published by Elsevier B.V. All rights reserved. 1. Introduction Portfolio optimization is the process of selection of wealth allo- cation to financial instruments held to meet certain pre-defined criteria. It is prudent to take risk into account when evaluating the optimal portfolio. Risk assessment and evaluation is in fact crit- ical. The uncertainty in risk assessment is expressed mathemati- cally using a random variable, a random vector or a random process and the risk is a functional (numerical surrogate) related to it. Choosing the appropriate numerical surrogate is what choos- ing the suitable (optimal) risk measure is about (Rockafellar et al., 2008). In order to apply risk assessment methodology in practice, the inference aspect using data has to be addressed. In the context of the risk measures of relevance to financial applications, this as- pect is still only vaguely discussed, especially in dynamic settings that describe the evolution of financial data in time. The relatively slow progress on the front of inference about dynamic stochastic risk could perhaps be explained by the fact that the resulting infer- ence problems are cast as difficult and non-standard optimization problems under constraints. These constraints make the model better correspond to reality but make the optimization challenging (Shapiro et al., 2009). Further difficulties arise when trying to con- struct confidence regions for the risk. It is known how to proceed when the estimator of the risk measure results from an uncon- strained optimization of a certain function like in the case of para- metric maximum likelihood estimators, M-estimators and the like. However, when we are dealing with non-explicitly defined estima- tors resulting from an involved constrained optimization problem, a new methodology about the asymptotic distribution of these estimators is required. Suppose we want to invest an amount A 0 in n assets. Denote by x i , i = 1, 2, ..., n, the portion of capital to be invested in each asset. Suppose, further, each asset has a return rate R i , i = 1, 2, ..., n. Of course, the return rates are unknown in advance and can be con- sidered to be random. A decision is to be made on the choice of x i , i = 1, 2, ..., n based on our beliefs about R i , i = 1, 2, ..., n. Sponta- neously, an investor can choose wealth allocation by maximizing the expected return rate. Such decision leads to concentrating the investment in the assets with the highest expected returns. This solution is very simplistic since it does not take into account unfavorable events. It leaves out all considerations regarding risk of losing if most profitable assets (defined as large expected return rate) start following a dropping path. The pressing question to be answered first is how should the risk be defined. Recently, there have been attempts to formulate universally acceptable conditions of consistency that a risk mea- sure should satisfy. The pioneering paper Artzner et al. (1999) has introduced the concept of a coherent risk measure which is meant to satisfy a set of properties. These are translation invari- ance, sub-additivity, positive homogeneity, and monotonicity. Later, Föllmer and Schied (2002) introduced a more general class of risk measures, called convex risk measures. In their definition, the sub-additivity has been replaced by the convexity property. 0377-2217/$ - see front matter Crown Copyright Ó 2012 Published by Elsevier B.V. All rights reserved. http://dx.doi.org/10.1016/j.ejor.2012.12.013 Corresponding author. Address: Department of Statistics, School of Mathemat- ics and Statistics, 2052 Sydney, NSW, Australia. Tel.: +61 (0)2 9385 7023; fax: +61 (0)2 9385 7123. E-mail address: [email protected] (S. Penev). European Journal of Operational Research 227 (2013) 190–198 Contents lists available at SciVerse ScienceDirect European Journal of Operational Research journal homepage: www.elsevier.com/locate/ejor

Upload: spiridon

Post on 03-Dec-2016

223 views

Category:

Documents


7 download

TRANSCRIPT

Page 1: Multistage optimization of option portfolio using higher order coherent risk measures

European Journal of Operational Research 227 (2013) 190–198

Contents lists available at SciVerse ScienceDirect

European Journal of Operational Research

journal homepage: www.elsevier .com/locate /e jor

Innovative Applications of O.R.

Multistage optimization of option portfolio using higher order coherentrisk measures

Yassine Matmoura a, Spiridon Penev b,⇑a Ecole Centrale, Paris, Franceb The University of New South Wales, Sydney, Australia

a r t i c l e i n f o a b s t r a c t

Article history:Received 6 January 2012Accepted 11 December 2012Available online 21 December 2012

Keywords:Coherent risk measuresDualityAverage value-at-riskMonte Carlo simulationKusuoka measureStochastic programming

0377-2217/$ - see front matter Crown Copyright � 2http://dx.doi.org/10.1016/j.ejor.2012.12.013

⇑ Corresponding author. Address: Department of Stics and Statistics, 2052 Sydney, NSW, Australia. Tel.:(0)2 9385 7123.

E-mail address: [email protected] (S. Penev).

Choosing a suitable risk measure to optimize an option portfolio’s performance represents a significantchallenge. This paper is concerned with illustrating the advantages of Higher order coherent risk mea-sures to evaluate option risk’s evolution. It discusses the detailed implementation of the resultingdynamic risk optimization problem using stochastic programming. We propose an algorithmic procedureto optimize an option portfolio based on minimization of conditional higher order coherent risk mea-sures. Illustrative examples demonstrate some advantages in the performance of the portfolio’s levelswhen higher order coherent risk measures are used in the risk optimization criterion.

Crown Copyright � 2012 Published by Elsevier B.V. All rights reserved.

1. Introduction strained optimization of a certain function like in the case of para-

Portfolio optimization is the process of selection of wealth allo-cation to financial instruments held to meet certain pre-definedcriteria. It is prudent to take risk into account when evaluatingthe optimal portfolio. Risk assessment and evaluation is in fact crit-ical. The uncertainty in risk assessment is expressed mathemati-cally using a random variable, a random vector or a randomprocess and the risk is a functional (numerical surrogate) relatedto it. Choosing the appropriate numerical surrogate is what choos-ing the suitable (optimal) risk measure is about (Rockafellar et al.,2008). In order to apply risk assessment methodology in practice,the inference aspect using data has to be addressed. In the contextof the risk measures of relevance to financial applications, this as-pect is still only vaguely discussed, especially in dynamic settingsthat describe the evolution of financial data in time. The relativelyslow progress on the front of inference about dynamic stochasticrisk could perhaps be explained by the fact that the resulting infer-ence problems are cast as difficult and non-standard optimizationproblems under constraints. These constraints make the modelbetter correspond to reality but make the optimization challenging(Shapiro et al., 2009). Further difficulties arise when trying to con-struct confidence regions for the risk. It is known how to proceedwhen the estimator of the risk measure results from an uncon-

012 Published by Elsevier B.V. All

atistics, School of Mathemat-+61 (0)2 9385 7023; fax: +61

metric maximum likelihood estimators, M-estimators and the like.However, when we are dealing with non-explicitly defined estima-tors resulting from an involved constrained optimization problem,a new methodology about the asymptotic distribution of theseestimators is required.

Suppose we want to invest an amount A0 in n assets. Denote byxi, i = 1, 2, . . ., n, the portion of capital to be invested in each asset.Suppose, further, each asset has a return rate Ri, i = 1, 2, . . ., n. Ofcourse, the return rates are unknown in advance and can be con-sidered to be random. A decision is to be made on the choice ofxi, i = 1, 2, . . ., n based on our beliefs about Ri, i = 1, 2, . . ., n. Sponta-neously, an investor can choose wealth allocation by maximizingthe expected return rate. Such decision leads to concentratingthe investment in the assets with the highest expected returns.This solution is very simplistic since it does not take into accountunfavorable events. It leaves out all considerations regarding riskof losing if most profitable assets (defined as large expected returnrate) start following a dropping path.

The pressing question to be answered first is how should therisk be defined. Recently, there have been attempts to formulateuniversally acceptable conditions of consistency that a risk mea-sure should satisfy. The pioneering paper Artzner et al. (1999)has introduced the concept of a coherent risk measure which ismeant to satisfy a set of properties. These are translation invari-ance, sub-additivity, positive homogeneity, and monotonicity.Later, Föllmer and Schied (2002) introduced a more general classof risk measures, called convex risk measures. In their definition,the sub-additivity has been replaced by the convexity property.

rights reserved.

Page 2: Multistage optimization of option portfolio using higher order coherent risk measures

Y. Matmoura, S. Penev / European Journal of Operational Research 227 (2013) 190–198 191

There is only a slight difference between the two definitions andvery often, as will be also the case in this article, the convex riskmeasures will be used as a counterpart of coherent risk measures.Formally, for random variables X, Y, the coherency of the risk g(with values on the real axis) requires:

� convexity g(kX + (1 � k)Y) 6 kg(X) + (1 � k)g(Y), k 2 [0,1].� monotonicity: if X 6 Y then g(X) P g(Y).� positive homogeneity: if k P 0 then g(kX) = kg(X).� translation invariance: if m 2 R then g(Y + m) = g(Y) �m.

Here the function g(X) represents a ‘‘safe equivalent’’ that off-sets the random variable and is designed to highlight most impor-tant risk information in X.

The convexity reflects the widely held view that diversificationdoes not increase risk. Consequently, new convex coherent mea-sures were introduced.

This line of research leads to new mathematical problems inconvex analysis, optimization and statistics (see the papers ofRockafellar and co-authors in Rockafellar and Uryasev (2002)and Rockafellar et al. (2008)). Research problems of interest arisewhen trying to optimize coherent risk measures (or their data-based counterparts) under practically relevant restrictions onthe random vectors involved in their description. Regarding prac-tical applications, the real interest is focussed on dynamic riskmeasures. These need to be considered on a filtered probabilityspace and to satisfy versions of the above four properties for suit-ably defined stopping times in a time-consistent manner. Despiterecent advances in the theory of such measures e.g., Acciaio andPenner (2011); Bion-Nadal (2009) and Ruszczynski (2010), the is-sue about statistical inference about such dynamic measures islagging behind.

A way to deal with both performance and risk is based onutility theory. A utility function is a transformation of randomoutcomes which gives information on outcomes based on riskaverse modeling. A precursor of this theory was developed inMarkowitz (1952, 1987). His formulation of the investment deci-sion making process lead to the by now commonly named’’mean–variance’’ model for portfolio optimization. This modelobviously suffers from non-robustness and various versions ofrobust optimization have been proposed (Ben-Tal and Nemirovski,1999). When applied in portfolio optimization context theseversions seem to significantly immunize the portfolio againstuncertainty in the asset returns. Further derivative insuranceguarantees, recently proposed in Zymler et al. (2011) allow fora better trade-off between weak and strong guarantees of theworst-case portfolio return. However the resulting portfolio opti-mization problems are still similar to the Markowitz model. Theycan not therefore overcome one of its main shortcomings,namely the fact that the variance as a measure of risk penalizesequally an excess over the mean and a shortfall of the samemagnitude.

More non-symmetric risk measures are better suited for finan-cial applications. A popular risk measure of this type is the Value atRisk VaR(a),i.e., simply the a-quantile of the loss distribution. How-ever its non-convexity makes it a non-coherent risk measure. Analternative to VaR is the Average Value at Risk (AVaR) (Acerbiand Tasche, 2002; Rockafellar and Uryasev, 2002). When theunderlying probability space on which the random variable X is de-fined, is atom-less, AVaR is a coherent risk measure and even whenthe probability space is atomic, AVaR can become coherent aftersome continuity adjustment. Hence it is bold to say that AVaR isa coherent risk measure and as such is preferred in algorithmic sto-chastic optimization.

The AVaR plays a central role in the theory of coherent risk mea-sures due to the fact that it is a building block in the construction of

every law invariant coherent risk measure via the Kusuoka repre-sentation. In particular, the main object of our study in this paper,the higher moment dual risk measures, allow for such a represen-tation, too. These are tail risk measures, first introduced in Krokh-mal (2007), with their Kusuoka representation derived inDentcheva et al. (2010).

In order to define a framework for algorithmic portfolio optimi-zation using these measures, we will introduce multistage riskaverse optimisation using stochastic programming. A slightly sim-ilar approach has been used recently in Topaloglou et al. (2011)where the authors use it to optimize the risk of a portfolio compris-ing of options and forwards. However, the risk measure used thereis AVaR (in fact they minimize the conditional value at risk (CVaR)but it is well known that these two measures are very similar andin fact for the case of return distributions not having point masses,they actually coincide). Our goal in this paper is to formalize theuse of higher order dual risk measures as a criterion in option port-folio optimization and to demonstrate on a given data set theadvantage of using these risk measures.

The plan of the paper is a follows. In Section 2, we will recall andreflect on properties of higher moment dual risk measures. We willthen introduce our framework of algorithmic portfolio optimisa-tion using stochastic programming in Section 3. In Section 4, wewill describe how to estimate financial risk measure for optionsand will optimize Vanilla options composed from S&P500 stocksusing higher order coherent risk measure, the mean–variancemodel, and AVaR (in their dynamic formulation), in order to illus-trate the advantages of the higher order coherent risk measures incomparison to the other two. To the best of our knowledge, for aportfolio of derivatives, such comparison has not been made inthe literature. In Section 5 we discuss numerical implementationissues and illustrative examples. Section 6 concludes.

2. Higher order dual risk measures

Let ðX;F ; PÞ be a probability space an p 2 [1,1). We denote by R

the extended real line. For random variables X;Y 2 LpðX;F ; PÞ anda coherent risk measure gð:Þ : LpðX;F ; PÞ ! R. In the definition ofcoherent risk measure given in Section 1, random outcomes repre-sent a reward. If we choose the random variables X, Y to represent aloss then the requirements about monotonicity and translationinvariance in the definition should clearly be modified as follows:

� monotonicity: if X 6 Y then g(X) 6 g(Y),� translation invariance: if m 2 R then g (Y + m) = g(Y) + m.

Of particular importance for the results in this paper are a spe-cial class of coherent risk measures, namely the higher order tailrisk measures. They have been investigated in Cheridito and Li(2009); Krokhmal (2007); Dentcheva et al. (2010). In the paperKrokhmal (2007) the name higher order risk measures was intro-duced to a set of measures that are in fact a special case of a moregeneral family considered in Cheridito and Li (2009). The dual rep-resentation of these measures (termed Kusuoka representation)was obtained in Dentcheva et al. (2010).

The higher order risk measures, to be introduced below, are be-lieved to be advantageous in comparison to many proceedingcoherent risk measures with a fixed tail cutoff points precisely be-cause their cut-off point is adjustable to the level of confidencea 2 (0,1). In addition, the values of p > 1 allow to incorporate infor-mation from the higher pth order moments of the variables whichleads to a more flexible risk assessment.

Some examples of portfolio optimization models were shown inKrokhmal (2007). He also has demonstrated some advantages ofthese measures in comparison to the classical mean–variance

Page 3: Multistage optimization of option portfolio using higher order coherent risk measures

192 Y. Matmoura, S. Penev / European Journal of Operational Research 227 (2013) 190–198

optimization model of Markowitz. Our goal in this paper is to dem-onstrate similar advantages in dynamic settings and in applicationto derivative portfolio optimization. The examples include dataform the S&P500 Index.

For a random variable X 2 L1ðX;F ; PÞ, we denote the cumula-tive distribution function (cdf) and the higher order cdf as

FXðgÞ ¼ P½X 6 g� ¼ Fð1ÞX ðgÞ;

FðkÞX ðgÞ ¼Z g

�1Fðk�1Þ

X ðaÞda for g 2 R; k P 2:ð1Þ

The left-continuous inverse of the cumulative distribution functionis defined as follows:

Fð�1ÞX ðaÞ ¼ inffg : FXðgÞP ag for 0 < a < 1:

The absolute Lorenz function Fð�2ÞX ð�Þ : R! R, is defined as the

cumulative quantile:

Fð�2ÞX ðaÞ ¼

Z a

0Fð�1Þ

X ðtÞdt for 0 < a 6 1: ð2Þ

We define Fð�2ÞX ð0Þ ¼ 0 and observe that Fð�2Þ

X ð1Þ ¼ E½X�.Similarly to Fð2ÞX ð�Þ, the function Fð�2Þ

X ð�Þ is well defined for anyrandom variable X 2 L1ðX;F ; PÞ. By construction, it is convex. Veryoften, in the econometric literature, the normed variant, the rela-tive Lorenz curve a # Fð�2Þ

X ðaÞ=E½X� is considered instead of the Lor-enz curve. Because of the norming at a = 1, the relative Lorenzfunction is a convex-shaped cumulative distribution function.

Popular measures of risk are VaR, which at level a is defined asVaRaðXÞ ¼ �Fð�1Þ

X ðaÞ, and the AVaR, which at level a is defined as

AVaRaðXÞ ¼ �1a

Fð�2ÞX ðaÞ ¼ 1

a

Z a

0VaRtðXÞdt:

The AVaR is well known to be a coherent risk measure (for a dem-onstration see for example Krokhmal (2007)).

A fundamental result in the theory of coherent measures of riskis the Kusuoka theorem (Kusuoka, 2001):

For a non-atomic space X, every law invariant, finite-valuedcoherent measure of risk on L1ðX;F ; PÞ can be represented asfollows:

gðXÞ ¼ supl2M

Z 1

0AVaRaðXÞlðdaÞ; ð3Þ

with some convex setM� Pðð0;1�Þ, the set of probability measureson (0,1].

Kusuoka’s result underscores the importance of AVaR in theconstruction of law invariant coherent risk measures.

We define g(X) to be a Kusuoka measure of risk, if there exists aconvex set M in the set Pðð0;1�Þ such that for all such X we have

gðXÞ ¼ supl2M

Z 1

0AVaRaðXÞlðdaÞ:

Kusuoka’s fundamental result above in fact states that undergeneral conditions, a coherent measure of risk on L1ðX;F ; PÞ is aKusuoka measure. This result was also extended for Lp spaces(p P 1) (Shapiro et al., 2009). It can be used in two ways: to showa Kusuoka representation for a measure that is known to be coher-ent or, by choosing the set M, to generate a new coherent riskmeasure.

In order to describe the intimate relation of the Kusuoka mea-sures of risk with the higher order risk measures, we notice the fol-lowing Fenchel duality result from Ogryczak and Ruszczynski(1999).

We start by defining Fð2ÞX ð:Þ:

Fð2ÞX ðgÞ ¼Z g

�1FXðaÞda ¼ E½ðg� XÞþ� for g 2 R: ð4Þ

Recall that for the function f : Rn ! R, its Fenchel conjugate isf � : Rn ! R : f �ðsÞ ¼ supxfhs; xi � f ðxÞg.

The crucial non-trivial fact is that not only are both Fð2ÞX ð�Þ andFð�2Þ

X ð�Þ convex functions but they are also Fenchel conjugates toeach other, that is:

Fð�2ÞX ¼ Fð2ÞX

h i�and Fð�2Þ

X ¼ Fð�2ÞX

h i�ð5Þ

holds. In particular, we get the representation of AVaRa(X):

AVaRaðXÞ ¼ �1a

Fð2ÞX

h i�ðaÞ ¼ �1

asupg2R

ga� Fð2ÞðgÞn o

¼ infg2R

1a

E½ðg� XÞþ� � g� �

: ð6Þ

If we set c = 1/a and take M to consist of a single measure l thatputs its whole mass on a fixed a 2 (0,1] then AVaRa(X) is obviouslya Kusuoka measure of risk. The above observations also suggest thatby replacing the one-norm in (6) by a p-norm (p > 1) we can gener-ally consider the following higher moment measures of risk:

gc;pðXÞ ¼ infg2Rfckðg� XÞþkp � gg; p > 1; c > 1: ð7Þ

In (7) the choice of the constant c would generally depend on a cho-sen a 2 (0,1) and, typically, c ¼ 1

a. Although not obvious, it has beendemonstrated in Krokhmal (2007) that the risk measures in (7) arecoherent, i.e., they satisfy the four conditions listed in theIntroduction.

These higher order measures have advantage in comparison topredecessors of the type

�EX þ bkðEX � XÞþkp;

where the tail cutoff point EX is fixed. For the new measures, thecutoff is adjustable via ahence they are called tail risk measures.

Let p 2 (1,1), q: 1/p + 1/q = 1, and choose a constant c P 1. Con-sider the risk measure gc,p(X). The main result from Dentcheva et al.(2010) states that gc,p(X) has a Kusuoka representation

gc;pðXÞ ¼ supl2Mq

Z 1

0AVaRaðXÞlðdaÞ ð8Þ

holds. In particular, when p = 2 we get

gc;2ðXÞ ¼ �gX þgX � Fð�1Þ

X

� �þ

��� ���2

2

gX � Fð�1ÞX

� �þ

��� ���1

;

where gX is the optimal solution in (7). If Y ¼ gX � Fð�1ÞX

� �þ

then

gc;2ðXÞ ¼ EðYÞ � gX þVarðYÞEðYÞ :

The last term in the right hand side is the Fano factor (Fano, 1947)which is itself an important measure of noise-to-signal ratio usedin Statistics. We note that gX may not represent a quantile of X sothat the above risk measures are of novel type.

3. Dynamic risk optimization using stochastic programming

The risk measure is a theoretical construction involving a func-tional defined on a random variable. The random variable itself istypically a transformation of a random vector. (For example, theportfolio’s value is a weighted combination of the values of theseparate stocks in the portfolio. Optimizing the risk means findingthe most appropriate weights.) Typically, there are some con-straints involved in this optimization. One obvious constraint isthat the sum of weights should be equal to one. Given data, thetheoretically defined functional could be replaced by an estimateand then optimization of this estimate with respect to the

Page 4: Multistage optimization of option portfolio using higher order coherent risk measures

Y. Matmoura, S. Penev / European Journal of Operational Research 227 (2013) 190–198 193

constraints is an ultimate goal to achieve. However, rarely would itbe the case that independent and identically distributed observa-tions of the random variable exist. Typically only one observation(the portfolio’s value) will be available at each point in time. Hencethe dynamics of the portfolio’s value is represented by a time ser-ies. Of course, more structure should be imposed on the time seriesmodel to allow for a tractable solution.

The proper framework to perform the optimization is in dy-namic setting. This section describes the successive steps in suchan analysis.

3.1. From 2-stage to multistage risk averse optimization

Recalling the motivating problem from Section 1, we have anamount A0 to be invested in n assets where we need to satisfy bothrequirements of minimum expected return and of safe investment.The respective optimization problem can be formulated as follows:

min gXn

i¼1

xiRi

!;

subject to EXn

i¼1

xiRi

" #P c; ðiÞ

Xn

i¼1

xi ¼ A0; ðiiÞ

x 2 C; ðiiiÞ

ð9Þ

with g(.) being the risk measure, c P 0, x = [x1, x2, . . ., xn]0 and C a setto be defined each time. (Henceforth 0 denotes transposition).

The first constraint (i) symbolises the requirement fora minimum expected return, the second constraint (ii) representsthe distribution of the whole amount A0. In this setting aconstraint (iii) C ¼ Rn

þ would imply that long positions only areallowed.

Problem (9) represents a one-stage risk averse optimization. Wenote that in this formulation, a static evaluation of the risk is to beperformed hence it is typically easy to solve the resulting optimi-zation problem. Next we would like to extend it to a dynamic prob-lem by including many steps via a stochastic process formalismand by defining new risk evaluation that is adjusted to the result-ing filtration.

Let X be a sample space equipped with two r-algebras F 1 � F 2

and a probability measure P on ðX;F 2Þ. A functional g : Lp

ðX;F 2; PÞ ! LpðX;F 1; PÞ, is called conditional risk mapping if itsatisfies:

1. Convexity: for all a 2 [0,1], for all Z; Z0 2 LpðX;F 2; PÞ:

gðaZ þ ð1� aÞZ0Þ 6 agðZÞ þ ð1� aÞgðZ0Þ:

2. Monotonicity: for all ðZ; Z0Þ 2 LpðX;F 2; PÞ: if Z P Z0 then g(Z) 6g(Z0).

3. Translation invariance: for all ðZ;YÞ 2 LpðX;F 2; PÞ�LpðX;F 1; PÞ:gðZ þ YÞ ¼ gðZÞ � Y .

4. Positive homogeneity: for all a P 0; Z 2 LpðX;F 2; PÞ : gðaZÞ ¼agðZÞ.

We observe immediately that if we choose trivially F 1 ¼ f£;Xgwe can recover the static coherent risk measure definition. Formore details about conditional risk mappings we refer to Shapiroet al. (2009, Section 6.7.2).

For a 2 (0,1) and for each x 2X, we can now define a randomvariable called conditional AVaR as follows:

AVaRaðXjF 1ÞðxÞ ¼ infYðxÞ2LpðX;F1 ;PÞ

YðxÞ þ 1a

E½ð�X � YÞþjF 1�ðxÞ� �

:

ð10Þ

Then, similarly to the approach in Section 2 we can define a higherorder moment coherent conditional risk measure for all a 2 (0,1),p > 1 as follows:

gðXjF 1ÞðxÞ ¼ infYðxÞ2LpðX;F1 ;PÞ

YðxÞ þ 1a

E ð�X � YÞpþjF 1� �1=pðxÞ

� �:

ð11Þ

To deal with a multistage modeling on LpðX;F ; PÞ let us con-sider filtrations F 1 � F 2 � � � � F T with F T ¼ F and F 1 ¼ f£;Xg.Denote for j ¼ 1;2; . . . ; T : nj ¼ LpðX;F j; PÞ. Then we denote a se-lected family of conditional risk mappings satisfying conditions(1)–(4) above:

gtþ1jF t: ntþ1 ! nt ; t ¼ 1;2; . . . ; T � 1:

The composition g2jF1 � � � gTjF T�1

: nT ! R is a coherent risk mea-sure (see Shapiro et al. (2009), Section 6.7.3). We can now formulatea multistage risk averse optimization problem. We note that in thisformulation, as opposed to the static formulation (9) above, we usesuperscripts xi, i = 1, 2, . . ., T to indicate the time evolution asfollows:

minx1 ;x2 ;...;xT

fZ1ðx1Þ þ g2jF1½Z2ðx2ðxÞ;xÞ þ � � � þ

gT�1jFT�2½ZT�1ðxT�1ðxÞ;xÞ þ gTjFT�1

½ZTðxTðxÞ;xÞ���g;subject to x1 2 H1 ¼ Rn; xtðxÞ 2 Htðxt�1ðxÞ;xÞ; t ¼ 2; . . . ; T;

with Zt such that Ztðxt ; :Þ 2 nt ; Ht � Rnt�1 �X:

ð12Þ

Problem (12) represents a very general formulation of the mul-tistage risk averse optimization problem. The advantage of thisgeneral formulation is that it can be specialized with any kind ofproducts from correlated spot assets to derivative products. Inthe next subsection, we will specialize this formulation for the pur-pose of option portfolio selection.

3.2. Numerical implementation for option portfolio selection

Denote by Rti

; t ¼ 1;2; . . . ; T the random process of rates of re-

turn for the ith asset (i = 1, 2, . . ., n). Hence Rt ¼ Rt1;R

t2; . . . ;Rt

n

� �0is a

random vector of rates of return of all assets in the portfolio at timet. We denote the r-algebra F t ¼ rðR1; . . . ;RtÞ. The random process(xt), adapted to the filtration F t , is xt ¼ xt

1; xt2; . . . ; xt

n

� �0 and it repre-sents the random vector of weights associated to the n assets attime t. Then the (random) return of the portfolio at time t isZtðxÞ ¼

Pni¼1xt

i ðxÞRti ðxÞ ¼ xt0RtðxÞ. The price process, that is, the

F t-predictable random process registering the value of the portfo-lio, will be denoted by (bt), t = 1, 2, . . ., T. For a specific realization rt

at time t of Rt, and for a specific selection ~xt of xt we have

btþ1 ¼ bt þXn

i¼1

~xti r

ti ;

with an initial value b0 = A0. We arrive at the following nested for-mulation of the portfolio selection problem:

minðxt Þf�gðZTÞg ¼ fg1½. . . gT�1jFT�2

½gTjFT�1½ZT ���g;

subject to EXn

i¼1

xti R

ti jF t�1

" #P c;

Xn

i¼1

xti ¼ bt ;

xt 2 Ct ; t ¼ 1;2; . . . ; T;

ð13Þ

with Ct a subset of Rn (usually Ct ¼ Rnþ) and �gð:Þ ¼

g1 g2jF1 . . . gTjF T�1

ð:Þ. Note that �gð:Þ is directly a coherent riskmeasure if the conditional risks are defined as the respective condi-tional expectations (Shapiro et al., 2009, p. 325).

Page 5: Multistage optimization of option portfolio using higher order coherent risk measures

194 Y. Matmoura, S. Penev / European Journal of Operational Research 227 (2013) 190–198

Next we introduce a general algorithm to implement the aboveoptimization process. It includes the following steps:

1. At t = 0 we know the vector of initial prices of assets

S0 ¼ S01; . . . ; S0

n

h i0and we choose b0 = 1.

2. For each fixed t = 1 to t = T � 1:(a) From the information on previous realizations (up to

moment (t � 1)) and using the assumptions on the assetprice evaluation, we get the distribution of Rt.

(b) From the distribution of Rt we draw an appropriately scaled

sample of size J: RtJ ¼

Rt1;1 . . . Rt

1;J. . . . . . . . .Rt

n;1 . . . Rtn;J

0@ 1A, where J denotes the

number of simulated price evolutions.(c) Using the generated samples in Rt

J we replace the associatedexpected values by their empirical estimators and solve theresulting optimization problems:

minðxt Þ

gtjF t�1ðxtÞ0Rt

J

� �n o;

subject to1J

XJ

j¼1

ðxtÞ0RtJ ej P c;

Xn

i¼1

xti ¼ bt ;

xt 2 Ct ; t ¼ 1;2; . . . ; T:

ð14Þ

Hereby we have denoted by ej the unit-length vector [0, 0, . . .,1, . . ., 0]0 with a one in jth position.

(d) We calculate the portfolio value at the next moment in timevia btþ1 ¼ bt þ ð~xtÞ0rt .

3. For t = T: Now that b = (b0, b1, . . ., bT) and ð~x ¼ ð~x1; . . . ; ~xT�1Þ areavailable we can analyze the price variation (delta) of the port-folio and the substitution of the optimal values of the optimiza-tion problem demonstrates the risk evolution.

4. Option portfolio optimization using higher order coherentrisk measures

We note that the algorithmic method in Section 3.2 is presentedin its very general form whereby we did not require specificassumptions on the price evolutions and on the kind of assets inthe portfolio. In this section, we will specialize the method whenapplying it for a portfolio of options on stocks. This will allow usto evaluate the distribution of Rt. We then apply the method byvarying the choice of the risk measures in order to illustrate theadvantage of the use of higher order coherent risk measures.

At each step of the optimization algorithm we need to find adistribution of the rate of returns of the options to substitute itin the estimation of the risk. The risk is to be optimized by choos-ing the optimal weights (the xt values) that represent the portfoliore-balancing. Further, we need the variation of the option prices ateach point in time. We achieve this by simulating many asset pathsbetween every two time points. For each path we calculate the op-tion price associated to the two dates and derive a distribution forthe variation of the option price between these days, and thereturn.

We now consequently introduce a suitable option pricing mod-el first, then we describe scenario simulations for the underlyingassets in order to derive the rate distribution. Finally, we outlineour approach in estimating the risk.

4.1. Option pricing model

There are many option pricing methods in the literature, withdifferent levels of sophistication. Here, we will be using a relatively

simple and not too much computer intensive method, yet which isflexible enough to evaluate the implied volatilities and correlationsbetween the assets. The method uses Monte Carlo simulationsbased on a multivariate GARCH (1,1) model to estimate thetime-varying covariance matrices. As the time for the Monte Carlocalculations in this model grows only linearly with the number ofassets, the added advantage is that the method can be applied witha large number of options. The detailed presentation of the quasi-likelihood method to estimate the time-varying covariance matri-ces is in Ledoit et al. (2003). We only summarize the steps in a min-imalist way in order to be able to describe the numerical procedurewe are employing.

Let uti be the realization of the rate of return of the asset i at

time t. We assume that for all i = 1, 2, . . ., n and t ¼1;2; . . . ; T : E Ut

i jF t�1� �

¼ 0. The GARCH-style model for the covari-ance matrices is defined by

Cov Uti ;U

tj jF t�1

� �¼: Cov t

ij ¼ cij þ aijut�1i ut�1

j þ bijCov t�1ij ; i – j;

Var Uti jF t�1

¼: rt

i2 ¼ r2

i þ aiiut�1i

2 þ biirt�1i

2:

ð15Þ

The interpretation in (15) is that the time-varying covariance Cov tij

between the ith and jth asset at time t depends on a long-termcovariance between these two assets and also on the history at timet � 1. Similar is the interpretation of the time varying variance inthe second Equation in (15). To ensure covariance stationarity, thecoefficients aij, bij, aii, bii satisfy the restrictions

0 < aij þ bij < 1; 0 < aii þ bii < 1; aij > 0; bij > 0;aii > 0; bii > 0:

We assume multivariate normality: Ut ¼ Ut1;U

t2; . . . ;

�Ut

n�0 N 0;Rtð Þ with

Rt ¼rt

12

. . . Cov t1;n

. . . . . . . . .

Cov tn;1 . . . rt

n2

0B@1CA:

Then the quasi-likelihood procedure for consistent estimation ofthe diagonal coefficients of Rt and of aii; bii; r2

i is well known(Campbell et al., 1997, Section 12.2). In Ledoit et al. (2003), asimple quasi-likelihood method is suggested to find consistentestimators of the off-diagonal coefficients by restricting eachtime attention to the 2 � 2 submatrices corresponding to thevariables Ui and Uj. For inference purposes, we follow a standardapproach whereby if L is certain period of time used to estimatethe above parameters by using daily returns then the estimatorsobtained are considered ‘‘valid’’ for the next time period oflength L.

To simplify the treatment, we assume, for now, that all optionsin the portfolio have the same exercise date T, and that we aredealing with European call and put options only. Let K = [K1, K2,. . ., Kn]0 be the vector containing the n strike prices and r be therisk-free rate. Then at time t < T the option price vector is

f ðtÞ ¼f1ðtÞ. . .

fnðtÞ

0B@1CA ¼ e�rðT�tÞ

E max � ST1 � K1

� �;0

� �jSt

1

h i. . .

E max � STn � Kn

� �;0

� �jSt

n

h i0BB@

1CCA; ð16Þ

where we use (+) for a call option and a (�) for a put option,and St

i denotes the price of the underlying asset at time t.When evaluating the option, we assume that the stock pricefollows:

STi ¼ St

i eðr�1

2

Xn

j¼1

eRtT ði;jÞÞðT�tÞþ

ffiffiffiffiffiffiT�tp

Ui

; ð17Þ

Page 6: Multistage optimization of option portfolio using higher order coherent risk measures

Y. Matmoura, S. Penev / European Journal of Operational Research 227 (2013) 190–198 195

with U ¼ ½U1;U2; . . . ;Un�0 N 0; eRtT

� �where each element of the

matrix eRtT is calculated as:

eRtTði;jÞ¼ ðT� tÞ cij

1�aij�bijþ 1�ðaijþbijÞT�t

�logðaijþbijÞðT� tÞ ðRtÞði;jÞ�cij

1�aij�bij

� ( )

(we refer to Hull (2009, p. 479)). The above assumption about thestock price evolution is generalizing the assumptions regardingthe Black–Scholes model in that now the variance is notconsidered as a constant anymore and the assets are allowedto be correlated. Having simulated many random vectorsUk ¼ Uk

1;Uk2; . . . ;

hUk

n�0; k ¼ 1;2; . . . ;B we calculate for each of them

the resulting price vectors fk(t) and then take the arithmeticmean of the B simulated vectors to evaluate f(t). We choseB = 10,000 in our simulations and we used the Cholesky trans-form of eRt

T and the antithetic variable method (Hull, 2009,p. 425) to reduce the variance.

4.2. Evolution scenarios

Since the returns distribution between t and t + 1 is difficult toevaluate explicitly, we adopt simulations at this step, too. We sim-ulate underlying stocks evolutions between t and t + 1.

The classical assumption is that stock increments follow Brown-ian motion, which leads to the standard log-normal model used inoption pricing. However, when historical data on the underlyingassets is available, we are in a position and we perform a moreaccurate modeling by estimating the density of the daily returnsnon-parametrically by using a kernel density estimation method.The advantage of this approach is that we do not have to makeany assumptions on the returns distribution.

For the multistage portfolio optimization duration time wechoose the same T as for the option’s life. This means that theend of the algorithmic optimization coincides with the end ofthe options’ life. We divide the interval [0,T] in equidistant pointsDt ¼ T

M. In our simulations, we use Dt = 5 (that is, a week oftrading). At each step kDt, we know SlDt, l = 1, 2, . . ., k and theprevious returns rlDt, l = 1, 2, . . ., (k � 1). Now we generate then � J matrix

eSðkþ1ÞDt ¼Sðkþ1ÞDt;1

1 . . . Sðkþ1ÞDt;J1

. . . . . . . . .

Sðkþ1ÞDt;1n . . . Sðkþ1ÞDt;J

n

0BB@1CCA

consisting of J columns of simulated prices at time (k + 1)Dt. Theprice simulations are done under the kernel density estimator ob-tained by using the historical data. Then for each column S(k+1)Dt,i

we calculate the associated option prices to get the matrix

gðkþ1Þ

E max � ST1 � K1

� �; 0

� �jSðkþ1ÞDt;1

1

h i. . . E max � ST

1 � K1

� �;0

� �jSðkþ1ÞDt;J

1

h i. . . . . . . . .

E max � STn � K1

� �; 0

� �jSðkþ1ÞDt;1

n

h i. . . E max � ST

n � K1

� �;0

� �jSðkþ1ÞDt;1

n

h i0BBB@

1CCCA;

with the scalar factor g(k+1) = e�r(T�(k+1)Dt). Each column of the ma-trix will be denoted ð~f ððkþ 1ÞDtÞjÞ; j ¼ 1;2; . . . ; J. At any momentkDt in time, we know (have observed) the vector f ðkDtÞ 2 Rn thuswe have a sample of return rates for the options at that time. Thishelps us to calculate the variation of the option price:

D~f ijðkDtÞ ¼ ~f iððkþ 1ÞDtÞj � fiðkDtÞ; i ¼ 1;2; . . . ;n; j

¼ 1;2; . . . ; J: ð18Þ

The sample of values (18) constitutes the scenario of evolution be-tween kDt and (k + 1)Dt.

Finally then we are in a position to evaluate the sample of ratesof return

RkDtij ¼

~f iððkþ 1ÞDtÞj � fiðkDtÞfiðkDtÞ ; 1 6 j 6 J; 1 6 i 6 n:

4.3. Risk evaluation

In accordance with (10) and (11) we have for Z 2 LpðX;F t ; PÞand a 2 (0,1) the conditional versions

CAVaRaðZÞ ¼ infY2LpðX;F t�1 ;PÞ

Y þ 1a

E½ð�Z � YÞþjY �� �

and for p > 1, the conditional higher moment coherent risk

CHMCRaðZÞ ¼ infY2LpðX;F t�1 ;PÞ

Y þ 1a

E ð�Z � YÞpþjY� �1=p

� �:

Suppose we are at time kDt (hence we know r(k�1)Dt and SkDt). Aswe have seen, an estimator of R(k+1)Dt is obtained using

RkDtij

� �; i ¼ 1;2; . . . ;n; j ¼ 1;2; . . . ; J. The corresponding risks can

be estimated as follows:

dCAVaRaðZkDtÞ ¼ infg2R

gþ 1a

1J

XJ

j¼1

�g�Xn

i¼1

xkDti RkDt

ij

� �þ

!( ); ð19Þ

dCHMCRaðZkDtÞ ¼ infg2R

gþ 1a

1

J1=p

XJ

j¼1

�g�Xn

i¼1

xkDti RkDt

ij

� � !p

þ

!1=p8<:

9=;:ð20Þ

5. Numerical implementation issues and illustrative examples

When p > 1, the optimization at each step k = 1, 2, . . ., M � 1 rep-resents a convex optimization problem with convex constraints,hence it has a unique solution. The case p = 1 actually representsa linear programming problem and can in principle be solved withthe simplex method although the general procedure for p > 1 canalso be applied for this particular case. We illustrate this for thecase of (20) [(19) is obviously similarly treated]. We see easily thatit is equivalent to:

ming;d;x

gþ 1a

1J1=p

XJ

j¼1

dpj

!1=p8<:

9=;;subject to

Xn

i¼1

xkDti ¼ bðtÞ;

1J

XJ

j¼1

Xn

i¼1

RkDtij xkDt

i P c;

� g�Xn

i¼1

RkDtij xkDt

i 6 dj;

dj P 0;

xkDti P k:

ð21Þ

Hereby we have denoted by d the vector in which all vectors dj, j = 1,2, . . ., J are stacked. If we only allow for long positions then we setk = 0. Else we can let k to take a negative value depending on theinvestor’s choice.

We also note that for p = 2 the problem is a second order coneprogramming problem. In this paper we use the CVX package tosolve the problem. The latter package is a freely downloadable Mat-lab software for specifying and solving convex programs (Grant andBoyd, 2011) hence it can be used for any choice of p > 1 in our case.

To illustrate the advantages of the higher order coherent riskmeasures, we use the stocks that are involved as components ofthe S&P500 index. We build options on n = 10 of these stocks andcombine these options in a portfolio to be optimized. The sample

Page 7: Multistage optimization of option portfolio using higher order coherent risk measures

0.5

1

1.5

2

2.5

0 5 10 15 20 25 30 35 40 45 50 55 60 65 70 75 80 85 90 95 100

Risk Averse Optimized portfolio evolutions

AVaR Optimized portfolio (euro options)SMCRM Optimized portfolio (euro options)

Mean Var Optimized portfolio (euro options)TMCRM Optimized portfolio (euro options)

Fig. 1. Weekly rebalanced portfolio levels from 2003/11/04 to 2005/09/22.

196 Y. Matmoura, S. Penev / European Journal of Operational Research 227 (2013) 190–198

is composed by using the Close Prices of S&P500 Stocks from 4thJanuary 2000 to 22nd September 2005 which gives 1500 quota-tions for each Stock. Initialization is done using quotations from4th January 2000 to 3rd November 2003. Using 1000 daily returns,we estimate the GARCH (1,1) multivariate coefficients and theCovariance Matrix (15). Next we calculate the 995 weekly returns.Using the remaining data from 4th November 2003 to 22nd Sep-tember 2005, we launch out-of-sample simulations. At this stagedata appear like continuously new quotations. Each week, the sam-ple of 995 weekly returns and the sample of 999 daily returns arerefreshed by deleting the five ‘‘oldest’’ and adding five new weeklyreturns. Then portfolio re-balancing is made using these samples.Thus at each loop of the algorithm (i.e., every week) the previous995 weekly returns are used to estimate the return distributionsof the stocks. Previous 999 daily returns are used to estimate theparameters of the GARCH (1,1) model. To illustrate the procedurenumerically, we choose a basket composed with calls in the moneydefined at the date 4th November 2003, set K = 0.8 � S0. In our

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0 5 10 15 20 25 30 35 40 45 50

Risk evolution of Risk Averse Optimized portfolio

Va

Fig. 2. Weekly rebalanced portfolio risk evo

example, we have T = 2005/09/22. We choose cij to coincide withthe risk-free rate a.s. and as a result the portfolio performs at leastas a risk free asset. For each Risk Measure we choose a level of con-fidence to have a fair comparison. We require these levels tosatisfy:

aðAVaRÞ < aðSMCRMÞ < aðTMCRMÞ;

with SMCRM denoting ‘‘Second Order Coherent Risk Measure’’(p = 2), and TMCRM denoting ‘‘Third Moment Coherent Risk Mea-sure’’ (p = 3). These choices are imposed by the Higher Order RiskMeasures definition. For a discussion, we refer to Example 2.2 inKrokhmal (2007) (by noting that (1 � a) used in his notationis denoted as a in this paper). We choose a(AVaR) = 0.01,a(SMCRM) = 0.07 and a(TMCRM) = 0.1. Finally we choosek(t) = �b(t)/n with n being the number of assets in the portfolio.With other words, we do allow short positions as long as they arelimited to the opposite value of the portfolio.

55 60 65 70 75 80 85 90 95 100

s (10 euro Call In The Money S&P 500 assets)

AVaR Optimized portfolioSMCRM Optimized portfolio

Mean Var Optimized portfolioTMCRM Optimized portfolio

R(CVaR dual value) Optimized portfolio

lutions from 2003/11/04 to 2005/09/22.

Page 8: Multistage optimization of option portfolio using higher order coherent risk measures

0.5

1

1.5

2

2.5

0 5 10 15 20 25 30 35 40 45 50 55 60 65 70 75 80 85 90 95 100

Delta evolution of Optimized portfolios (10 euro Call In The Money S&P 500 assets)

AVaR Optimized portfolio (euro options)SMCRM Optimized portfolio (euro options)

Mean Var Optimized portfolio (euro options)TMCRM Optimized portfolio (euro options)

Fig. 3. Weekly rebalanced portfolio delta evolution from 2003/11/04 to 2005/09/22.

Y. Matmoura, S. Penev / European Journal of Operational Research 227 (2013) 190–198 197

Fig. 1 represents the evolution of the portfolio’s levels. Everyportfolio starts with the value 1 and follows its path that is drivenby the risk optimization methodology. We can see that at the endof the 2 years of out-of-sample simulations, the best performing isthe portfolio based on TMCRM (p = 3) optimization. Second bestwas the result obtained via SMCRM (p = 2) optimization, third bestis the AVaR optimization (p = 1) and the worst result is obtainedwhen optimization using the Markowitz model is performed.

Fig. 2 shows the risk evolution of each portfolio. We can verifyagain that virtually at each point in time, the SMCRM, TMCRM andAVaR results are about equal and are uniformly higher than thoseobtained when using VaR. It can also be seen that the SMCRM andTMCRM results are very volatile.

Fig. 3 reports the Delta evolution of each portfolio. Delta hasbeen calculated by utilizing the Monte Carlo simulations used tocalculate option price. The percentage of variation of underlyingwas settled at 1%. We can see that Delta grows very strongly fromthe 80th week for SMCRM, TMCRM and AVaR portfolios. Moreoverthe Delta is very volatile for every portfolio.

These results highlight the limitation of the Markowitz modelwhen it deals with no cash products. Complications appear whenwe treat even simple vanilla options. Simple minimization of var-iance is not efficient to reduce future portfolio losses. The higherorder coherent risk measures, by focusing on the tail behavior ofthe losses, take into account more efficiently the information pro-vided by the past. However these portfolios exhibit larger varia-tions which are difficult to predict. An investor who wants amore stable investment could in addition apply a reduction of glo-bal variance method to these portfolios by dividing every weightapplied on the different options by their realized volatility. In thisway the global volatility can be kept in control.

We also observe heavy dependence on delta. A way to reducethis dependence can be to take an opposite position on Index witha level of delta. This method can reduce performance but will con-siderably reduce dependence on market moves.

6. Conclusion

This paper developed application of higher order risk measuresto the optimization of an option portfolio. Starting with the discus-

sion of coherent risk measures in static and dynamic setting, wediscussed the family of higher order coherent risk measuresthrough their general definition. This family can be used to esti-mate risk of every process which follows uncertain paths. To dealwith a general portfolio optimization we introduced the multistagerisk averse optimization using Stochastic Programming. The lattermethod is a generalization of the well-known 2-stage optimizationwhich minimizes a risk functional on portfolio weights, imposing aminimum expected return. An important advantage of this optimi-zation is that it can be applied with every instrument as long as wecan get a return distribution. We chose to apply the method to Va-nilla European options. As the members of an S&P500 Index areintensively correlated, we took into account these correlationsand the time dependence through the use of the flexible Multivar-iate GARCH (1,1) covariance model in option pricing. After gettinga sample of underlying paths between two instants we were ableto obtain the options weekly returns. These evolution scenarioswere implemented in the multistage optimization using each ofthe Markowitz, AVaR and the HMCRM estimators. Results obtainedwith European Calls show a wide efficiency of HCMRM when esti-mating options risk. The methodology outlined in the paper can beextended to any kind of options. Different option pricing have to beused to get option returns especially for path dependant optionsbut the algorithmic method is generally applicable. An investorwho uses American options can use a tree method to estimatethe variation of option prices, for example.

There are further avenues to be explored. The global perfor-mance of risk optimization strategies could be improved by addingmore layers on the risk control. One possibility is to add a deltahedging with a daily frequency to reduce the market expositionof the portfolio. A Vega and Gamma control can also be added byde-leveraging and leveraging the global exposition in a weekly ormonthly basis. One more important avenue is, by using again thehigher order coherent risk measures instead of the more traditionalones, to study the risk evolution of a hedging strategy during thelife cycle of options.

The demonstrated effects in Section 5 are data specific. Replicat-ing the empirical part of the study using newer data and differenttime spans, varying the choice of the constant c in (7), changing thesubset of stocks from the S&P index and other variations of the dataanalysis are possible. These activities are left as a future avenue of

Page 9: Multistage optimization of option portfolio using higher order coherent risk measures

198 Y. Matmoura, S. Penev / European Journal of Operational Research 227 (2013) 190–198

research. However the presented case study indicates the great po-tential and the promising performance of the higher order coher-ent risk measures as a criterion in optimizing the performance ofoption portfolios.

Acknowledgements

The authors are very grateful to the editor and to the threeexternal referees for their valuable input on an earlier submission.Their very useful comments and suggestions helped us to improveboth the style and the presentation of the paper.

References

Acciaio, B., Penner, I., 2011. Dynamic risk measures. In: Di Nunno, G., Øksendal, B.(Eds.), Advanced Mathematical Methods for Finance. Springer, Heidelberg, NewYork, pp. 1–34.

Acerbi, C., Tasche, D., 2002. On the coherence of expected shortfall. Journal ofBanking and Finance 26, 1487–1503.

Artzner, P., Delbaen, F., Eber, J.-M., Heath, D., 1999. Coherent measures of risk.Mathematical Finance 9, 203–228.

Ben-Tal, A., Nemirovski, A., 1999. Robust solutions of uncertain linear programs.Operations Research Letters 25, 1–13.

Bion-Nadal, J., 2009. Time consistent dynamic risk processes. Stochastic Processesand their Applications 119, 633–654.

Campbell, J., Lo, A., MacKinlay, A.C., 1997. The Econometrics of Financial Markets.Princeton University Press, Princeton.

Cheridito, P., Li, T.H., 2009. Risk measures on Orlicz hearts. Mathematical Finance19, 189–214.

Dentcheva, D., Penev, S., Ruszczynski, A., 2010. Kusuoka representation of higherorder dual risk measures. Annals of Operations Research 181, 325–335.

Grant, M., Boyd, S., 2011. CVX: Matlab software for disciplined convexprogramming, version 1.21. <http://cvxr.com/cvx/>.

Fano, U., 1947. Ionization yield of radiations: II. The fluctuations of the number ofions. Physical Review 72, 26–29.

Föllmer, H., Schied, A., 2002. Convex measures of risk and trading constraints.Finance and Stochastics 6, 429–447.

Krokhmal, P., 2007. Higher moment coherent risk measures. Quantitative Finance 7,373–387.

Hull, J., 2009. Options, Futures and Other Derivatives. Pearson Education, NewJersey.

Kusuoka, S., 2001. On law invariant coherent risk measures. Advances inMathematical Economics 3, 83–95.

Ledoit, O., Santa-Clara, P., Wolf, M., 2003. Flexible multivariate GARCH modelingwith an application to international stock markets. The Review of Economicsand Statistics 85, 735–747.

Markowitz, H.M., 1952. Portfolio selection. Journal of Finance 7, 77–91.Markowitz, H.M., 1987. Mean–Variance Analysis in Portfolio Choice and Capital

Markets. Blackwell, Oxford.Ogryczak, W., Ruszczynski, A., 1999. From stochastic dominance to mean-risk

models: semideviations and risk measures. European Journal of OperationalResearch 116, 33–50.

Rockafellar, R., Uryasev, S., Zabarankin, M., 2008. Risk tuning with generalized linearregression. Mathematics of Operations Research 33, 712–729.

Rockafellar, R., Uryasev, S., 2002. Conditional value-at-risk for general lossdistributions. Journal of Banking & Finance 26, 1443–1471.

Ruszczynski, A., 2010. Risk-averse dynamic programming for Markov decisionprocesses. Mathematical Programming 125, 235–261.

Shapiro, A., Dentcheva, D., Ruszczynski, A., 2009. Lectures on StochasticProgramming: Modelling and Theory. SIAM.

Topaloglou, N., Vladimirou, H., Zenios, S., 2011. Optimizing international portfolioswith options and forwards. Journal of Banking & Finance 35, 3188–3201.

Zymler, S., Rustem, B., Kuhn, D., 2011. Robust portfolio optimization with derivativeinsurance guarantees. European Journal of Operational Research 210, 410–424.