LP Solvable Models for Portfolio Optimization: A Survey and Comparison - Part I

Similar documents
On LP Solvable Models for Portfolio Selection

Efficient optimization of the reward-risk ratio with polyhedral risk measures

A SECOND ORDER STOCHASTIC DOMINANCE PORTFOLIO EFFICIENCY MEASURE

Portfolio optimization with a copula-based extension of conditional value-at-risk

Włodzimierz Ogryczak. Warsaw University of Technology, ICCE ON ROBUST SOLUTIONS TO MULTI-OBJECTIVE LINEAR PROGRAMS. Introduction. Abstract.

Reformulation of chance constrained problems using penalty functions

Multiple criteria linear programming model for portfolio selection

Robustness and bootstrap techniques in portfolio efficiency tests

Department of Economics and Management University of Brescia Italy WORKING PAPER. Discrete Conditional Value-at-Risk

Valid Inequalities and Restrictions for Stochastic Programming Problems with First Order Stochastic Dominance Constraints

COHERENT APPROACHES TO RISK IN OPTIMIZATION UNDER UNCERTAINTY

Multicriteria models for fair resource allocation 1

Heuristic algorithms for the portfolio selection problem with minimum transaction lots

Inverse Stochastic Dominance Constraints Duality and Methods

Portfolio optimization with stochastic dominance constraints

Birgit Rudloff Operations Research and Financial Engineering, Princeton University

Operations Research Letters. On a time consistency concept in risk averse multistage stochastic programming

RISK AND RELIABILITY IN OPTIMIZATION UNDER UNCERTAINTY

Department of Social Systems and Management. Discussion Paper Series

Robustness in Stochastic Programs with Risk Constraints

ORIGINS OF STOCHASTIC PROGRAMMING

Bayesian Optimization

Finanzrisiken. Fachbereich Angewandte Mathematik - Stochastik Introduction to Financial Risk Measurement

THE REFERENCE POINT METHOD WITH LEXICOGRAPHIC MIN-ORDERING OF INDIVIDUAL ACHIEVEMENTS

On Kusuoka Representation of Law Invariant Risk Measures

Financial Optimization ISE 347/447. Lecture 21. Dr. Ted Ralphs

Stochastic Programming with Multivariate Second Order Stochastic Dominance Constraints with Applications in Portfolio Optimization

Decomposability and time consistency of risk averse multistage programs

MFM Practitioner Module: Risk & Asset Allocation. John Dodson. February 18, 2015

Particle swarm optimization approach to portfolio optimization

Markowitz Efficient Portfolio Frontier as Least-Norm Analytic Solution to Underdetermined Equations

Stochastic Dominance and Prospect Dominance with Subjective Weighting Functions

Beyond average and minimax in MILP

Miloš Kopa. Decision problems with stochastic dominance constraints

Reward-Risk Portfolio Selection and Stochastic Dominance

Coherent Risk Measures. Acceptance Sets. L = {X G : X(ω) < 0, ω Ω}.

Fundamentals in Optimal Investments. Lecture I

Relaxations of linear programming problems with first order stochastic dominance constraints

Minimax and risk averse multistage stochastic programming

Multivariate Stress Testing for Solvency

Aggregate Risk. MFM Practitioner Module: Quantitative Risk Management. John Dodson. February 6, Aggregate Risk. John Dodson.

Combinatorial Data Mining Method for Multi-Portfolio Stochastic Asset Allocation

Coherent risk measures

Stochastic Optimization with Risk Measures

Inequality measures and equitable approaches to location problems q

MEASURES OF RISK IN STOCHASTIC OPTIMIZATION

Comparing downside risk measures for heavy tailed distributions

MFM Practitioner Module: Risk & Asset Allocation. John Dodson. January 25, 2012

Testing Downside-Risk Efficiency Under Distress

Risk optimization with p-order conic constraints

OPTIMIZATION APPROACHES IN RISK MANAGEMENT AND FINANCIAL ENGINEERING

Stochastic dominance and risk measure: A decision-theoretic foundation for VaR and C-VaR

Complexity of Bilevel Coherent Risk Programming

Motivation General concept of CVaR Optimization Comparison. VaR and CVaR. Přemysl Bejda.

SOME HISTORY OF STOCHASTIC PROGRAMMING

Risk-Averse Dynamic Optimization. Andrzej Ruszczyński. Research supported by the NSF award CMMI

Robust Multicriteria Risk-Averse Stochastic Programming Models

Pareto Efficiency in Robust Optimization

Portfolios that Contain Risky Assets 17: Fortune s Formulas

Information, Utility & Bounded Rationality

Upper bound for optimal value of risk averse multistage problems

ИЗМЕРЕНИЕ РИСКА MEASURING RISK Arcady Novosyolov Institute of computational modeling SB RAS Krasnoyarsk, Russia,

Expected Shortfall is not elicitable so what?

Set-Valued Risk Measures and Bellman s Principle

The Subdifferential of Convex Deviation Measures and Risk Functions

Modern Portfolio Theory with Homogeneous Risk Measures

Research Article Continuous Time Portfolio Selection under Conditional Capital at Risk

On a class of minimax stochastic programs

On Kusuoka representation of law invariant risk measures

A Robust Portfolio Technique for Mitigating the Fragility of CVaR Minimization

Distributionally Robust Convex Optimization

Robust Optimization: Applications in Portfolio Selection Problems

Tail Value-at-Risk in Uncertain Random Environment

Examples. 1.1 Cones in Vector Spaces

MFM Practitioner Module: Risk & Asset Allocation. John Dodson. February 4, 2015

Expected Shortfall is not elicitable so what?

Are Probabilities Used in Markets? 1

Spectral Measures of Uncertain Risk

Convex optimization problems. Optimization problem in standard form

Theoretical Foundation of Uncertain Dominance

Higher Moment Coherent Risk Measures

POLYNOMIAL-TIME IDENTIFICATION OF ROBUST NETWORK FLOWS UNDER UNCERTAIN ARC FAILURES

Risk Aversion and Coherent Risk Measures: a Spectral Representation Theorem

Generalized quantiles as risk measures

Distributionally Robust Discrete Optimization with Entropic Value-at-Risk

Multivariate GARCH models.

IR / July. On Stochastic Dominance and Mean-Semideviation Models

Regularly Varying Asymptotics for Tail Risk

Value at Risk and Tail Value at Risk in Uncertain Environment

MEAN-ABSOLUTE DEVIATION PORTFOLIO SELECTION MODEL WITH FUZZY RETURNS. 1. Introduction

Time inconsistency of optimal policies of distributionally robust inventory models

Asymptotic distribution of the sample average value-at-risk

In this work we introduce a class of measures for evaluating the quality of financial positions based on their

Worst-Case Conditional Value-at-Risk with Application to Robust Portfolio Management

The Canonical Model Space for Law-invariant Convex Risk Measures is L 1

Robust multi-sensor scheduling for multi-site surveillance

Semidefinite and Second Order Cone Programming Seminar Fall 2012 Project: Robust Optimization and its Application of Robust Portfolio Optimization

The newsvendor problem with convex risk

On Markov Properties in Evidence Theory

An Application of Cointegration Analysis on Strategic Asset Allocation

Transcription:

LP Solvable Models for Portfolio Optimization: A Survey and Comparison - Part I Renata Mansini,W lodzimierz Ogryczak, M. Grazia Speranza May 1, 2001 Abstract The Markowitz model of portfolio optimization quantifies the problem in a lucid form of only two criteria: the mean, representing the expected outcome, and the risk, a scalar measure of the variability of outcomes. The classical Markowitz model uses the variance as the risk measure, thus resulting in a quadratic optimization problem. Following Sharpe s work on linear approximation to the mean variance model, many attempts have been made to linearize the portfolio optimization problem. There were introduced several alternative risk measures which are computationally attractive as (for discrete random variables) they result in solving linear programming (LP) problems. The LP solvability is very important for applications to real-life financial decisions where the constructed portfolios have to meet numerous side constraints and take into account transaction costs. The variety of LP solvable portfolio optimization models presented in the literature generates a need for their classification and comparison. It is the main goal of our work. The first part provides a systematic overview of the LP solvable models with a wide discussion of their theoretical properties. The second part will present results of an experimental comparison of the models. Keywords: Portfolio optimization, mean risk model, linear programming University of Brescia, Department of Electronics for Automation, via Branze 38, 25123 Brescia, Italy Warsaw University of Technology, Institute of Control and Computation Engineering, Nowowiejska 15/19, 00-665 Warsaw, Poland University of Brescia, Department of Quantitative Methods, C. da S.Chiara 48/B, 25122 Brescia, Italy

1 Introduction The portfolio optimization problem considered in this paper follows the original Markowitz formulation and is based on a single period model of investment. At the beginning of a period, an investor allocates the capital among various securities, thus assigning a nonnegative weight (share of the capital) to each security. During the investment period, a security generates a random rate of return. This results in a change of capital invested (observed at the end of the period) which is measured by the weighted average of the individual rates of return. Let J = {1, 2,...,n} denote a set of securities considered for an investment. For each security j J, its rate of return is represented by a random variable R j with a given mean µ j = E{R j }. Further, let x =(x j ) j=1,2,...,n denote a vector of decision variables x j expressing the weights defining a portfolio. To represent a portfolio, the weights must satisfy a set of constraints that form a feasible set P. The simplest way of defining a feasible set is by a requirement that the weights must sum to one, i.e. n j=1 x j =1andx j 0forj =1,...,n. Hereafter, it is assumed that P is a general LP feasible set given in a canonical form as a system of linear equations with nonnegative variables. Each portfolio x defines a corresponding random variable R x = n j=1 R jx j that represents the portfolio rate of return. The mean rate of return for portfolio x is given as: µ(x) =E{R x } = n µ j x j. Following the seminal work by Markowitz [14], the portfolio optimization problem is modeled as a mean risk bicriteria optimization problem where µ(x) is maximized and some risk measure ϱ(x) is minimized. In the original Markowitz model [14] the risk is measured by the standard deviation or variance: σ 2 (x) =E{(µ(x) R x ) 2 }. Several other risk measures have been later considered thus creating the entire family of mean risk (Markowitz-type) models. While the original Markowitz model forms a quadratic programming problem, following Sharpe [26], many attempts have been made to linearize the portfolio optimization procedure (c.f., [29] and references therein). The LP solvability is very important for applications to real-life financial decisions where the constructed portfolios have to meet numerous side constraints (including the minimum transaction lots [13]) and to take into account transaction costs [7, 10]. Certainly, in order to guarantee that the portfolio takes advantage of diversification, no risk measure can be a linear function of x. Nevertheless, a risk measure can be LP computable in the case of discrete random variables, i.e., in the case of returns defined by their realizations under the specified scenarios. We will consider T scenarios with probabilities p t (where t =1,...,T). We will assume that for each random variable R j its realization r jt under the scenario t is known. Typically, the realizations are derived from historical data treating T historical periods as equally probable scenarios (p t =1/T ). The realizations of the portfolio return R x are given as y t = n j=1 r jtx j and the expected value µ(x) can be computed as: n µ(x) = y t p t = r jt x j p t. Similarly, several risk measures can be LP computable with respect to the realizations y t.the mean absolute deviation was very early considered in the portfolio analysis ([27] and references j=1 j=1 1

therein) while quite recently Konno and Yamazaki [9] presented and analyzed the complete portfolio LP solvable optimization model based on this risk measure the so-called MAD model. Yitzhaki [31] introduced the mean risk model using Gini s mean (absolute) difference as the risk measure. For a discrete random variable represented by its realizations y t, the Gini s mean difference is LP computable (when minimized). Recently, Young [32] analyzed the LP solvable portfolio optimization model based on risk defined by the worst case scenario (minimax approach), while Ogryczak [18] introduced the multiple criteria LP model covering all the above as special aggregation techniques. The Markowitz model is frequently criticized as not consistent with axiomatic models of preferences for choice under risk [24]. Models consistent with the preference axioms are based on the relations of stochastic dominance or on expected utility theory [30, 12]. The relations of stochastic dominance allow a pairwise comparison of given portfolios but do not offer any computational recipe to analyze the portfolio selection problem. The expected utility theory, when used for the portfolio selection problem, is restrictive in modeling preferences of investors. If the rates of return are normally distributed, then most of the LP computable risk measures become proportional to the standard deviation σ(x) [11, pp. 1216 1217]. Hence the corresponding LP solvable mean risk models are then equivalent to the Markowitz mean variance model. However, the LP solvable mean risk models do not require any specific type of return distributions. Moreover, opposite to the mean variance approach, for general random variables some consistency with the stochastic dominance relations was shown for the Gini s mean difference [31], for the MAD model [20] and for many other LP solvable models as well [18]. It is often argued that the variability of the rate of return above the mean should not be penalized since the investors are concerned of an underperformance rather than the overperformance of a portfolio. This led Markowitz [15] to propose downside risk measures such as (downside) semivariance to replace variance as the risk measure. Consequently, one observes growing popularity of downside risk models for portfolio selection [28]. Some authors pointed out that the MAD model opens up opportunities for more specific modeling of the downside risk [5, 29]. In fact, most of the LP solvable models may be viewed as based on some downside risk measures. Moreover, the models may be extended with some piece-wise linear penalty (risk) functions to provide opportunities for more specific modeling of the downside risk [4, 8, 16]. The variety of LP solvable portfolio optimization models presented in the literature generates a need for their classification and comparison. This is the major goal of this paper. We provide a systematic overview of the models with a wide discussion of their theoretical properties. The paper is organized as follows. In the next section we show how various LP computable performance measures can be derived from shortfall criteria related to the stochastic dominance. Section 3 gives a detailed revue of the LP solvable portfolio optimization models we examine. Finally, in Section 4 some concluding remarks are stated. 2 Shortfall criteria and performance measures In this section we first recall the concepts of shortfall criteria and stochastic dominance. Then, we show how various possible portfolio performance measures can be derived from shortfall criteria and that some are consistent with the stochastic dominance relations. Some of the performance measures are risk measures (to be minimized) and some are safety measures (to be maximized). We also show how these measures become LP computable in the case of returns 2

defined by discrete random variables. 2.1 Shortfall criteria and stochastic dominance The notion of risk is related to a possible failure of achieving some targets. It was formalized as the so-called safety-first strategies [25] and later led to the concept of below-target risk measures [6] or shortfall criteria. The simplest shortfall criterion for the specific target value τ is the mean below-target deviation δ τ (x) =E{max{τ R x, 0}} (1) which, when minimized, is LP computable for returns represented by their realizations y t as: δ τ = d t p t subject to d t τ y t, d t 0 for t =1,...,T. We show that the concept of mean below-target deviation is related to the second degree stochastic dominance relation [30] which is based on an axiomatic model of risk-averse preferences [24]. In stochastic dominance, uncertain returns (random variables) are compared by pointwise comparison of functions constructed from their distribution functions. The first function F x (1) is given as the right-continuous cumulative distribution function of the rate of return F x (1) (η) =F x (η) =P{R x η} and it defines the weak relation of the first degree stochastic dominance (FSD) as follows : R FSD R F R (η) F R (η) for all η. The second function is derived from the first as: F x (2) (η)= η F x (ξ) dξ for real numbers η, and defines the (weak) relation of second degree stochastic dominance (SSD): R x SSD R x F (2) x (η) F (2) x (η) for all η. We say that portfolio x dominates x under the SSD (R x SSD R x ), if F (2) x (η) F (2) x (η) for all η, with at least one strict inequality. A feasible portfolio x 0 Pis called SSD efficient if there is no x Psuch that R x SSD R x 0.IfR x SSD R x, thenr x is preferred to R x within all risk-averse preference models where larger outcomes are preferred. Note that the SSD relation covers increasing and concave utility functions, while the first stochastic dominance is less specific as it covers all increasing utility functions [12], thus neglecting a risk averse attitude. It is therefore a matter of primary importance that a model for portfolio optimization be consistent with the SSD relation, in the sense that R x SSD R x implies that the performance measure in x takes a value not worse than (lower than or equal to, in the case of a risk measure) in x. The consistency with the SSD relation implies that an optimal portfolio is SSD efficient. Function F x (2), used to define the SSD relation, can also be presented as follows [20]: F x (2) (η)=p{r x η}e{η R x R x η} = E{max{η R x, 0}} = δ η (x). 3

Hence, the SSD relation can be seen as a dominance for mean below-target deviations from all possible targets. The mean below-target deviation from a specific target (1) represents only a single criterion. One may consider several, say m, targets τ 1 >τ 2 >... > τ m and use the weighted sum of the shortfall criteria as a risk measure { m m m } w k δτk (x)= w k E{max{τ k R x, 0}} = E w k max{τ k R x, 0} (2) k=1 k=1 where w k (for k =1,...,m) are positive weights which maintains LP computability of the measure (when minimized). Actually, the measure (2) can be interpreted as a single mean below-target deviation applied with a penalty function: E{u(max{τ 1 R x, 0})} where u is increasing and convex piece-wise linear penalty function with breakpoints b k = τ 1 τ k and slopes s k = w 1 +...+ w k, k =1,...,m. Such a piece-wise linear penalty function is used in the Russel-Yasuda-Kasai financial planning model [4] to define the corresponding risk measure. 2.2 MAD and downside versions When an investment situation involves minimal acceptable returns, then the below-target deviation and its extensions are considered to be good risk measures [6]. However, when the mean portfolio return is used as a target, then in (1) the mean µ(x) can be used instead of the fixed target τ. This results in the risk measure known as the downside mean semideviation from the mean k=1 δ(x)=e{max{µ(x) R x, 0}} = F (2) x (µ(x)). (3) The downside mean semideviation is always equal to the upside one and therefore we refer to it hereafter as to the mean semideviation. The mean semideviation is a half of the mean absolute deviation from the mean (see [29]), i.e. δ(x) =E{ R x µ(x) } =2 δ(x). Hence the corresponding mean risk model is equivalent to the MAD model. For a discrete random variable represented by its realizations y t, the mean semideviation (3), when minimized, is LP computable as: δ = d t p t subject to d t µ(x) y t,d t 0 t =1,...,T. Due to the use of distribution dependent target value µ(x), the mean semideviation cannot be directly considered an SSD shortfall criterion. However, as shown by Ogryczak and Ruszczyński [20], the mean semideviation is closely related to the graph of F x (2). The function F x (2) is continuous, convex, nonnegative and nondecreasing. The graph F x (2) (η), referred to as the Outcome Risk (O R) diagram, has two asymptotes which intersect at the point (µ(x), 0) (Figure 1). Specifically, the η-axis is the left asymptote and the line η µ(x) is the right asymptote. In the case of a deterministic (risk-free) return (R x = µ(x)), the graph of F x (2) (η) coincides with the asymptotes, whereas any uncertain return with the same expected value µ(x) yields a graph above (precisely, not below) the asymptotes. The space between the curve (η, F x (2) (η)), 4

and its asymptotes represents the dispersion (and thereby the riskiness) of R x in comparison to the deterministic return µ(x). Therefore, it is called the dispersion space. The mean semideviation turns out to be the largest vertical diameter of the dispersion space while the variance represents its doubled area [20]. δ(x) µ(x) F x (2) (η) η µ(x) η Figure 1: The O R diagram and the mean semideviation Every shortfall risk measure or, more precisely, every pair of a target value τ and the corresponding downside deviation defines also the quantity of mean below-target underachievement τ δ τ (x)=e{τ max{τ R x, 0}} = E{min{R x,τ}}. The latter portfolio performance measure can be considered a safety measure as the larger values are preferred. In the case of a fixed target τ one gets τ δ τ (x ) τ δ τ (x )iff δ τ (x ) δ τ (x ). Hence, the minimization of the mean below-target deviation (risk measure) and the maximization of the corresponding mean below-target underachievement (safety measure) are equivalent. The latest property is no longer valid when µ(x) is used as the target. One may introduce the safety measure of mean downside underachievement µ(x) δ(x)=e{µ(x) max{µ(x) R x, 0}} = E{min{R x,µ(x)}} (4) but the minimization of the mean semideviation is, in general, not equivalent to the maximization of the mean downside underachievement. Note that, as shown in [20], R x SSD R x implies the inequality µ(x ) δ(x ) µ(x ) δ(x ) while the corresponding inequality on the mean semideviations δ(x ) δ(x ) may not be valid. Thus, the mean downside underachievement is consistent with the SSD relation, while the consistency is not guaranteed for the mean semideviation. For better modeling of the downside risk, one may consider a risk measure defined by the mean semideviation applied with a piece-wise linear penalty function [8] to penalize larger downside deviations. It turns out, however, that for maintaining both the LP computability and SSD consistency [16], the breakpoints (or additional target values) must be located at the corresponding mean downside underachievements (4). Namely, when using m distribution dependent targets µ 1 (x) =µ(x), µ 2 (x),..., µ m (x) and the corresponding mean semideviations δ 1 (x) = δ(x), δ 2 (x),..., δ m (x) defined recursively according to the formulas: k 1 δ k (x) = E{max{µ k (x) R x, 0}} = E{max{µ(x) δ i (x) R x, 0}} 5 i=1

µ k+1 (x) = µ k (x) δ k (x)=µ(x) k δ i (x) =E{min{R x,µ k (x)}}, one may combine the semideviations by the weighted sum to the measure δ (m) (x) = i=1 m w k δk (x) 1 = w 1 w 2 w m 0 (5) k=1 as in the m MAD model [16]. Actually, the measure can be interpreted as a single mean semideviation (from the mean) applied with a penalty function: δ(m) (x) =E{u(max{µ(x) R x, 0})} where u is increasing and convex piece-wise linear penalty function with breakpoints b k = µ(x) µ k (x) andslopess k = w 1 +...+ w k, k =1,...,m. Therefore, we will refer to the measure δ (m) (x) astothemean penalized semideviation. The mean penalized semideviation (5) defines the corresponding safety measure µ(x) δ (m) (x) which may be expressed directly as the weighted sum of the mean downside underachievements µ k (x): (m) µ(x) δ w (x)=(w 1 w 2 )µ 2 (x)+(w 2 w 3 )µ 3 (x)+...+(w m 1 w m )µ m (x)+w m µ m+1 (x) (6) where the weights coefficients are nonnegative and total to 1. This safety measure was shown [16] to be SSD consistent in the sense that R x SSD R x implies µ(x ) δ w (m) (x ) µ(x ) δ w (m) (x ). 2.3 Minimax and the worst conditional expectation For a discrete random variable represented by its realizations y t,theworst realization M(x) = min y t (7),...,T is a well appealing safety measure, while the maximum (downside) semideviation (x)=µ(x) M(x) = max (µ(x) y t) (8),...,T represents the corresponding risk measure. The latter is well defined in the O R diagram (Fig. 1) as it represents the maximum horizontal diameter of the dispersion space. The measure M(x) is known to be SSD consistent and it was applied to portfolio optimization by Young [32]. A natural generalization of the measure M(x) is the worst conditional expectation defined as the mean of the specified size (quantile) of worst realizations. For the simplest case of equally probable scenarios (p t =1/T ), one may define the worst conditional expectation M k T (x) asthe mean return under the k worst scenarios. In general, the worst conditional expectation and the worst conditional semideviation for any real parameter 0 <β 1(replacingthequotientk/T) are defined as and M β (x) = 1 β β 0 F ( 1) x (α)dα for 0 <β 1 (9) β (x)=µ(x) M β (x) for 0 <β 1 (10) where F x ( 1) (p) =inf{η : F x (η) p} is the left-continuous inverse of the cumulative distribution function F x. For any 0 <β 1 the conditional worst realization M β (x) is an SSD consistent 6

measure. Actually, the conditional worst expectations provide an alternative characterization of the SSD relation [19, 21] in the sense of the following equivalence: R x SSD R x M β (x ) M β (x ) for all 0 <β 1. (11) Note that M 1 (x) =µ(x) andm β (x) tends to M(x) forβ approaching 0. The worst conditional expectation is closely related to the so-called Tail or Conditional Value-at-Risk (CVaR) which was recently considered in the portfolio optimization [23]. By the theory of convex conjugent (dual) functions [22], the worst conditional expectation may be defined by optimization [21]: M β (x) = max η R [ η 1 β F x (2) (η) ] =max η R [ η 1 ] β E{max{η R x, 0}} where η is a real variable taking the value of β quantile Q β (x) at the optimum. Hence, the worst conditional expectations and the corresponding worst conditional semideviations express the results of the O R diagram analysis according to a slant direction defined the slope β (Fig. 2). (12) F x ( 2) (β) F x (2) (η) M β (x) Q β (x) µ(x) slope 1 slope β η Figure 2: Quantile safety measures in the O-R diagram For a discrete random variable represented by its realizations y t, problem (12) becomes an LP. Thus [ ] M β (x) = max η 1 d t β p t s.t. d t η y t, d t 0 for t =1,...,T (13) whereas the worst conditional semideviations may be computed as the corresponding differences from the mean ( β (x) =µ(x) M β (x)) or directly as: β (x) = min (d + t + 1 β β d t )p t s.t. d t d + t = η y t, d + t,d t 0 for t =1,...,T where η is an auxiliary (unbounded) variable. Note that for β =0.5 one has 1 β = β. Hence, 0.5 (x) represents the mean absolute deviation from the median, the risk measure suggested by 7

Sharpe [27]. The LP problem for computing this measure takes the form: 0.5 (x) = min (d t + d + t )p t s.t. d t d + t = η y t, d + t,d t 0 for t =1,...,T. 2.4 Gini s mean difference Yitzhaki [31] introduced the mean-risk model using Gini s mean (absolute) difference as the risk measure. For a discrete random variable represented by its realizations y t,thegini s mean difference Γ(x) = 1 y t y t p t p t (14) 2 t =1 t =1 is obviously LP computable (when minimized). In the case of equally probable T scenarios with p t =1/T the Gini s mean difference may be expressed as the weighted average of the worst conditional semideviations k (x)fork =1,...,T T [18]. Exactly, using weights w k =(2k)/T 2 for k =1, 2,...,T 1 andw T =1/T =1 T 1 k=1 w k, one gets Γ(x) = T k=1 w k k (x). On the other hand, for general discrete distributions, directly T from the definition (14): Γ(x) = (y t y t )p t p t = F x (2) (y t )p t = δ yt (x)p t. t =1 t :y t <y t Hence, Γ(x) can be interpreted as the weighted sum of multiple mean below-target deviations (2) but both the targets and the weights are distribution dependent. This corresponds to an interpretation of Γ(x) as the integral of F x (2) with respect to the probability measure induced by R x [21]. Thus although not representing directly any shortfall criterion, the Gini s mean difference is a combination of the basic shortfall criteria. Note that the Gini s mean difference defines the corresponding safety measure µ(x) Γ(x) =E{R x R x } (15) which is the expectation of the minimum of two i.i.d.r.v. R x [31] thus representing the mean worse return. This safety measure is SSD consistent [31, 21] in the sense that R x SSD R x implies µ(x ) Γ(x ) µ(x ) Γ(x ). 3 LP solvable models In this section we present the complete set of LP solvable models we consider and their LP formulation. 3.1 Risk and safety measures Following Markowitz [14], the portfolio optimization problem is modeled as a mean risk bicriteria optimization problem max{[µ(x), ϱ(x)] : x P} (16) 8

where the mean µ(x) is maximized and the risk measure ϱ(x) is minimized. A feasible portfolio x 0 P is called efficient solution of problem (16) or µ/ϱ efficient portfolio if there is no x P such that µ(x) µ(x 0 )andϱ(x) ϱ(x 0 ) with at least one inequality strict. The original Markowitz model [14] uses the standard deviation σ(x) as the risk measure. As shown in the previous section, several other risk measures may be used instead of the standard deviation thus generating the corresponding LP solvable mean-risk models. In this paper we restrict our analysis to the risk measures which, similar to the standard deviation, are shift independent dispersion parameters. Thus, they are equal to 0 in the case of a risk free portfolio and take positive values for any risky portfolio. This excludes the direct use of the mean belowtarget deviation (1) and its extensions with penalty functions (2). Nevertheless, as shown in Section 2, there is a gamut of LP computable risk measures fitting the requirements. In Section 2 we have seen that in the literature some of the LP computable measures are dispersion type risk measures and some are safety measures, which, when embedded in an optimization model, are maximized instead of being minimized. Moreover, we have shown that each risk measure ϱ(x) has a well defined corresponding safety measure µ(x) ϱ(x) and viceversa. Although the risk measures are more natural, due to the consolidated familiarity with Markowitz model, we have seen that the safety measures, contrary to the dispersion type risk measures, satisfy axioms of the so-called coherent risk measurement [1]. Moreover, they are SSD consistent in the sense that R x SSD R x implies µ(x ) ϱ(x ) µ(x ) ϱ(x ) [16, 20, 19, 31, 32]. The practical consequence of the lack of SSD consistency can be illustrated by two portfolios x and x (with rate of return given in percents): P{R x = ξ} = { 1, ξ =1.5 0, otherwise 1/2, ξ =3.5 P{R x = ξ} = 1/2, ξ =4.5 0, otherwise Note that the risk free portfolio x with the guaranteed result 1.5 is obviously worse than the risky portfolio x giving 3.5 or 4.5. Certainly, in all models consistent with the preference axioms [1, 12, 30], portfolio x is dominated by x,inparticularr x SSD R x. When a dispersion type risk measure ϱ(x) is used, then both the portfolios may be efficient in the corresponding mean risk model since for each such measure ϱ(x ) > 0 while ϱ(x ) = 0. This is a common flaw of all Markowitz-type mean risk models where risk is measured with some dispersion measure. It is interesting to note that, in order to overcome this weakness of the Markowitz model already in 1964 Baumol [2] suggested to consider a safety measure, which he called the expected gain-confidence limit criterion, µ(x) λσ(x) to be maximized instead of the σ(x) minimization. Thus, on the basis of the above remarks, for each risk measure, it is reasonable to consider also an alternative mean-safety bicriteria model: max{[µ(x),µ(x) ϱ(x)] : x P}. (17) Summarizing, in the computational part of this paper we will consider the full set of risk and safety measures presented in Table 1. Note that the MAD model was first introduced [9] with the risk measure of mean absolute deviation δ(x) whereas the mean semideviation δ(x) we consider is half of it. This is due to the fact that the resulting optimization models are equivalent and that the model with the semideviation is more efficient (see [29]). For the MAD model, the safety measure represents 9

Table 1: Risk and safety measures Risk measure ϱ(x) Safety measure µ(x) ϱ(x) MAD model [9] δ(x) (3) E{min{Rx,µ(x)}} (4) m MAD model [16] δ(m) (x) (5) µ(x) δ (m) (x) (6) Minimax model [32] (x) (8) M(x) (7) CVaR model [23] β (x) (9) M β (x) (10) GMD model [31] Γ(x) (14) E{R x R x} (15) the mean downside underachievement. For the m MAD model the two measures represent the mean penalized semideviation and the weighted sum of the mean downside underachievements, respectively. The Minimax model was considered and tested [32] with the safety measure of the worst realization M(x) but it was also analyzed with the maximum semideviation (x) [18]. The CVaR model was considered with the safety measure of the worst conditional expectation [23]. The risk measure represents the worst conditional semideviation. Yitzhaki [31] introduced the GMD model with the Gini s mean difference Γ(x) but he also analyzed the model implementation with the corresponding safety measure of the mean worse return E{R x R x }. As shown in the previous section, all the risk measures we consider may be derived from the shortfall criteria of SSD. However, they are quite different in modeling of the downside risk aversion. Definitely, the strongest in this respect is the maximum semideviation (x) used in the Minimax model. It is a strict worst case measure where only the worst scenario is taken into account. The CVaR model allows to extend the analysis to a specified β quantile of the worst returns. The measure of worst conditional semideviation β (x) offers a continuum of models evolving from the strongest downside risk aversion (β close to 0) to the complete risk neutrality (β = 1). The MAD model with risk measured by the mean (downside) semideviation from the mean is formally a downside risk model. However, due to the symmetry of mean semideviations from the mean [20], it is equally appropriate to interpret the MAD model as an upside risk model. Actually, the m MAD model has been introduced to incorporate downside risk modeling capabilities into the MAD model. The Gini s mean difference although related to all the worst conditional semideviations is, similar to the mean absolute deviation, a symmetric risk measure (in the sense that for R x and R x it has exactly the same value). Note that having µ(x ) µ(x )andϱ(x ) ϱ(x ) with at least one inequality strict, one gets µ(x ) ϱ(x ) >µ(x ) ϱ(x ). Hence, a portfolio dominated in the mean-risk model (16) is also dominated in the corresponding mean-safety model (17). In other words, the efficient portfolios of problem (17) form a subset of the entire µ/ϱ efficient set. Due to the SSD consistency of the safety measures, except for portfolios with identical mean and risk measure, every portfolio belonging to this subset is SSD efficient. Although very important, the SSD efficiency is only a theoretical property. For specific types of distributions or feasible sets the subset of portfolios with guaranteed SSD efficiency may be larger [19, 20]. Hence, the mean-safety model (17) may be too restrictive in some practical investment decisions. In conclusion, we believe the computational testing of the models resulting from all the risk and safety measures summarized in Table 1 is an important part of the models comparison. 10

3.2 How to solve bicriteria problems In order to compare on real-life data the performance of various mean risk models, one needs to deal with specific investor preferences expressed in the models. There are two ways of modeling risk averse preferences and therefore two major approaches to handle bicriteria mean risk problems (16). First, having assumed a trade-off coefficient λ between the risk and the mean, the so-called risk aversion coefficient, one may directly compare real values µ(x) λϱ(x) and find the best portfolio by solving the optimization problem: max {µ(x) λϱ(x):x P}. (18) Various positive values of parameter λ allow to generate various efficient portfolios. By solving the parametric problem (18) with changing λ>0onegets the so-called critical line approach [15]. Due to convexity of risk measures ϱ(x) with respect to x, λ>0provides the parameterization of theentiresetoftheµ/ϱ efficient portfolios (except of its two ends which are the limiting cases). Note that (1 λ)µ(x)+λ(µ(x) ϱ(x)) = µ(x) λϱ(x). Hence, bounded trade-off 0 <λ 1inthe Markowitz-type mean risk model (16) corresponds to the complete weighting parameterization of the model (17). The critical line approach allows to select an appropriate value of the risk aversion coefficient λ and the corresponding optimal portfolio through a graphical analysis in the mean-risk image space. Unfortunately, in practical investment situations, the risk aversion coefficient does not provide a clear understanding of the investor preferences. The commonly accepted approach to implementation of the Markowitz-type mean risk model is then based on the use of a specified lower bound µ 0 on expected returns which results in the following problem: min{ϱ(x): µ(x) µ 0, x P}. (19) This bounding approach provides a clear understanding of investor preferences and a clear definition of solution portfolios to be used in the models comparison. Therefore, we use the bounding approach (19) in our analysis. Due to convexity of risk measures ϱ(x) with respect to x, by solving the parametric problem (19) with changing µ 0 [ min µ j, max µ j] one gets various efficient portfolios. Actually, j=1,...,n j=1,...,n for µ 0 smaller than the expected return of the MRP (the minimum risk portfolio defined as solution of min ϱ(x)) problem (19) generates always the MRP while larger values of µ 0 provide x P the parameterization of the µ/ϱ efficient set. As a complete parameterization of the entire µ/ϱ efficient set, the approach (19) generates also those portfolios belonging to the subset of efficient solutions of the corresponding mean safety problem (17). The latter correspond to larger values of bound µ 0 exceeding the expected return of the portfolio defined as solution of max[µ(x) ϱ(x)]. Thus, opposite to the critical line approach, having a specified value of x P parameter µ 0 one does not know if the optimal solution of (19) is also an efficient portfolio with respect to the corresponding mean safety model (17) or not. Therefore, when using the bounding approach to the mean risk models (16), we need to consider explicitly a separate problem max{µ(x) ϱ(x): µ(x) µ 0, x P} (20) for the corresponding mean safety model (17). 11

3.3 The LP models We provide here the detailed LP formulations for all the models we have analyzed. For each type of model, the pair of problems (19) and (20), we have analyzed, can be stated as: max{αµ(x) ϱ(x) : µ(x) µ 0, x P} (21) covering the minimization of risk measure ϱ(x) (19) for α = 0 while for α = 1 it represents the maximization of the corresponding safety measure µ(x) ϱ(x) (20). Both optimizations are considered with a given lower bound on the expected return µ(x) µ 0. All the models (21) contain the following linear constraints: x P and z µ 0 (22) n n µ j x j z =0 and r jt x j y t =0 fort =1,...,T (23) j=1 j=1 where z is an unbounded variable representing the mean return of the portfolio x and y t (t = 1,...,T) are unbounded variables to represent the realizations of the portfolio return under the scenario t. In addition to these common variables and constraints, each model contains its specific linear constraints to define the risk or safety measure. MAD models. The standard MAD model [9], when implemented with the mean semideviation as the risk measure (ϱ(x)= δ(x)), leads to the following LP problem: maximize αz z 1 subject to (22) (23) and z 1 p t d 1t = 0 (24) d 1t z + y t 0, d 1t 0 for t =1,...,T (25) where nonnegative variables d 1t represent downside deviations from the mean under several scenarios t and z 1 is a variable to represent the mean semideviation itself. The latter can be omitted by using the direct formula for mean semideviation in the objective function instead of equation (24). The above LP formulation uses T +1variablesandT + 1 constraints to model the mean semideviation. In order to incorporate downside risk aversion by techniques of the m MAD model [16], one needs to repeat constraints of type (24) (25) for each penalty level k =2,...,m. maximize subject to m αz z 1 w k z k k=2 (22) (23), (24) (25) and for k =2,...,m: z k p t d kt =0 k 1 d kt z + z i + y t 0, d kt 0 for t =1,...,T i=1 12

This results in the LP formulation that uses m(t + 1) variables and m(t + 1) constraints to model the m-level penalized mean semideviation. Minimax models. For any 0 <β<1 the CVaR model [23] with ϱ(x) = β (x), due to (12), may be implemented as the following LP problem: maximize y (1 α)z 1 β p t d t subject to (22) (23) and d t y + y t 0, d t 0 for t =1,...,T It is very similar to that of MAD but while defining the downside deviations an independent free variable y is used instead of z representing the mean. Recall that the optimal value of y represents the value of β-quantile. T +1 variablesand T constraints are used here to model the worst conditional semideviation. As the limiting case when β tends to 0 one gets the standard Minimax model [32]. The latter may be additionally simplified by dropping the explicit use of the deviational variables: maximize y (1 α)z subject to (22) (23) and y t y 0 for t =1,...,T thus resulting in T constraints and a single variable used to model the maximum semideviation. GMD model. The model with risk measured by the Gini s mean difference (ϱ(x)=γ(x)) [31], due to the relation Γ(x) =µ(x) E{R x R x }, may be implemented as follows: maximize (α 1)z + subject to T 1 y t p 2 t +2 (22) (23) and u t t y t, u t t y t t =1 t =t +1 u t t p t p t for t =1,...,T 1; t = t +1,...,T where u t t are unbounded variables to represent min{y t,y t }. The above LP formulation uses T (T 1)/2 variables and T (T 1) constraints to model the Gini s mean difference. The direct formulation of the GMD model according to (14) takes the form: maximize αz t =1 t t p t p t d t t subject to (22) (23) and d t t y t y t, d t t 0 for t,t =1,...,T; t t which contains T (T 1) nonnegative variables d t t and T (T 1) inequalities to define them. The symmetry property d t t = d t t is here ignored and, therefore, the number of variables is doubled in comparison to the previous model. However, variables d t t are associated with the singleton coefficient columns. Hence, while solving the dual instead of the original primal, the corresponding dual constraints take the form of simple upper bounds (SUB) [17] which are handled implicitly outside the LP matrix. In other words, the dual contains T (T 1) variables but the number of constraints (excluding the SUB structure) is then proportional to T. Such a dual approach may dramatically improve the LP model efficiency in the case of large number of scenarios. Certainly, one cannot take the advantages of solving dual in the case of some discrete constraints defining the portfolios set P. 13

4 Concluding remarks The classical Markowitz model uses the variance as the risk measure, thus resulting in a quadratic optimization problem. Several alternative risk measures were later introduced which are computationally attractive as (for discrete random variables) they result in solving linear programming (LP) problems. The LP solvability is very important for applications to real-life financial decisions where the constructed portfolios have to meet numerous side constraints and take into account transaction costs. A gamut of LP solvable portfolio optimization models has been presented in the literature thus generating a need for their classification and comparison. In this paper we have provided a systematic overview of the models with a wide discussion of their theoretical properties. We have shown that all the risk measures used in the LP solvable models can be derived from the SSD shortfall criteria. The presented formulations show that all the mean-risk models, we consider, can easily be implemented within the LP methodology. In order to implement the corresponding risk measures, the simplest models require only T (the number of scenarios) auxiliary variables and inequalities. The m MAD model, providing more detailed downside risk modeling capabilities, requires also more complex LP formulations. The number of auxiliary variables and constraints is there multiplied by m (the combination size) but still remain proportional to the number of scenarios T. The Gini s mean difference requires essentially T 2 auxiliary variables and constraints, but taking advantage of the dual formulation allows to reduce the auxiliary structure size. Theoretical properties, although crucial for understanding the modeling concepts, provide only a very limited background for comparison of the final optimization models. Computational results are known only for individual models and not all the models were tested in a real-life decision environment. The second part of this paper will present a comprehensive experimental study comparing practical performances of various LP solvable portfolio optimization models on real-life stock market data. References [1] P. Artzner, F. Delbaen, J.-M. Eber and D. Heath, Coherent measures of risk, Mathematical Finance 9 (1999) 203 228. [2] W.J. Baumol, An expected gain-confidence limit criterion for portfolio selection, Management Science 10 (1964) 174 182. [3] V.S. Bawa, Stochastic Dominance: A Research Bibliography, Management Science 28 (1982) 698 712. [4] D.R. Carino, D.H. Myers and W.T. Ziemba, Concepts, Technical Issues and Uses of the Russel-Yasuda Kasai Financial Planning Model, Operations Research 46 (1998) 450 463. [5] C.D. Feinstein and M.N. Thapa, A Reformulation of a Mean Absolute Deviation Portfolio Optimization Model, Management Science 39 (1993) 1552 1553. [6] P.C. Fishburn, Mean Risk Analysis with Risk Associated with Below Target Returns, The American Economic Review 67 (1977) 116 126. 14

[7] H. Kellerer, R. Mansini and M.G. Speranza, Selecting portfolios with fixed costs and minimum transaction lots, Annals of Operations Research, in print. [8] H. Konno, Piecewise Linear Risk Function and Portfolio Optimization, Journal of the Operations Research Society of Japan 33 (1990) 139 156. [9] H. Konno and H. Yamazaki, Mean absolute deviation portfolio optimization model and its application to Tokyo stock market, Management Science 37 (1991) 519 531. [10] H. Konno and A. Wijayanayake, Portfolio optimization problem under concave transaction costs and minimal transaction unit constraints, Mathematical Programming 89 (2001) 233 250. [11] W.H. Kruskal and J.M. Tanur (Eds.), International Encyclopedia of Statistics, Macmillan, London, 1978. [12] H. Levy, Stochastic dominance and expected utility: survey and analysis, Management Science 38 (1992) 555 593. [13] R. Mansini and M.G. Speranza, Heuristic algorithms for the portfolio selection problem with minimum transaction lots, European Journal of Operational Research 114 (1999) 219 233. [14] H.M. Markowitz, Portfolio selection, Journal of Finance 7 (1952) 77 91. [15] H.M. Markowitz, Portfolio Selection: Efficient Diversification of Investments, John Wiley & Sons, New York, 1959. [16] W. Michalowski and W. Ogryczak, Extending the MAD portfolio optimization model to incorporate downside risk aversion, Naval Research Logistics 48 (2001) 185-200. [17] J.L. Nazareth, Computer Solutions of Linear Programs, Oxford Univ. Press, New York, 1987. [18] W. Ogryczak, Multiple Criteria Linear Programming Model for Portfolio Selection, Annals of Operations Research 97 (2000) 143 162. [19] W. Ogryczak, Risk Measurement: Mean Absolute Deviation versus Gini s Mean Difference. In: Decision Theory and Optimization in Theory and Practice, G. Wanka (ed.), pp. 33 51, Shaker Verlag, Aachen, 2000. [20] W. Ogryczak and A. Ruszczyński, From Stochastic Dominance to Mean Risk Models: Semideviations as Risk Measures, European Journal of Operational Research 116 (1999) 33 50. [21] W. Ogryczak and A. Ruszczyński, Dual Stochastic Dominance and Related Mean Risk Models, Research Report RRR 10 01, RUTCOR, Rutgers University, Piscataway, NJ, 2001. [22] R.T. Rockafellar, Convex Analysis, Princeton Univ. Press, Princeton, NJ, 1970. [23] R.T. Rockafellar and S. Uryasev, Optimization of Conditional Value-at-Risk, Journal of Risk 2 (2000) 21 41. 15

[24] M. Rothschild and J.E. Stiglitz, Increasing Risk: I. A Definition, Journal of Economic Theory 2 (1969) 225 243. [25] A.D. Roy, Safety-first and the holding of assets, Econometrica 20 (1952) 431 449. [26] W.F. Sharpe, A linear programming approximation for the general portfolio analysis problem, Journal of Financial and Quantitative Analysis 6 (1971) 1263 1275. [27] W.F. Sharpe, Mean-Absolute Deviation Characteristic Lines for Securities and Portfolios, Management Science 18 (1971) B1 B13. [28] F.A. Sortino and H.J. Forsey, On the Use and Misuse of Downside Risk, Journal of Portfolio Management, Winter 1996 35 42. [29] M.G. Speranza, Linear programming models for portfolio optimization, Finance 14 (1993) 107 123. [30] G.A Whitmore and M.C. Findlay (eds.), Stochastic Dominance: An Approach to Decision Making Under Risk, D.C.Heath, Lexington MA 1978. [31] S. Yitzhaki, Stochastic dominance, mean variance, and Gini s mean difference, American Economic Revue 72 (1982) 178 185. [32] M.R. Young, A minimax portfolio selection rule with linear programming solution, Management Science 44 (1998) 673 683. [33] S.A. Zenios and P. Kang, Mean Absolute Deviation Portfolio Optimization for Mortgage- Backed Securities, Annals of Operations Research 45 (1993) 433 450. 16

Latest changes 1. The front page reedited and the last sentences of the abstract adjusted to the explicite use of Part I. 2. Figures moved into the text. 3. References to two papers on m-mad reduced to only one in NRL. 4. A tentative beginning of the concluding section added. 5. computationality replaced with computability