Buered Probability of Exceedance: Mathematical Properties and Optimization Algorithms
|
|
- Berniece Waters
- 7 years ago
- Views:
Transcription
1 Buered Probability of Exceedance: Mathematical Properties and Optimization Algorithms Alexander Mafusalov, Stan Uryasev RESEARCH REPORT Risk Management and Financial Engineering Lab Department of Industrial and Systems Engineering 303 Weil Hall, University of Florida, Gainesville, FL s: First draft: October 2014, This draft: October 2014 Correspondence should be addressed to: Stan Uryasev Abstract This paper introduces a new probabilistic characteristic called buered probability of exceedance (bpoe). This characteristic is an extension of so-called buered probability of failure and it is equal to one minus superdistribution function. Paper provides ecient calculation formulas for bpoe. bpoe is proved to be a quasi-convex function of random variable w.r.t. the regular addition operation and a concave function w.r.t. the mixture operation; it is a monotonic function of random variable. bpoe is proved to be a strictly decreasing function of the parameter on the interval between the mathematical expectation and the essential supremum. Multiplicative inverse of the bpoe is proved to be a convex function of parameter, and a piecewise-linear function in the case of discretely distributed random variable. Minimization of the bpoe can be reduced to a convex program for a convex feasible region and to LP for a polyhedral feasible region. A family of bpoe minimization problems and family of the corresponding CVaR minimization problems share the same frontier of optimal solutions and optimal values. Keywords: probability of failure, probability of exceedance, buered probability of failure, superdistribution, superquantile, Conditional Value-at-Risk, CVaR, parametric simplex method 1. Introduction This paper uses the notation CVaR α (X) for conditional-value-at-risk (CVaR) for a random variable X and a condence level α [0, 1], explored in [5]. To have a more consice notation, an alternative name superquantile q α (X) is used, similar to a regular quantile q α (X). That is, q α (X) = CVaR α (X). Notation q(α; X) is used to present superquantile q α (X) = q(α; X) as a function of parameter α. For example, q 1 (x; X) should be interpreted as an inverse function of superquantile as a function of α. Probability of exceedance is dened as p x (X) = P (X > x) = 1 F X (x), where F X (x) is a distribution function. In engineering applications it is usual to see an optimization problem with probability of exceedance in constraints or as an objective. Paper [4] suggests as an alternative to the probability of failure, which is p(x) = P (X > 0), the
2 buered probability of failure, which is a value p(x) such that q p(x) (X) = 0. This paper denes buered probability of exceedance p x (X) in a way that p 0 (X) = p(x) and p x (X) = p 0 (X x). To dene buered probability of exceedandce, we introduce the following mathematical notions from paper [3]. For any random variable X with distribution function F X (x) there is an auxilary random variable X = q(f X (X); X) with distribution function F X(x) = F X (x), called superdistribution function, and 1, for x sup X; F X (x) = q 1 (x; X), for EX < x < sup X; 0, otherwise, where q 1 (x; X) is an inverse of the function q(α; X) as a function of α. Denition 1. For a random variable X and x R, buered probability of exceedance is dened as follows 0, for x sup X; p x (X) = 1 F X (x) = 1 q 1 (x; X), for EX < x < sup X; 1, otherwise. Book [9] considers Chebyshev-type family of inequalities with CVaR deviation and shows that the tightest inequality in the family is obtained for α = p x (X), and the tightest inequality itself reduces to p x (X) p x (X). (1) Inequality (1) is similar to q α (X) q α (X). Inequality (1) is one of the motivations for introducing buered probability of exceedance instead of regular probability of exceedance. Paper [4] uses inequality (1) to argue that the buered probability of failure is a conservative estimate of the probability of failure. Similarly, buered probability of exceedance is a conservative estimate of the probability of exceedance. Section 2 proves several formulas for ecient calculation of p x (X). Section 3.1 investigates mathematical properties of p x (X) w.r.t. parameter x. Section 3.2 establishes mathematical properties of p x (X) w.r.t. random variable X. Section 4 studies minimization of p x (X) over a feasible region X X. 2. Calculation Formulas for BPOE Note that, since q α (X) x = q α (X x) for any constant x, see e.g. [6], then X x = X x and F X (x) = F X x (0). Therefore, p x (X) = p 0 (X x). The following proposition is a slightly modied proposition from paper [1], studying applications of buered probability of exceedance in classication. Proposition 1. For a random variable X and x R, buered probability of exceedance equals { 0, if x = sup X; p x (X) = (2) min a 0 E[a(X x) + 1] +, otherwise. 2
3 Proof. In the denition of buered probability of exceedance we have three cases: 1. p x (X) = 1 q 1 (x; X) when EX < x < sup X, 2. p x (X) = 1 when x < EX or x = EX < sup X, 3. p x (X) = 0 when x sup X. Let us prove the proposition case by case. 1. Let EX < x < sup X, and take x = 0. Since q α (X) is a strictly increasing function of α on α [0, 1 P (X = sup X)], then equation q p (X) = 0 has a unique solution p for EX < x < sup X. Then, p 0 (X) = p such that min c c + 1 p E[X c]+ = 0. Since q α (X) is an increasing function of parameter α, then we can reformulate p 0 (X) = min p p such that min c c + 1 p E[X c]+ 0. Therefore, p 0 (X) = min p,c p s.t. c + 1 p E[X c]+ 0. Optimal c < 0, since c c + 1 p E[X c ] + 0, and c = 0 implies sup X 0, which is not the case we consider. Therefore, p 0 (X) = min p,c s.t. p p c c + E [ 1 c X c ] + 0. c Since c < 0, then c c and, therefore, 1 = 1. Further, denoting a =, we have c p 0 (X) = min p,a>0 p s.t. E[aX + 1] + p. p 0 (X) = min a 0 E[aX + 1]+. Note that change a > 0 to a 0 includes value 1 to the feasible region, which does not aect the case considered. Finally, since p x (X) = p 0 (X x), then p x (X) = min a 0 E[a(X x) + 1]+. 2. When EX x, we have E[a(X x) + 1] + ae(x x) Note also that E[a(X x) + 1] + = 1 for a = 0. Therefore, min a 0 E[a(X x) + 1] + = For x = sup X, by the formula, p x (X) = 0. Consider x > sup X, i.e., X x ε < 0. Taking a = 1 ε makes a(x x) 1, therefore, min a 0 E[a(X x) + 1] + = 0. Corollary 1. For EX < x < sup X, p x (X) = 1 q 1 (x; X) = min c<x 3 E[X c] +. (3) x c
4 Furthermore, for x = q α (X), where α (0, 1), it is valid that and, consequently, q α (X) arg min c<x E[X c] +, x c p x (X) = E[X q α(x)] + q α (X) q α (X). Proof. Since EX < x < sup X, then q 1 (x; X) (0, 1), therefore, a = 0 is not optimal for min a 0 E[a(X x)+1] +. Therefore, change of variable a 1 leads to an equivalent x c program: [ ] + 1 min E[a(X x) + a 0 1]+ = min E E[X c] + (X x) + 1 = min. c<x x c c<x x c Note that if x = q α (X), then p x (X) = 1 q 1 (x; X) = 1 α. Since q α (X) = q α (X) α E[X q α] +, then p x (X) = 1 α = E[X q α(x)] + q α (X) q α (X), that is, q α (X) arg min c<x E[X c] + x c. Let X be a discretely distributed random variable with atoms {x i } N i=1, and probabilities {p i } N i=1, where x i x i+1, i = 1,..., N 1, and N is either nite or N =. For condence levels α j = j i=1 pi, where j = 0,..., N, let us denote corresponding superquantiles x j = N i=j+1 xi p i /(1 α j ), with x N = x N for nite N and x N = lim i x i for N =. Then, p x (X) = 1 for x x 0 = EX, p x (X) = 0 for x x N = sup X, and p x j(x) = 1 α j for j = 0,..., N 1. Corollary 2. p x (X) = E[X xj+1 ] + x x j+1 = for x j < x < x j+1, where j = 0,..., N 1. N i=j+1 pi [x i x j+1 ] + x x j+1, (4) Proof. Note that for x j < q α (X) < x j+1 we have α j < α < α j+1, therefore, q α (X) = x j+1. Therefore, formula (4) is implied by Corollary 1. Buered probability is calculated with a simple formula for a set of specic values x j, j = 0,..., N. The following proposition presents a formula for calculation bpoe in intermediate values, i.e., for x such that x j < x < x j+1. Such value x can be also represented as a weighted combination of values x j and x j+1 : x = µ x j + (1 µ) x j+1, for some µ (0, 1). Corollary 3. For µ (0, 1) and j = 0,..., N 1 ( p(µ x j + (1 µ) x j+1 µ ; X) = p( x j ; X) + 1 µ ) 1, p( x j+1 ; X) i.e. 1/ p x (X) is a piecewise-linear function of x. 4
5 Proof. Corollary 2 implies 1 p x (X) = x x j+1 N i=j+1 pi [x i x j+1 ] +, for x j < x < x j+1, where j = 0,..., N 1. Therefore, since p x (X) is continuous for x [EX, sup X), see Proposition 2, then 1/ p x (X) is a piecewise-linear function of x. 3. Mathematical Properties of bpoe 3.1. Properties of bpoe w.r.t. Parameter x Proposition 2. Distribution F X(x) = F X (x) has no more than one atom at sup X = sup X with probability P ( X = sup X) = P (X = sup X). Proof. Note that if for α 1 < α 2 we have q α1 (X) = q α2 (X), then, by denition of CVaR, { min c + 1 } { E[X c] + = min c + 1 } E[X c] +. c 1 α 1 c 1 α 2 For each value of c, if c < sup X, then E[X c] + > 0 and c α 1 E[X c] + < c + { } 1 1 α 2 E[X c] +. Therefore, arg min c c α 1 E[X c] + = sup X. It proves that q α (X) as a function of α can have only one interval of constancy, which is for α [1 P (X = sup X), 1]. For the interval α [0, 1 P (X = sup X)] the function q α (X) is strictly increasing in α. This implies that if superdistribution has an atom, then there are two possible locations. The rst case, x = EX, but q 0 (X) = EX, therefore, FX (EX 0) = F X (EX + 0) = 0, and EX is a continuity point of superdistribution. The second case, x = sup X, then lim x sup X 0 FX (x) = 1 P (X = sup X). Since F X (sup X) = 1, and X X, see [3], then sup X = sup X and P ( X = sup X) = P (X = sup X). Corollary 4. For any random variable X, buered probability of exceedance p x (X) is a continuous strictly decreasing function of x at the interval x [EX, sup X). Proof. bpoe equals 1 q 1 (x; X) for x [EX, sup X). The function q(α; X) is strictly increasing continuous for α [0, 1 P (X = sup X)] (see, e.g., proof of Proposition 2). Therefore, for x (EX, sup X) the function q 1 (x; X) is a strictly increasing continuous function of x. The point x = EX can be added to the interval of continuity, since we have proved that it is a continuity point of FX (x) = q 1 (x; X). Corollary 5. Buered probability of exceedance p x (X) is a non-increasing right-continuous function of x with no more than one point of discontinuity. Proof. Immediately follows from denition p x (X) = 1 F X (x) and Proposition 2. Proposition 3. Function 1 1 F X (x) = 1 p x (X) is a convex function w.r.t. x. Moreover, it is piecewise-linear for discretely distributed X. 5
6 Proof. Consider interval EX < x < sup X, where formula (3) is valid. Then, 1 p x (X) = 1/ min c<x E[X c] + x c = max c<x x c E[X c] +. Note that since max c<x (x c)/e[x c] + > 0, then max c<x x c E[X c] = max + c<x [x c] + E[X c] = max + c [x c] + E[X c] +. The last expression max c {[x c] + /E[X c] + } is convex over x as a maximum over the family of convex functions of x. p x (X) is a continuous non-increasing function on x (, sup X), therefore, 1/ p x (X) is a continuous non-decreasing function on x (, sup X). Then, extending the interval from (EX, sup X) to (, sup X) does not violate convexity of 1/ p x (X), since 1/ p x (X) = 1, i.e., constant, for x (, EX]. Further extending of the interval from (, sup X) to (, + ), i.e., R, will not violate convexity either, since 1/ p x (X) = + for x sup X. That is, 1/ p x (X) is a convex function of x. Suppose that X is discretely distributed. Again, 1/ p x (X) = 1 for x (, EX], and that is the rst interval of linearity. Consider probability atom with value x which random variable X takes with probability p. Denote α 1 = F X (x ) = P (X < x ), α 2 = F X (x) = P (X x ) = α 1 + p and x i = q α i(x) for i = 1, 2. Then for x 1 < x < x 2 we have x = q α (X) with α (α 1, α 2 ), therefore, q α (X) = x. Applying Corollary 1 we nd that 1/ p x (X) = (x x )/E[X x ] + for x 1 < x < x 2. Therefore, 1/ p x (X) is linear on x 1 < x < x 2. This way, all the atom probability intervals of type (F X (x ), F X (x )) [0, 1] will project into the intervals of type ( x 1 ; x 2 ) (EX, sup X) between corresponding superquantiles, covering all the interval (EX, sup X). Therefore, 1/ p x (X) is a piecewiselinear function on x (, sup X), and 1/ p x (X) = + on x [sup X, + ) Properties of bpoe w.r.t. Random Variable Proposition 4. Buered probability is a closed quasi-convex function of random variable (w.r.t. addition operation), i.e., the set {X p x (X) p} is a closed convex set of random variables for any p R. Furthermore, for p [0, 1), p x (X) p q 1 p (X) x. Proof. If p 1, then the inequality p x (X) p holds for any x and X. Therefore, the level-set {X p x (X) p} is a closed convex set. For p < 0, {X p x (X) p} =. Consider p [0, 1). Suppose p x (X) p, then p x (X) = p ε for some ε 0. Then, either q 1 px(x)(x) = q 1 p+ε (X) = x, therefore, q 1 p (X) x, or sup X x, therefore, q 1 p (X) q 1 (X) x. Conversely, if q 1 p (X) x, then either q 1 p+ε (X) = x for some ε 0, or sup X x. In the rst case, p x (X) = p ε p, and p x (X) p q 1 p (X) x. If sup X x, then p x (X) = 0 p. Function q 1 p (X) is a closed convex function of X, therefore, the set {X q 1 p (X) x} is closed convex. Then, the set {X p x (X) p} is closed convex. 6
7 Example 1. Buered probability of exceedance is not a convex function of random variable (w.r.t. addition operation), i.e., in general, p x (λx + (1 λ)y ) λ p x (X) + (1 λ) p x (Y ). Counterexample is as follows. Take x = 0 and { 1, with probability 1/2, X = 1, with probability 1/2. Take Y 0, λ = 1/2. Note that p 0 (X) = 1, since q 0 (X) = 0, p 0 (Y ) = 0. Note also that λx + (1 λ)y = X/2, therefore, p 0 (λx + (1 λ)y ) = 1 1/2 = λ p 0 (X) + (1 λ) p 0 (Y ). Denote by B λ the Bernoulli random variable with probability λ being equal to 1, i.e., { 1, with probability λ, B λ = 0, with probability 1 λ. Denote the mixture of random variables with coecient λ as λx (1 λ)y = XB λ + Y (1 B λ ), where B λ is independent of X and Y. In words, a mixture of random veriables with coecient λ is a random variable which takes a value of the rst random variable with probability λ, and a value of the second random variable with probability (1 λ). Mixture operation results from the addition operation over measures. Suppose µ and ν are measures. Then, scaled measure λµ is a measure satisfying (λµ)(a) = λµ(a) for any measurable set A, λ R. Sum of measures µ + ν is a measure satisfying (µ + ν)(a) = µ(a) + ν(a) for any measurable set A. Random variable X denes measure µ X on (R, B(R)) such that µ X (A) = P (X A), for any A B(R). Conversely, any nonnegative measure µ X on (R, B(R)) such that µ X (R) = 1 denes a random variable. Suppose that random variables X and Y correspond to measures µ X and µ Y. Measure µ Z = λµ X + (1 λ)µ Y for λ [0, 1] is a nonnegative measure on (R, B(R)) and µ Z (R) = λµ X (R) + (1 λ)µ Y (R) = 1. Therefore, µ Z denes the random variable Z. We call Z a mixture of random variables X and Y with coecient λ and denote Z = λx (1 λ)y. In particular, F λx (1 λ)y (z) = λf X (z)+(1 λ)f Y (z), where F Z is a cumulative distribution function of the random variable Z. Proposition 5. (1 α) q α (X) is a concave function of (X, α) w.r.t. mixture operation and addition operation correspondingly, i.e., (1 (λα 1 + (1 λ)α 2 )) q (λα1 +(1 λ)α 2 )(λx (1 λ)y ) λ [(1 α 1 ) q α1 (X)] + (1 λ) [(1 α 2 ) q α2 (Y )]. 7
8 Proof. Denote α M = λα 1 +(1 λ)α 2. Then, with denitions of CVaR and λx (1 λ)y, we have (1 α M ) q αm (λx (1 λ)y ) = min c { (1 αm )c + E[B λ X + (1 B λ )Y c] +} = = min c { (1 αm )c + E[X c] + I(B λ = 1) + E[Y c] + I(B λ = 0) }. Since B λ is independent of X and Y, then E[X c] + I(B λ = 1) = E[X c] + EI(B λ = 1) = λe[x c] +. Then, { (1 α M ) q αm (λx (1 λ)y ) = min (1 αm )c + λe[x c] + + (1 λ)e[y c] +} { c min λ(1 α1 )c 1 + λe[x c 1 ] + + (1 λ)(1 α 2 )c 2 + (1 λ)e[y c 2 ] +} = c 1,c 2 = λ(1 α 1 ) q α1 (X) + (1 λ)(1 α 2 ) q α2 (Y ). The following statement is similar to a proposition in [2], which has motivated Proposition 5 in the rst place. Here we show how this proposition can be proved from Proposition 5 as a corollary. Corollary 6. Let X(x, p) be a discretely distributed random variable, taking values x = (x 1,..., x m ) with probabilities p = (p 1,..., p m ), p i 0, m i=1 p i = 1. Then function q α (X(x, p)) is a concave function of p. Proof. Note that if p M = λp 1 + (1 λ)p 2, then F X(x,pM )(x) = λf X(x,p1 )(x) + (1 λ)f X(x,p2 )(x). Therefore, X(x, p M ) = λx(x, p 1 ) (1 λ)x(x, p 2 ). Then Proposition 5 implies the concavity of q α (X(x, p)) w.r.t. vector p. The following proposition is similar to the one in [10]. proposition can be proved from Proposition 5 as a corollary. Here we show how this Corollary 7. Let random variable X p have a distribution F (x; p) = m i=1 p if i (x), where F i (x) for i = 1,..., m are the distribution functions, and p = (p 1,..., p m ) R m, p i 0, m i=1 p i = 1. Then function q α (X p ) is a concave function of p. Proof. Note that if p M = λp 1 + (1 λ)p 2, then F (x; p M ) = λf (x; p 1 ) + (1 λ)f (x; p 2 ). Therefore, X pm = λx p1 (1 λ)x p2. Then Proposition 5 implies the concavity of q α (X p ) w.r.t. vector p. Proposition 6. Buered probability of exceedance is a concave function of random variable w.r.t. mixture operation, i.e., p x (λx (1 λ)y ) λ p x (X)+(1 λ) p x (Y ), λ (0, 1). Proof. Suppose p x (X) = α 1 and p x (Y ) = α 2, then there are three possible cases. First, α 1 = α 2 = 1, then EX x and EY x, therefore, EλX (1 λ)y = λex + (1 λ)ey x, and p x (λx (1 λ)y ) = λ p x (X) + (1 λ) p x (Y ) = 1. Second, α 1 = α 2 = 0, then sup X x and sup Y x, and sup λx (1 λ)y max{sup X, sup Y } x. 8
9 Therefore, p x (λx (1 λ)y ) = λ p x (X) + (1 λ) p x (Y ) = 0. Third, in all other cases λα 1 + (1 λ)α 2 (0, 1). By Proposition 4, p x (X) p q 1 p (X) x for p (0, 1), therefore, p x (X) > p q 1 p (X) > x and, furthermore, p x (X) > p q 1 p (X) > x, sup X > x. Note also that p x (X) = p (0, 1) q 1 p (X) = x, sup X > x. Then, is equivalent to Note that p x (λx (1 λ)y ) λ p x (X) + (1 λ) p x (Y ) = λα 1 + (1 λ)α 2 (0, 1), q 1 (λα1 +(1 λ)α 2 )(λx (1 λ)y ) x, sup λx (1 λ)y > x. α 1 q 1 α1 (X) α 1 x. (5) If α 1 = 0, then 0 0. If α 1 = 1, then q 1 1 (X) = EX x. If α 1 (0, 1), then q 1 α1 (X) = x. Similarly, α 2 q 1 α2 (Y ) α 2 x. (6) Implying Proposition 5 and inequalities (5), (6), we get q λ(1 α1 )+(1 λ)(1 α 2 )(λx (1 λ)y ) λα 1 q 1 α1 (X) + (1 λ)α 2 q 1 α2 (Y ) λα 1 + (1 λ)α 2 λα 1x + (1 λ)α 2 x = x. λα 1 + (1 λ)α 2 Since α 1 > 0 or α 2 > 0, then sup X > x or sup Y > x, therefore sup λx (1 λ)y = max{sup X, sup Y } > x, which nishes the proof. It was mentioned that distribution functions are linear w.r.t. mixture operation: F λx (1 λ)y (x) = λf X (x) + (1 λ)f Y (x). Note that Proposition 6 proves that superdistribution functions are convex w.r.t. mixture operation: FλX (1 λ)y (x) λ F X (x) + (1 λ) F Y (x). Proposition 7. Buered probability of exceedance p x (X) is a monotonic function of random variable, i.e., p x (Y ) p x (Z) for Y Z almost surely. Proof. Suppose x sup Y and x sup Z. Then min a 0 E[aY +1] + min a 0 E[aZ +1] +, therefore, p x (Y ) p x (Z). Suppose x = sup Y, then p x (Y ) = 0 p x (Z). Suppose x = sup Z, then x sup Y, therefore, p x (Y ) = 0 = p x (Z). 4. Optimization Problems with bpoe 4.1. Two Families of Optimization Problems Denote program P(x), Denote program Q(α), P(x) : min p x (X) s.t. X X. Q(α) : min q α (X) s.t. X X. 9
10 For a set of random variables X, dene e X = inf X X EX, s X = inf sup X. X X Proposition 8. Let X 0 X be an optimal solution to P(x 0 ), where e X < x 0 s X. Then, X 0 is an optimal solution to Q(1 p x0 (X 0 )). Proof. Denote p = p x0 (X 0 ). Since x 0 > e X, then p x0 (X 0 ) < 1. If p x0 (X 0 ) = 0, then sup X 0 x 0, but x 0 s X, and sup X 0 s X by denition, therefore, sup X 0 = x 0 = s X. If 0 < p x0 (X 0 ) < 1, then, by Denition 1 of bpoe, q 1 p (X 0 ) = x 0. Suppose that X 0 is not an optimal solution to Q(1 p x0 (X 0 )), then there exists X X such that q 1 p (X ) < x 0. Since x 0 s X, then p > 0 and sup X x 0. Therefore, there exists p < p such that q 1 p (X ) = x 0, since q 1 p (X) is a continuous non-increasing function of p. There are two possible cases. First, if sup X = x 0, then p x0 (X ) = 0 < p, X 0 is not an optimal solution to P(x 0 ), contradiction. Second, if sup X > x 0, then p x0 (X ) = p < p, X 0 is not an optimal solution to P(x 0 ), contradiction. Two intervals for x 0 are not covered in Proposition 8. Note that for x 0 e X the optimal value for P(x 0 ) is 1, therefore, any feasible solution is an optimal solution. As for the interval x > s X, optimal value for P(x 0 ) is 0. If s X < sup X 0 x 0, then p x0 (X 0 ) = 0, and it is optimal for P(x 0 ), but q 1 (X 0 ) > s X, and it is not optimal for Q(1). Proposition 9. Let X 0 X be an optimal solution to Q(α 0 ). Then X 0 is an optimal solution to P( q α0 (X 0 )), unless sup X 0 > q α0 (X 0 ) and there exists X X such that 1. sup X = q α0 (X 0 ), 2. P (X = sup X ) 1 α 0. Proof. Denote x 0 = q α0 (X 0 ). First, suppose sup X 0 = x 0. Then p x0 (X 0 ) = 0, and X 0 is an optimal solution to P(x 0 ). Second, suppose that sup X 0 > x 0 and that exists X X such that p x0 (X ) < p x0 (X 0 ). Since x 0 = q α0 (X 0 ) and sup X 0 > x 0, then p x0 (X 0 ) = 1 α 0. Suppose sup X > x 0, then q α (X ) is strictly increasing on [0, 1 p x0 (X )]. Therefore, q α0 (X ) < q 1 px0 (X )(X ) = x 0, which implies that X 0 is not an optimal solution to Q(α 0 ), contradiction. Consequently, sup X = x 0. Suppose P (X = x 0 ) < 1 α 0. Then, q α (X ) is strictly increasing on [0, 1 P (X = x 0 )], and q α0 (X ) < x 0. Therefore, X 0 is not an optimal solution to Q(α 0 ), contradiction. Therefore, P (X = x 0 ) 1 α 0. Intuiton behind Proposition 9 is as follows. Note that X is also an optimal solution to Q(α 0 ). Therefore, we have two optimal solutions to right tail expectation minimization problem. The dierence between optimal solutions X and X 0 is that X is constant in its right 1 α 0 tail, and X 0 is not, since q α0 (X 0 ) < sup X 0. Proposition 9 implies that X is an optimal solution to P( q α0 (X 0 )), while X 0 is not. Which is a very natural risk-averse decision. This implies that, for certain problems, formulations of type P(x) provide more reasonable solutions than formulations of type Q(α). Corollary 8. Let X be a set of random variables, such that sup X = for all X X. Then, program families P(x), for x > e X, and Q(α), for 0 < α < 1, have the same set of optimal solutions. That is, if X 0 is optimal for P(x 0 ), then X 0 is optimal for Q(1 p x0 (X 0 )). Conversely, if X 0 is optimal for Q(α 0 ), then X 0 is optimal for P( q α0 (X 0 )). 10
11 Proof. Proposition 8 implies that if e X < x 0 s X =, then if X 0 is optimal for P(x 0 ), then X 0 is optimal for Q(1 p x0 (X 0 )). Note that since e X < x 0 < s X, then p x0 (X 0 ) (0, 1). Proposition 9 implies that if X 0 is optimal for Q(α 0 ), then X 0 is optimal for P( q α0 (X 0 )), unless exists X X such that sup X = q α0 (X 0 ). Which is impossible since sup X = > q α0 (X 0 ). Note that since α 0 (0, 1), then e X < q α0 (X 0 ) <. Assumption sup X = + for all X X in Corollary 8 might be too strong for some practical problems, where it is a common practice for all random variables to be dened on a nite probability space generated by system observations. Let us describe sets of optimal points (x, α) for problem families P(x) and Q(α). Dene f P (x) = min X X p x(x), f Q (α) = min X X q α(x). Then, sets of all optimal points of P(x) and Q(α) families are S P = {(x, α) f P (x) = 1 α}, S Q = {(x, α) f Q (α) = x}. Finally, reduced sets of optimal points are S P = {(x, α) S P e X x s X }, S Q = {(x, α) S Q x < s X } {(s X, 1)}. For any random variable X X there is a set S X = {(x, α) q α (X) = x}. Let us dene union of such sets for X X as S X = S X = {(x, α) exists X : q α (X) = x}. X X Naturally, we prefer random variables with superquantile as small as possible for a xed condence level and with condence level as big as possible for a xed superquantile value. Therefore, for the set S X we dene a Pareto front, which is often called an ecient frontier in nance, as follows: S X = {(x, α) S X x < x or α > α for all (x, α ) S X, (x, α ) (x, α)}. Proposition 10. S P S Q = S P = S Q = S X. Proof. Let us start with S P S Q = S X. Notice that (x, α) S P (x, α ) S X α α, (7) (x, α) S Q (x, α) S X x x. (8) Clearly, right sides of (7) and (8) hold for S X, which implies S X S P S Q. Suppose S X S P S Q, i.e., for some (x, α) S P S Q there exists (x, α ) S X such that x x, α α and (x, α ) (x, α). Notice that if (x, α) S P S Q, then (x, α ) S X α α and (x, α) S X x x. Then, x < x and α > α. Consider random variable X which has generated point (x, α ). Since q X (α ) < x, then q X (α) < x. Therefore, there exists ( q X (α), α) S X with q X (α) < x, while (x, α) S Q and (8) holds. Contradiction. Let us prove S P = S Q. Suppose (x, α) S P and x e X, then we can use Proposition 8 to conclude that (x, α) S Q. If x = e X, then p x (X) = 0, therefore, 11
12 (x, α) = (e X, 0) S Q. Let (x, α) S Q. If x < s X, then there is no X such that sup X = x, therefore, we can use Proposition 9 and conclude that (x, α) S P. Finally, let us prove S P S Q = S P. Since S P S P, S Q S Q and S P = S Q, then S P S P S Q. Suppose (x, α) S P S Q. If x < s X, then (x, α) S P. If x = s X, then α = 1, because there is only one α for any x in S P. Then (x, α) = (s X, 1) S P. Point with x > s X can not be in S Q since f Q (α) = min X X q α (X) min X X sup X = s X. Therefore, S P S Q S P, which nalizes the proof Parametric Simplex Algorithm for CVaR and bpoe Minimization Suppose that we are interested in solution to P 1 (x) for x s X. Suppose also we have an algorithm for solving P 2 (α), i.e. we can calculate function f 2 (α) for any α [0, 1]. Then we can nd an approximation to f 1 (x) by calculating f 2 (α) several times. First, calculate f 2 (1) = x X. If x > f 2 (1), then problem P 1 (x) is inecient. If x = f 2 (1), then f 1 (x) = 0. If x < f 2 (1), continue. Calculate f 2 (0). If x < f 2 (0), then P 1 (x) is infeasible. If x = f 2 (0), then f 1 (x) = 1. If x > f 2 (0), continue. Set a = 0, b = 1. Inequality 1 b < f 1 (x) < 1 a holds. We will calculate f 2 ((a + b)/2) at each step of the binary search procedure to make dierence b a as small as we need. Suppose that problem P 2 (α) can be expressed as a linear program. Let X 1,..., X n be a set of random variables discretely distributed on the common set of m scenarios, with scenario probabilities p 1,..., p m. Let random variable X i take value x j i under scenario j. Let λ = (λ 1,..., λ n ) be a set of decision variables such that X X X = n λ i X i, for some λ Λ, where Λ R n is a polyhedral set. Then P 2 (α) is equivalent to ( n ) min q α λ i X i λ i=1 i=1 s.t. λ Λ. With minimization form of CVaR, which is, q α (x) = min c { c α E[X c]+}, we reformulate P 2 (α) as min c,λ c α s.t. λ Λ. [ m n ] + p j λ i x j i c i=1 j=1 We will slightly adapt the parametric simplex method, see e.g. [8], [7], to solve this problem for all values α [0, 1]. To start, we need to obtain a basic feasible solution for one of extreme values, say α = 0. To get a solution for α = 0 we need to nd a random variable with minimal expectation. For example, if λ 0 and n i=1 λ i = 1, then we nd EX i = j p jx j i for all i and then take i = arg min i EX i. Then the optimal solution is λ 0 such that λ i = 1, λ j = 0 for j i. After obtaining the rst solution, denote µ = 1 and express reduced costs for 1 α nonbasic decision variables as linear functions of µ. Since the solution is optimal for 12
13 µ 0 = 1, all reduced costs are nonnegative at µ = 1. We do not have dependence on µ in constraints, that is why if µ is changing, solution remains primal feasible, but may become dual infeasible. Let us nd the biggest parameter value µ 1 at which reduced costs are still nonnegative. For µ 0 µ µ 1 solution λ 0 remains optimal. When µ > µ 1 some reduced costs became negative, that is why we make primal pivots until we nd new optimal solution λ 1. Then we express reduced costs as linear functions of µ, nd the next critical value µ 2 at which some costs reduce to 0, and so forth, see [8] for detailes. At some point all reduced costs will become nonnegative, no matter how big µ is, it means that the current solution is optimal for µ up to +, or up to α = 1. As a result of the algorithm we have: a sequence of parameters 1 = µ 0,..., µ M which corresponds to sequence α 0 = 0,..., α M = 1 1/µ M, α M+1 = 1; a sequence of optimal solutions λ 0,..., λ M ; a sequence of optimal objective values f 2(α i ) (f 2 (α M+1 ) = f 2 (α M )). To calculate f 1 (x), nd the interval [α j, α j+1 ] such that f 2 (α j ) x f 2 (α j+1 ). Then optimal solution is λ j = (λ j 1,..., λ j n) and f 1 (x) = p x ( n i=1 λj i X i) Finite Probability Space Applications Proposition 11. Let X be a convex set of random variables. inf X X p x (X), there are two cases: Then for the problem 1. If inf X X sup X is attained for some X X, and x = min X X sup X = sup X, then min X X p x (X) = 0 with optimal solution X. 2. Problem inf X X p x (X) can be reformulated as a following problem: inf E[Y + Y Y 1]+, where Y = cl cone(x x) is a closed convex cone. Proof. If Case 1 is not valid, then with Proposition 1 we conclude that inf p x(x) = inf min E[a(X x) + X X X X a 0 1]+ = inf E[a(X x) + X X,a 0 1]+. Denote Y = a(x x). Since X is convex, then constraints X X, a 0 are equivalent to Y cone(x x). Suppose that sequence {Y i } i=1 cone(x x) converges weakly to Y Y = cl cone(x x). Since cone(x x) is a convex set, and for convex sets weak and L 1 convergences are equivalent, then sequence {Y i } i=1 is L 1 -converging to Y. Therefore, lim E[Y i + 1] + = E[Y + 1] +. i Then, nally, inf p x(x) = inf E[Y + X X Y Y 1]+. Denote Π m = {q = (q 1,..., q m ) T R m q i 0, i = 1,..., m; m i=1 q i = 1}. Denote Π m + = {q = (q 1,..., q m ) T R m q i > 0, i = 1,..., m; m i=1 q i = 1}. For the following proposition we suppose X to be a set of random variables dened on the common nite probability space with the vector of elementary events' probabilies p = (p 1,..., p m ) T Π m +. Denote by the set S R m the set of vectors such that any random 13
14 variable X X takes values x 1,..., x m with probabilities p 1,..., p m correspondingly, for some x = (x 1,..., x m ) T S. Then X being closed convex set is equivalent to S being a closed convex set. Let us say that random variable X takes values from x = (x 1,..., x m ) R m with probabilities p = (p 1,..., p m ) Π m if X is disctretely distributed over m atoms and takes value x i with probability p i for i = 1,..., m. Corollary 9. Let X be a set of random variables such that X X X takes values from x S with probabilities p, where S R m is a convex set, p Π m +. Then for the problem inf X X p x (X), there are two cases: 1. If inf x S max i x i is attained for some x S, and x = min x S max i x i = max i x i, then min X X p x (X) = 0 with optimal solution X taking values from x. 2. Problem inf X X p x (X) can be reformulated as a following problem: inf y C pt [y + e] +, where C = cl cone(s xe) is a closed convex cone, e = (1,..., 1) T R m. Proof. Let us apply Proposition 11 to the specic case of nite probability space. If we consider the Case 2, then problem inf X X p x (X) can be reformulated inf y C pt [y + e] +, where C = cl cone(s xe) is a closed convex cone, since S is convex. Corollary 10. Let X be a set of random variables such that X X X takes values from x S with probabilities p, where S = {x Ax b} R m, p Π m +. Then for the problem inf X X p x (X), there are two cases: 1. If x = min x S max i x i, then min X X p x (X) = 0 with optimal solution X taking values from x such that max i x i = x. 2. Problem inf X X p x (X) can be reformulated as an LP: inf p T z (9) s.t. z y + e, (10) Ay a(b xae) 0, (11) z 0, a 0. (12) Proof. Corollary 9 implies that for the Case 2 problem inf X X p x (X) can be reformulated as inf y C p T [y + e] +, with C = cl cone(s xe). Note that S xe = {x A(x + xe) b} = {x Ax b xae}. 14
15 Note also that cone(s xe) = {ax Ax b xae, a > 0} {0} = {y Ay a(b xae), a > 0} {0}. Therefore, cl cone(s xe) = {y Ay a(b xae), a 0} {0} = {y Ay a(b xae), a 0}. Finally, introducing z = [y + e] +, we obtain reformulation (9)(12). Consider random real-valued function f(w; X), where w R k and X T = (X 1,..., X n ) is a random vector of dimension n. It is assumed here that variables X 1,..., X n can be observed, but can not be controlled. It is also assumed that value f(w; X) can be controlled by the vector w W R k. Proposition 12. Let f(w; X) be a convex function of w. Then p x (f(w; X)) is a quasiconvex function of w. Proof. Convexity of f implies f(w M ; X) λf(w 1 ; X) + (1 λ)f(w 2 ; X), for w M = λw 1 + (1 λ)w 2. Then, using monotonicity of p x (X), see Proposition 7, p x (f(w M ; X)) p x (λf(w 1 ; X) + (1 λ)f(w 2 ; X). p x (X) is a quasi-convex function of X, see Proposition 4. Quasi-convexity of a function p x (X) is equivalent to p x (λx 1 + (1 λ)x 2 ) max{ p x (X 1 ), p x (X 2 )}. Then, Therefore, p x (λf(w 1 ; X) + (1 λ)f(w 2 ; X)) max{ p x (f(w 1 ; X)), p x (f(w 2 ; X))}. p x (f(w M ; X)) max{ p x (f(w 1 ; X)), p x (f(w 2 ; X))}, i.e., p x (f(w; X)) is a quasi-convex function of w. Proposition 13. Let X be a random vector and let f(w; X) be a convex positive-homogeneous function of w. Assume that convergence of {w i } implies L 1 -convergence of {f(w i ; X)}. Let W be a convex set. Then for the problem inf w W p x (f(w; X)) there are two possible cases: 1. If inf w W sup f(w; X) is attained for some w W, and x = min w W sup f(w; X) = sup f(w ; X), then min w W p x (f(w; X)) = 0 with optimal solution w. 2. Problem inf w W p x (f(w; X)) can be reformulated as a convex programming problem: inf E[ f(v; X) + 1] +, v V where v T = (v 1,..., v k+1 ) R k+1, f(v; X) = f((v1,..., v k ) T ; X) v k+1, and V = cl cone(w {x}) is a closed convex cone. 15
16 Proof. In the Case 2 we can reformulate problem inf w W p x (f(w; X)) as follows: inf p x(f(w; X)) = inf E[a(f(w; X) x) + w W a 0,w W 1]+. Denote v T = (v 1,..., v k+1 ) R k+1 and f(v; X) = f((v 1,..., v k ) T ; X) v k+1. Then, inf p x(f(w; X)) = inf E[a f(v; X) + 1] +. w W a 0,v W {x} Note that f(v; X) is also a convex positive-homogeneous function of v. Then, inf p x(f(w; X)) = inf E[ f(av; X) + 1] +. w W a 0,v W {x} Note that since W is a convex set, then W {x} is also a convex set. Therefore, a 0, v W {x} av cone(w {x}). Note that feasible region can be extended to V = cl cone(w {x}): since convergence of {w i } implies L 1 -convergence of f(w; X), then convergence of v i v implies L 1 - convergence of f(v i ; X) L1 f(v; X), therefore, E[ f(v i ; X) + 1] + E[ f(v; X) + 1] +. Finally, inf p x(f(w; X)) = inf E[ f(v; X) + 1] +. w W v V Corollary 11. Let X = (X 1,..., X n, 1) T be a random vector, with the last component being a constant 1, and E X i < for i = 1,..., n. Let W R n+1 be a convex set. Then for the problem inf w W p x (w T X) there are two possible cases as follows: 1. If inf w W sup w T X is attained for some w W, and x = min w W sup w T X = sup(w ) T X, then min w W p x (w T X) = 0 with optimal solution w. 2. Problem inf w W p x (w T X) can be reformulated as the following convex programming problem: inf v V E[vT X + 1] +, where V = cl cone(w xe n+1 ) is a closed convex cone, where e n+1 = (0,..., 0, 1) T R n+1. Proof. Let us show that this corollary follows from Proposition 13. First, f(w; X) = n i=1 w ix i + w n+1 is convex and positive-homogeneous w.r.t. w. Second, suppose that w j w. Then E (w j ) T X w T X w j n+1 w n+1 + n w j i w i E X i 0, i=1 since E X i <. Therefore, convergence of w implies L 1 -convergence of f(w; X). Note that in this particular case of function f there is no need to introduce a new parameter. It is sucient to shift feasible region for w n+1 by x: W = W xen+1. Further change of variables v = aw, and setting V = cl cone( W ), as it is done in Proposition 13, nalizes the proof. 16
17 Corollary 12. Let X = (X 1,..., X n, 1) T be a random vector, with the last component being a constant 1, and E X i < for i = 1,..., n. Let W = {w Aw b} R n+1. Then for the problem inf w W p x (w T X) there are two possible cases: 1. If inf w W sup w T X is attained for some w W, and x = min w W sup w T X = sup(w ) T X, then min w W p x (w T X) = 0 with optimal solution w. 2. Problem inf w W p x (w T X) can be reformulated as the linear programming problem: inf E[v T X + 1] +, (13) s.t. Av T a(b xae n+1 ) 0, (14) a 0. (15) Proof. Let us prove that this corollary follows from Corollary 11. Note that Note further that W xe n+1 = {w Aw + xae n+1 b}. cone(w xe n+1 ) = {v Av + axae n+1 ab, a > 0} {0}. Finally, V = cl cone(w xe n+1 ) = {v Av a(b xae n+1 ) 0, a 0}. Corollary 13. Let X be a random vector taking values x 1,..., x m R n with probabilities p = (p 1,..., p m ) Π m +. Let f(w; X) be a convex positive-homogeneous function of w R k. Let W R k be a convex set. Then for the problem inf w W p x (f(w; X)) there are two possible cases: 1. If inf w W max j f(w; x j ) is attained for some w W, and x = min w W max j f(w; x j ) = max j f(w ; x j ), then min w W p x (f(w; X)) = 0 with optimal solution w. 2. Problem inf w W p x (f(w; X)) can be reformulated as the convex programming problem: inf v V m p j [ f(v; x j ) + 1] +, j=1 where v T = (v 1,..., v k+1 ) R k+1, f(v; X) = f((v1,..., v k ) T ; X) v k+1, and V = cl cone(w {x}) is a closed convex cone. Proof. Note that since there are nitely many scenarios for random vector X, then for w i w, due to continuity of function f w.r.t. w, we have max j f(w i ; x j ) f(w; x j ) 0. That is, convergence on w implies L 1 -convergence of f(w; X). Therefore, this corollary follows directly from Proposition References [1] Norton, M., and Uryasev, S. AUC and Buered AUC Maximization. University of Florida, Research Report, in preparation, [2] Pavlikov, K., and Uryasev, S. CVaR Distance between Distributions and Applications. University of Florida, Research Report, in preparation,
18 [3] Rockafellar, R. T., and Royset, J. O. Random variables, monotone relations and convex analysis. Mathematical Programming B, accepted. [4] Rockafellar, R. T., and Royset, J. O. On buered failure probability in design and optimization of structures. Reliability Engineering and System Safety 95, 5 (2010), [5] Rockafellar, R. T., and Uryasev, S. Conditional value-at-risk for general loss distributions. Journal of Banking and Finance (2002), [6] Rockafellar, R. T., and Uryasev, S. The Fundamental Risk Quadrangle in Risk Management, Optimization and Statistical Estimation. Surveys in Operations Research and Management Science 18 (2013). [7] Ruszczynski, A., and Vanderbei, R. J. Frontiers of Stochastically Nondominated Portfolios. Econometrica 71, 4 (2003), [8] Vanderbei, R. J. Linear Programming: Foundations and Extensions. International series in operations research & management science. Kluwer Academic, [9] Zabarankin, M., and Uryasev, S. Statistical decision problems. Selected concepts and portfolio safeguard case studies. Springer Optimization and Its Applications 85. New York, NY: Springer. [10] Zdanovskaya, V., Pavlikov, K., and Uryasev, S. Estimation of Mixtures of Continuous Distributions: Mixtures of Normal Distributions and Applications. University of Florida, Research Report, in preparation,
Duality of linear conic problems
Duality of linear conic problems Alexander Shapiro and Arkadi Nemirovski Abstract It is well known that the optimal values of a linear programming problem and its dual are equal to each other if at least
More information2.3 Convex Constrained Optimization Problems
42 CHAPTER 2. FUNDAMENTAL CONCEPTS IN CONVEX OPTIMIZATION Theorem 15 Let f : R n R and h : R R. Consider g(x) = h(f(x)) for all x R n. The function g is convex if either of the following two conditions
More informationMathematical finance and linear programming (optimization)
Mathematical finance and linear programming (optimization) Geir Dahl September 15, 2009 1 Introduction The purpose of this short note is to explain how linear programming (LP) (=linear optimization) may
More informationTwo-Stage Stochastic Linear Programs
Two-Stage Stochastic Linear Programs Operations Research Anthony Papavasiliou 1 / 27 Two-Stage Stochastic Linear Programs 1 Short Reviews Probability Spaces and Random Variables Convex Analysis 2 Deterministic
More informationTOPIC 4: DERIVATIVES
TOPIC 4: DERIVATIVES 1. The derivative of a function. Differentiation rules 1.1. The slope of a curve. The slope of a curve at a point P is a measure of the steepness of the curve. If Q is a point on the
More informationSeparation Properties for Locally Convex Cones
Journal of Convex Analysis Volume 9 (2002), No. 1, 301 307 Separation Properties for Locally Convex Cones Walter Roth Department of Mathematics, Universiti Brunei Darussalam, Gadong BE1410, Brunei Darussalam
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES Contents 1. Random variables and measurable functions 2. Cumulative distribution functions 3. Discrete
More informationMaximum Likelihood Estimation
Math 541: Statistical Theory II Lecturer: Songfeng Zheng Maximum Likelihood Estimation 1 Maximum Likelihood Estimation Maximum likelihood is a relatively simple method of constructing an estimator for
More informationNo: 10 04. Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics
No: 10 04 Bilkent University Monotonic Extension Farhad Husseinov Discussion Papers Department of Economics The Discussion Papers of the Department of Economics are intended to make the initial results
More informationInternational Doctoral School Algorithmic Decision Theory: MCDA and MOO
International Doctoral School Algorithmic Decision Theory: MCDA and MOO Lecture 2: Multiobjective Linear Programming Department of Engineering Science, The University of Auckland, New Zealand Laboratoire
More informationLinear Programming in Matrix Form
Linear Programming in Matrix Form Appendix B We first introduce matrix concepts in linear programming by developing a variation of the simplex method called the revised simplex method. This algorithm,
More informationLinear Programming Notes V Problem Transformations
Linear Programming Notes V Problem Transformations 1 Introduction Any linear programming problem can be rewritten in either of two standard forms. In the first form, the objective is to maximize, the material
More information1 Norms and Vector Spaces
008.10.07.01 1 Norms and Vector Spaces Suppose we have a complex vector space V. A norm is a function f : V R which satisfies (i) f(x) 0 for all x V (ii) f(x + y) f(x) + f(y) for all x,y V (iii) f(λx)
More information1 Introduction. Linear Programming. Questions. A general optimization problem is of the form: choose x to. max f(x) subject to x S. where.
Introduction Linear Programming Neil Laws TT 00 A general optimization problem is of the form: choose x to maximise f(x) subject to x S where x = (x,..., x n ) T, f : R n R is the objective function, S
More information24. The Branch and Bound Method
24. The Branch and Bound Method It has serious practical consequences if it is known that a combinatorial problem is NP-complete. Then one can conclude according to the present state of science that no
More informationChapter 4 Lecture Notes
Chapter 4 Lecture Notes Random Variables October 27, 2015 1 Section 4.1 Random Variables A random variable is typically a real-valued function defined on the sample space of some experiment. For instance,
More informationconstraint. Let us penalize ourselves for making the constraint too big. We end up with a
Chapter 4 Constrained Optimization 4.1 Equality Constraints (Lagrangians) Suppose we have a problem: Maximize 5, (x 1, 2) 2, 2(x 2, 1) 2 subject to x 1 +4x 2 =3 If we ignore the constraint, we get the
More informationWalrasian Demand. u(x) where B(p, w) = {x R n + : p x w}.
Walrasian Demand Econ 2100 Fall 2015 Lecture 5, September 16 Outline 1 Walrasian Demand 2 Properties of Walrasian Demand 3 An Optimization Recipe 4 First and Second Order Conditions Definition Walrasian
More informationLecture Notes on Measure Theory and Functional Analysis
Lecture Notes on Measure Theory and Functional Analysis P. Cannarsa & T. D Aprile Dipartimento di Matematica Università di Roma Tor Vergata cannarsa@mat.uniroma2.it daprile@mat.uniroma2.it aa 2006/07 Contents
More information(Basic definitions and properties; Separation theorems; Characterizations) 1.1 Definition, examples, inner description, algebraic properties
Lecture 1 Convex Sets (Basic definitions and properties; Separation theorems; Characterizations) 1.1 Definition, examples, inner description, algebraic properties 1.1.1 A convex set In the school geometry
More informationRandom variables P(X = 3) = P(X = 3) = 1 8, P(X = 1) = P(X = 1) = 3 8.
Random variables Remark on Notations 1. When X is a number chosen uniformly from a data set, What I call P(X = k) is called Freq[k, X] in the courseware. 2. When X is a random variable, what I call F ()
More informationE3: PROBABILITY AND STATISTICS lecture notes
E3: PROBABILITY AND STATISTICS lecture notes 2 Contents 1 PROBABILITY THEORY 7 1.1 Experiments and random events............................ 7 1.2 Certain event. Impossible event............................
More informationDate: April 12, 2001. Contents
2 Lagrange Multipliers Date: April 12, 2001 Contents 2.1. Introduction to Lagrange Multipliers......... p. 2 2.2. Enhanced Fritz John Optimality Conditions...... p. 12 2.3. Informative Lagrange Multipliers...........
More informationGambling Systems and Multiplication-Invariant Measures
Gambling Systems and Multiplication-Invariant Measures by Jeffrey S. Rosenthal* and Peter O. Schwartz** (May 28, 997.. Introduction. This short paper describes a surprising connection between two previously
More informationCASH FLOW MATCHING PROBLEM WITH CVaR CONSTRAINTS: A CASE STUDY WITH PORTFOLIO SAFEGUARD. Danjue Shang and Stan Uryasev
CASH FLOW MATCHING PROBLEM WITH CVaR CONSTRAINTS: A CASE STUDY WITH PORTFOLIO SAFEGUARD Danjue Shang and Stan Uryasev PROJECT REPORT #2011-1 Risk Management and Financial Engineering Lab Department of
More informationSome representability and duality results for convex mixed-integer programs.
Some representability and duality results for convex mixed-integer programs. Santanu S. Dey Joint work with Diego Morán and Juan Pablo Vielma December 17, 2012. Introduction About Motivation Mixed integer
More informationThe Heat Equation. Lectures INF2320 p. 1/88
The Heat Equation Lectures INF232 p. 1/88 Lectures INF232 p. 2/88 The Heat Equation We study the heat equation: u t = u xx for x (,1), t >, (1) u(,t) = u(1,t) = for t >, (2) u(x,) = f(x) for x (,1), (3)
More informationProximal mapping via network optimization
L. Vandenberghe EE236C (Spring 23-4) Proximal mapping via network optimization minimum cut and maximum flow problems parametric minimum cut problem application to proximal mapping Introduction this lecture:
More informationNotes on metric spaces
Notes on metric spaces 1 Introduction The purpose of these notes is to quickly review some of the basic concepts from Real Analysis, Metric Spaces and some related results that will be used in this course.
More informationMATH 425, PRACTICE FINAL EXAM SOLUTIONS.
MATH 45, PRACTICE FINAL EXAM SOLUTIONS. Exercise. a Is the operator L defined on smooth functions of x, y by L u := u xx + cosu linear? b Does the answer change if we replace the operator L by the operator
More informationST 371 (IV): Discrete Random Variables
ST 371 (IV): Discrete Random Variables 1 Random Variables A random variable (rv) is a function that is defined on the sample space of the experiment and that assigns a numerical variable to each possible
More informationA NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION
1 A NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION Dimitri Bertsekas M.I.T. FEBRUARY 2003 2 OUTLINE Convexity issues in optimization Historical remarks Our treatment of the subject Three unifying lines of
More informationLECTURE 5: DUALITY AND SENSITIVITY ANALYSIS. 1. Dual linear program 2. Duality theory 3. Sensitivity analysis 4. Dual simplex method
LECTURE 5: DUALITY AND SENSITIVITY ANALYSIS 1. Dual linear program 2. Duality theory 3. Sensitivity analysis 4. Dual simplex method Introduction to dual linear program Given a constraint matrix A, right
More informationMetric Spaces. Chapter 7. 7.1. Metrics
Chapter 7 Metric Spaces A metric space is a set X that has a notion of the distance d(x, y) between every pair of points x, y X. The purpose of this chapter is to introduce metric spaces and give some
More informationA Uniform Asymptotic Estimate for Discounted Aggregate Claims with Subexponential Tails
12th International Congress on Insurance: Mathematics and Economics July 16-18, 2008 A Uniform Asymptotic Estimate for Discounted Aggregate Claims with Subexponential Tails XUEMIAO HAO (Based on a joint
More information1 if 1 x 0 1 if 0 x 1
Chapter 3 Continuity In this chapter we begin by defining the fundamental notion of continuity for real valued functions of a single real variable. When trying to decide whether a given function is or
More informationRegulatory Impacts on Credit Portfolio Management
Regulatory Impacts on Credit Portfolio Management Ursula Theiler*, Vladimir Bugera **, Alla Revenko **, Stanislav Uryasev ** *Risk Training, Carl-Zeiss-Str. 11, D-83052 Bruckmuehl, Germany, mailto: theiler@risk-training.org,
More informationWhat is Linear Programming?
Chapter 1 What is Linear Programming? An optimization problem usually has three essential ingredients: a variable vector x consisting of a set of unknowns to be determined, an objective function of x to
More informationAdvanced Microeconomics
Advanced Microeconomics Ordinal preference theory Harald Wiese University of Leipzig Harald Wiese (University of Leipzig) Advanced Microeconomics 1 / 68 Part A. Basic decision and preference theory 1 Decisions
More informationOptimal shift scheduling with a global service level constraint
Optimal shift scheduling with a global service level constraint Ger Koole & Erik van der Sluis Vrije Universiteit Division of Mathematics and Computer Science De Boelelaan 1081a, 1081 HV Amsterdam The
More informationContinuity of the Perron Root
Linear and Multilinear Algebra http://dx.doi.org/10.1080/03081087.2014.934233 ArXiv: 1407.7564 (http://arxiv.org/abs/1407.7564) Continuity of the Perron Root Carl D. Meyer Department of Mathematics, North
More informationLinear Programming. March 14, 2014
Linear Programming March 1, 01 Parts of this introduction to linear programming were adapted from Chapter 9 of Introduction to Algorithms, Second Edition, by Cormen, Leiserson, Rivest and Stein [1]. 1
More information1 Solving LPs: The Simplex Algorithm of George Dantzig
Solving LPs: The Simplex Algorithm of George Dantzig. Simplex Pivoting: Dictionary Format We illustrate a general solution procedure, called the simplex algorithm, by implementing it on a very simple example.
More information1. Prove that the empty set is a subset of every set.
1. Prove that the empty set is a subset of every set. Basic Topology Written by Men-Gen Tsai email: b89902089@ntu.edu.tw Proof: For any element x of the empty set, x is also an element of every set since
More informationPractical Guide to the Simplex Method of Linear Programming
Practical Guide to the Simplex Method of Linear Programming Marcel Oliver Revised: April, 0 The basic steps of the simplex algorithm Step : Write the linear programming problem in standard form Linear
More informationMINIMIZATION OF ENTROPY FUNCTIONALS UNDER MOMENT CONSTRAINTS. denote the family of probability density functions g on X satisfying
MINIMIZATION OF ENTROPY FUNCTIONALS UNDER MOMENT CONSTRAINTS I. Csiszár (Budapest) Given a σ-finite measure space (X, X, µ) and a d-tuple ϕ = (ϕ 1,..., ϕ d ) of measurable functions on X, for a = (a 1,...,
More informationMath 4310 Handout - Quotient Vector Spaces
Math 4310 Handout - Quotient Vector Spaces Dan Collins The textbook defines a subspace of a vector space in Chapter 4, but it avoids ever discussing the notion of a quotient space. This is understandable
More informationOnline Appendix to Stochastic Imitative Game Dynamics with Committed Agents
Online Appendix to Stochastic Imitative Game Dynamics with Committed Agents William H. Sandholm January 6, 22 O.. Imitative protocols, mean dynamics, and equilibrium selection In this section, we consider
More informationLargest Fixed-Aspect, Axis-Aligned Rectangle
Largest Fixed-Aspect, Axis-Aligned Rectangle David Eberly Geometric Tools, LLC http://www.geometrictools.com/ Copyright c 1998-2016. All Rights Reserved. Created: February 21, 2004 Last Modified: February
More informationDuality in General Programs. Ryan Tibshirani Convex Optimization 10-725/36-725
Duality in General Programs Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: duality in linear programs Given c R n, A R m n, b R m, G R r n, h R r : min x R n c T x max u R m, v R r b T
More information4.6 Linear Programming duality
4.6 Linear Programming duality To any minimization (maximization) LP we can associate a closely related maximization (minimization) LP. Different spaces and objective functions but in general same optimal
More informationOn closed-form solutions of a resource allocation problem in parallel funding of R&D projects
Operations Research Letters 27 (2000) 229 234 www.elsevier.com/locate/dsw On closed-form solutions of a resource allocation problem in parallel funding of R&D proects Ulku Gurler, Mustafa. C. Pnar, Mohamed
More informationA FIRST COURSE IN OPTIMIZATION THEORY
A FIRST COURSE IN OPTIMIZATION THEORY RANGARAJAN K. SUNDARAM New York University CAMBRIDGE UNIVERSITY PRESS Contents Preface Acknowledgements page xiii xvii 1 Mathematical Preliminaries 1 1.1 Notation
More informationDuality in Linear Programming
Duality in Linear Programming 4 In the preceding chapter on sensitivity analysis, we saw that the shadow-price interpretation of the optimal simplex multipliers is a very useful concept. First, these shadow
More informationLinear Algebra I. Ronald van Luijk, 2012
Linear Algebra I Ronald van Luijk, 2012 With many parts from Linear Algebra I by Michael Stoll, 2007 Contents 1. Vector spaces 3 1.1. Examples 3 1.2. Fields 4 1.3. The field of complex numbers. 6 1.4.
More informationContinued Fractions and the Euclidean Algorithm
Continued Fractions and the Euclidean Algorithm Lecture notes prepared for MATH 326, Spring 997 Department of Mathematics and Statistics University at Albany William F Hammond Table of Contents Introduction
More informationMathematical Methods of Engineering Analysis
Mathematical Methods of Engineering Analysis Erhan Çinlar Robert J. Vanderbei February 2, 2000 Contents Sets and Functions 1 1 Sets................................... 1 Subsets.............................
More informationRecovery of primal solutions from dual subgradient methods for mixed binary linear programming; a branch-and-bound approach
MASTER S THESIS Recovery of primal solutions from dual subgradient methods for mixed binary linear programming; a branch-and-bound approach PAULINE ALDENVIK MIRJAM SCHIERSCHER Department of Mathematical
More informationDiscrete Optimization
Discrete Optimization [Chen, Batson, Dang: Applied integer Programming] Chapter 3 and 4.1-4.3 by Johan Högdahl and Victoria Svedberg Seminar 2, 2015-03-31 Todays presentation Chapter 3 Transforms using
More informationLinear Programming I
Linear Programming I November 30, 2003 1 Introduction In the VCR/guns/nuclear bombs/napkins/star wars/professors/butter/mice problem, the benevolent dictator, Bigus Piguinus, of south Antarctica penguins
More informationLinear Programming: Theory and Applications
Linear Programming: Theory and Applications Catherine Lewis May 11, 2008 1 Contents 1 Introduction to Linear Programming 3 1.1 What is a linear program?...................... 3 1.2 Assumptions.............................
More informationDefinition and Properties of the Production Function: Lecture
Definition and Properties of the Production Function: Lecture II August 25, 2011 Definition and : Lecture A Brief Brush with Duality Cobb-Douglas Cost Minimization Lagrangian for the Cobb-Douglas Solution
More informationPacific Journal of Mathematics
Pacific Journal of Mathematics GLOBAL EXISTENCE AND DECREASING PROPERTY OF BOUNDARY VALUES OF SOLUTIONS TO PARABOLIC EQUATIONS WITH NONLOCAL BOUNDARY CONDITIONS Sangwon Seo Volume 193 No. 1 March 2000
More informationCONTINUED FRACTIONS AND PELL S EQUATION. Contents 1. Continued Fractions 1 2. Solution to Pell s Equation 9 References 12
CONTINUED FRACTIONS AND PELL S EQUATION SEUNG HYUN YANG Abstract. In this REU paper, I will use some important characteristics of continued fractions to give the complete set of solutions to Pell s equation.
More informationMATH4427 Notebook 2 Spring 2016. 2 MATH4427 Notebook 2 3. 2.1 Definitions and Examples... 3. 2.2 Performance Measures for Estimators...
MATH4427 Notebook 2 Spring 2016 prepared by Professor Jenny Baglivo c Copyright 2009-2016 by Jenny A. Baglivo. All Rights Reserved. Contents 2 MATH4427 Notebook 2 3 2.1 Definitions and Examples...................................
More informationLecture 3. Linear Programming. 3B1B Optimization Michaelmas 2015 A. Zisserman. Extreme solutions. Simplex method. Interior point method
Lecture 3 3B1B Optimization Michaelmas 2015 A. Zisserman Linear Programming Extreme solutions Simplex method Interior point method Integer programming and relaxation The Optimization Tree Linear Programming
More informationLecture 2: The SVM classifier
Lecture 2: The SVM classifier C19 Machine Learning Hilary 2015 A. Zisserman Review of linear classifiers Linear separability Perceptron Support Vector Machine (SVM) classifier Wide margin Cost function
More informationLinear Programming: Chapter 11 Game Theory
Linear Programming: Chapter 11 Game Theory Robert J. Vanderbei October 17, 2007 Operations Research and Financial Engineering Princeton University Princeton, NJ 08544 http://www.princeton.edu/ rvdb Rock-Paper-Scissors
More informationThe sample space for a pair of die rolls is the set. The sample space for a random number between 0 and 1 is the interval [0, 1].
Probability Theory Probability Spaces and Events Consider a random experiment with several possible outcomes. For example, we might roll a pair of dice, flip a coin three times, or choose a random real
More informationx a x 2 (1 + x 2 ) n.
Limits and continuity Suppose that we have a function f : R R. Let a R. We say that f(x) tends to the limit l as x tends to a; lim f(x) = l ; x a if, given any real number ɛ > 0, there exists a real number
More informationLecture Notes on Elasticity of Substitution
Lecture Notes on Elasticity of Substitution Ted Bergstrom, UCSB Economics 210A March 3, 2011 Today s featured guest is the elasticity of substitution. Elasticity of a function of a single variable Before
More informationAdvanced Lecture on Mathematical Science and Information Science I. Optimization in Finance
Advanced Lecture on Mathematical Science and Information Science I Optimization in Finance Reha H. Tütüncü Visiting Associate Professor Dept. of Mathematical and Computing Sciences Tokyo Institute of Technology
More informationRANDOM INTERVAL HOMEOMORPHISMS. MICHA L MISIUREWICZ Indiana University Purdue University Indianapolis
RANDOM INTERVAL HOMEOMORPHISMS MICHA L MISIUREWICZ Indiana University Purdue University Indianapolis This is a joint work with Lluís Alsedà Motivation: A talk by Yulij Ilyashenko. Two interval maps, applied
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 14 10/27/2008 MOMENT GENERATING FUNCTIONS
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 14 10/27/2008 MOMENT GENERATING FUNCTIONS Contents 1. Moment generating functions 2. Sum of a ranom number of ranom variables 3. Transforms
More information4. Continuous Random Variables, the Pareto and Normal Distributions
4. Continuous Random Variables, the Pareto and Normal Distributions A continuous random variable X can take any value in a given range (e.g. height, weight, age). The distribution of a continuous random
More informationALMOST COMMON PRIORS 1. INTRODUCTION
ALMOST COMMON PRIORS ZIV HELLMAN ABSTRACT. What happens when priors are not common? We introduce a measure for how far a type space is from having a common prior, which we term prior distance. If a type
More informationVector and Matrix Norms
Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a non-empty
More informationProbability Generating Functions
page 39 Chapter 3 Probability Generating Functions 3 Preamble: Generating Functions Generating functions are widely used in mathematics, and play an important role in probability theory Consider a sequence
More informationGLOBAL OPTIMIZATION METHOD FOR SOLVING MATHEMATICAL PROGRAMS WITH LINEAR COMPLEMENTARITY CONSTRAINTS. 1. Introduction
GLOBAL OPTIMIZATION METHOD FOR SOLVING MATHEMATICAL PROGRAMS WITH LINEAR COMPLEMENTARITY CONSTRAINTS N.V. THOAI, Y. YAMAMOTO, AND A. YOSHISE Abstract. We propose a method for finding a global optimal solution
More informationSingle item inventory control under periodic review and a minimum order quantity
Single item inventory control under periodic review and a minimum order quantity G. P. Kiesmüller, A.G. de Kok, S. Dabia Faculty of Technology Management, Technische Universiteit Eindhoven, P.O. Box 513,
More informationNonlinear Optimization: Algorithms 3: Interior-point methods
Nonlinear Optimization: Algorithms 3: Interior-point methods INSEAD, Spring 2006 Jean-Philippe Vert Ecole des Mines de Paris Jean-Philippe.Vert@mines.org Nonlinear optimization c 2006 Jean-Philippe Vert,
More information1.2 Solving a System of Linear Equations
1.. SOLVING A SYSTEM OF LINEAR EQUATIONS 1. Solving a System of Linear Equations 1..1 Simple Systems - Basic De nitions As noticed above, the general form of a linear system of m equations in n variables
More informationConvex analysis and profit/cost/support functions
CALIFORNIA INSTITUTE OF TECHNOLOGY Division of the Humanities and Social Sciences Convex analysis and profit/cost/support functions KC Border October 2004 Revised January 2009 Let A be a subset of R m
More informationLecture 13: Martingales
Lecture 13: Martingales 1. Definition of a Martingale 1.1 Filtrations 1.2 Definition of a martingale and its basic properties 1.3 Sums of independent random variables and related models 1.4 Products of
More informationFurther Study on Strong Lagrangian Duality Property for Invex Programs via Penalty Functions 1
Further Study on Strong Lagrangian Duality Property for Invex Programs via Penalty Functions 1 J. Zhang Institute of Applied Mathematics, Chongqing University of Posts and Telecommunications, Chongqing
More information1 The Brownian bridge construction
The Brownian bridge construction The Brownian bridge construction is a way to build a Brownian motion path by successively adding finer scale detail. This construction leads to a relatively easy proof
More informationLinear Programming. April 12, 2005
Linear Programming April 1, 005 Parts of this were adapted from Chapter 9 of i Introduction to Algorithms (Second Edition) /i by Cormen, Leiserson, Rivest and Stein. 1 What is linear programming? The first
More informationChapter 7. Continuity
Chapter 7 Continuity There are many processes and eects that depends on certain set of variables in such a way that a small change in these variables acts as small change in the process. Changes of this
More informationLinear Programming. Widget Factory Example. Linear Programming: Standard Form. Widget Factory Example: Continued.
Linear Programming Widget Factory Example Learning Goals. Introduce Linear Programming Problems. Widget Example, Graphical Solution. Basic Theory:, Vertices, Existence of Solutions. Equivalent formulations.
More informationAn axiomatic approach to capital allocation
An axiomatic approach to capital allocation Michael Kalkbrener Deutsche Bank AG Abstract Capital allocation techniques are of central importance in portfolio management and risk-based performance measurement.
More informationOptimization under uncertainty: modeling and solution methods
Optimization under uncertainty: modeling and solution methods Paolo Brandimarte Dipartimento di Scienze Matematiche Politecnico di Torino e-mail: paolo.brandimarte@polito.it URL: http://staff.polito.it/paolo.brandimarte
More information5. Continuous Random Variables
5. Continuous Random Variables Continuous random variables can take any value in an interval. They are used to model physical characteristics such as time, length, position, etc. Examples (i) Let X be
More information14.451 Lecture Notes 10
14.451 Lecture Notes 1 Guido Lorenzoni Fall 29 1 Continuous time: nite horizon Time goes from to T. Instantaneous payo : f (t; x (t) ; y (t)) ; (the time dependence includes discounting), where x (t) 2
More informationBANACH AND HILBERT SPACE REVIEW
BANACH AND HILBET SPACE EVIEW CHISTOPHE HEIL These notes will briefly review some basic concepts related to the theory of Banach and Hilbert spaces. We are not trying to give a complete development, but
More informationINDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS
INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS STEVEN P. LALLEY AND ANDREW NOBEL Abstract. It is shown that there are no consistent decision rules for the hypothesis testing problem
More information15 Kuhn -Tucker conditions
5 Kuhn -Tucker conditions Consider a version of the consumer problem in which quasilinear utility x 2 + 4 x 2 is maximised subject to x +x 2 =. Mechanically applying the Lagrange multiplier/common slopes
More informationInner Product Spaces
Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and
More informationPractice with Proofs
Practice with Proofs October 6, 2014 Recall the following Definition 0.1. A function f is increasing if for every x, y in the domain of f, x < y = f(x) < f(y) 1. Prove that h(x) = x 3 is increasing, using
More information160 CHAPTER 4. VECTOR SPACES
160 CHAPTER 4. VECTOR SPACES 4. Rank and Nullity In this section, we look at relationships between the row space, column space, null space of a matrix and its transpose. We will derive fundamental results
More informationHøgskolen i Narvik Sivilingeniørutdanningen STE6237 ELEMENTMETODER. Oppgaver
Høgskolen i Narvik Sivilingeniørutdanningen STE637 ELEMENTMETODER Oppgaver Klasse: 4.ID, 4.IT Ekstern Professor: Gregory A. Chechkin e-mail: chechkin@mech.math.msu.su Narvik 6 PART I Task. Consider two-point
More information