NonLife Insurance Mathematics. Christel Geiss and Stefan Geiss Department of Mathematics and Statistics University of Jyväskylä


 Griffin Campbell
 3 years ago
 Views:
Transcription
1 NonLife Insurance Mathematics Christel Geiss and Stefan Geiss Department of Mathematics and Statistics University of Jyväskylä July 29, 215
2 2
3 Contents 1 Introduction Some facts about probability Claim number process models The homogeneous Poisson process The renewal process The inhomogeneous Poisson process The total claim amount process S(t) The CramérLundbergmodel The renewal model Properties of S(t) Premium calculation principles Used principles Claim size distributions Examples The QQplot Modern premium calculation principles The exponential principle The quantile principle The Esscher principle The distribution of S(t) Mixture distributions Applications in insurance The Panjer recursion
4 4 CONTENTS 7.5 Approximation of F S(t) Monte Carlo approximations of F S(t) Reinsurance treaties 49 9 Probability of ruin The risk process Bounds for the ruin probability An asymptotics for the ruin probability Problems 81 A The LebesgueStieltjes integral 95 A.1 The RiemannStieltjes integral A.2 The LebesgueStieltjes integral
5 1. Introduction Insurance Mathematics is sometimes divided into life insurance, health insurance and nonlife insurance. Life insurance includes for instance life insurance contracts and pensions where long terms are covered. Nonlife insurance comprises insurances against fire, water damage, earthquake, industrial catastrophes or car insurance, for example. Nonlife insurances cover in general a year or other fixed time periods. Health insurance is special because it is differently organized in each country. The course material is based on the textbook NonLife Insurance Mathematics by Thomas Mikosch [6] The problem to solve We will consider the following situation. 1. Insurance contracts (or policies ) are sold. This is the income of the insurance company. 2. At times T i, T 1 T 2... claims happen. The times T i are called the claim arrival times. 3. The ith claim arriving at time T i causes the claim size X i. Task: Find a stochastic model for the T i s and X i s to compute or estimate how much an insurance company should demand for its contracts and how much initial capital of the insurance company is required to keep the probability of ruin below a certain level. 5
6 6 CHAPTER 1. INTRODUCTION 1.1 Some facts about probability We shortly recall some definitions and facts from probability theory which we need in this course. For more information see [9] or [2], for example. (i) A probability space is a triple (Ω, F, P), where Ω is a nonempty set, F is a σalgebra consisting of subsets of Ω and P is a probability measure on (Ω, F). (ii) A function f : Ω R is called a random variable if and only if for all intervals (a, b), < a < b < the preimage f 1 ((a, b)) : {ω Ω : a < f(ω) < b} F. (iii) The random variables f 1,..., f n are independent if and only if P(f 1 B 1,..., f n B n ) P(f 1 B 1 ) P(f n B n ) for all B k B(R), k 1,..., n. (Here B(R) denotes the Borel σ algebra.) If the f i s have discrete values, i.e. f i : Ω {x 1, x 2, x 3,...}, then the random variables f 1,..., f n are independent if and only if P(f 1 k 1,..., f n k n ) P(f 1 k 1 ) P(f n k n ) for all k i {x 1, x 2, x 3...}. (iv) If f 1,..., f n are independent random variables such that f i has the density function h i (x), i.e. P(f i (a, b)) b h a i(x)dx, then P((f 1,..., f n ) B) 1I B (x 1,..., x n )h 1 (x) h n (x n )dx 1 dx n R n for all B B(R n ). The σalgebra B(R n ) is the Borel σalgebra, which is the smallest σalgebra containing all the open rectangles (a 1, b 1 )... (a n, b n ). The function 1I B (x) is the indicator function for the set B, which is defined as { 1 if x B 1I B (x) if x B.
7 1.1. SOME FACTS ABOUT PROBABILITY 7 (v) A random variable f : Ω {, 1, 2,...} is Poisson distributed with parameter λ > if and only if λ λk P(f k) e k!. This is often written as f P ois(λ). (vi) A random variable g : Ω [, ) is exponentially distributed with parameter λ > if and only if for all a < b P(g (a, b)) λ b a 1I [, ) (x)e λx dx. The picture below shows the density λ1i [, ) (x)e λx for λ 3. density for lambda3 y x
8 8 CHAPTER 1. INTRODUCTION
9 2. Models for the claim number process N(t) In the following we will introduce three processes which are used as claim number processes: the Poisson process, the renewal process and the inhomogeneous Poisson process. 2.1 The homogeneous Poisson process with parameter λ > Definition (homogeneous Poisson process). A stochastic process N (N(t)) t [, ) is a Poisson process if the following conditions are fulfilled: (P1) N() a.s. (almost surely), i.e. P({ω Ω : N(, ω) }) 1. (P2) N has independent increments, i.e. if t < t 1 <... < t n, (n 1), then N(t n ) N(t n 1 ), N(t n 1 ) N(t n 2 ),..., N(t 1 ) N(t ) are independent. (P3) For any s and t > the random variable N(t+s) N(s) is Poisson distributed, i.e. λt (λt)m P(N(t + s) N(s) m) e, m, 1, 2,... m! (P4) The paths of N, i.e. the functions (N(t, ω)) t [, ) for fixed ω are almost surely right continuous and have left limits. One says N has càdlàg (continue à droite, limite à gauche) paths. 9
10 1 CHAPTER 2. CLAIM NUMBER PROCESS MODELS Lemma Assume W 1, W 2,... are independent and exponentially distributed with parameter λ >. Then, for any x > we have n 1 P(W W n x) 1 e λx k (λx) k. k! This means the sum of independent exponentially distributed random variables is a Gamma distributed random variable. Proof: Exercise. Definition Let W 1, W 2,... be independent and exponentially distributed with parameter λ >. Define T n : W W n and ˆN(t, ω) : #{i 1 : T i (ω) t}, t. Lemma For each n, 1, 2,... and for all t > it holds P({ω Ω : ˆN(t, ω) n}) e λt (λt)n, n! i.e. ˆN(t) is Poisson distributed with parameter λt. Proof: From the definition of ˆN it can be concluded that {ω : ˆN(t, ω) n} {ω : Tn (ω) t < T n+1 (ω)} {ω : T n (ω) t} \ {ω : T n+1 (ω) t} Because of T n T n+1 we have the inclusion {T n+1 t} {T n t}. This implies P( ˆN(t) n) P(T n t) P(T n+1 t) n 1 1 e λt (λt) k 1 + e λt k! k λt (λt)n e. n! n (λt) k k k!
11 2.1. THE HOMOGENEOUS POISSON PROCESS 11 Theorem (a) ˆN(t) t [, ) is a Poisson process with parameter λ >. (b) Any Poisson process N(t) with parameter λ > can be written as Proof: N(t) #{i 1, T i t}, t, where T n W W n, n 1, and W 1, W 2,... are independent and exponentially distributed with λ >. (a) We check the properties of the Definition (P1) From (vi) of Section 1.1 we get for any x > that P(W 1 > ) P(W 1 (, )) λ 1I [, ) (y)e λy dy 1. This implies that ˆN(, ω) if only < T 1 (ω) W 1 (ω) but W 1 > holds almost surely. Hence ˆN(, ω) a.s. (P2) We only show that ˆN(s) and ˆN(t) ˆN(s) are independent: i.e. P( ˆN(s) l, ˆN(t) ˆN(s) m)p( ˆN(s) l)p( ˆN(t) ˆN(s) m) (1) for l, m. The general case can be shown similarly. It holds P( ˆN(s) l, ˆN(t) ˆN(s) m) P( ˆN(s) l, ˆN(t) m + l) P(T l s < T l+1, T l+m t < T l+m+1 ) By defining functions f 1, f 2, f 3 and f 4 as f 1 : T l f 2 : W l+1 f 3 : W l W l+m f 4 : W l+m+1,
12 12 CHAPTER 2. CLAIM NUMBER PROCESS MODELS and h 1,..., h 4 as the respective densities, it follows that P(T l s < T l+1, T l+m t < T l+m+1 ) P(f 1 s < f 1 + f 2, f 1 + f 2 + f 3 t < f 1 + f 2 + f 3 + f 4 ) P( f 1 < s, s f 1 < f 2 <, f 3 < t f 1 f 2, : I 1 s t x 1 x 2 s x 1 h 4 (x 4 )dx 4 h 3 (x 3 )dx 3 t x 1 x 2 x } 3 {{ } I 4 (x 1,x 2,x 3 ) } {{ } I 3 (x 1,x 2 ) t (f 1 + f 2 + f 3 ) < f 4 < ) } {{ } I 2 (x 1 ) h 2 (x 2 )dx 2 h 1 (x 1 )dx 1 By direct computation and rewriting the density function of f 4 W l+m+1, I 4 (x 1, x 2, x 3 ) λe λx 4 1I [, ) (x 4 )dx 4 t x 1 x 2 x 3 e λ(t x 1 x 2 x 3 ). Here we used t x 1 x 2 x 3 >. This is true because the integration w.r.t. x 3 implies < x 3 < t x 1 x 2. The density of f 3 W l W l+m is Therefore, I 3 (x 1, x 2 ) x m 2 3 h 3 (x 3 ) λ m 1 (m 2)! 1I [, )(x 3 )e λx 3. t x 1 x 2 λ m 1 x m 2 3 (m 2)! e λx 3 e λ(t x 1 x 2 x 3 ) dx 3 1I [,t x1 )(x 2 )e λ(t x 1 x 2 ) λ m 1 (t x 1 x 2 ) m 1. (m 1)!
13 2.1. THE HOMOGENEOUS POISSON PROCESS 13 The density of f 2 W l+1 is This implies h 2 (x 2 ) 1I [, ) (x 2 )λe λx 2. I 2 (x 1 ) 1I [,t x1 )(x 2 )e λ(t x1 x2) λ m 1 (t x 1 x 2)m 1 (m 1)! s x 1 λ m e λ(t x 1) (t s)m. m! Finally, from Lemma we conclude I 1 s λ m e λ(t x 1) λ m λ l λt (t s)m e ( m! (λs) l If we sum l! (t s)m m! s l l! λ l x l 1 1 ) ( ) (λ(t s)) e λs m e λ(t s) m! P( ˆN(s) l)p( ˆN(t s) m). λe λx 2 dx 2 (l 1)! 1I [, )(x 1 )e λx 1 dx 1 P( ˆN(s) l, ˆN(t) ˆN(s) m) P( ˆN(s) l)p( ˆN(t s) m) over l N we get and hence (1). P( ˆN(t) ˆN(s) m) P( ˆN(t s) m) (2) (P3) follows from Lemma and (2). (P4) is clear from the construction. (b) The proof is an exercise.
14 14 CHAPTER 2. CLAIM NUMBER PROCESS MODELS Poisson, lambda5 X T Poisson, lambda1 X T 2.2 The renewal process To model windstorm claims, for example, it is not good to use the Poisson process because windstorm claims happen rarely, sometimes with years in between. The Pareto distribution, for example, which has the distribution function ( ) α κ F (x) 1 κ + x
15 2.2. THE RENEWAL PROCESS 15 with parameters α, κ > would fit better. For a Pareto distributed random variable it is more likely to have large values than for an exponential distributed random variable. Definition (Renewal process). Assume that W 1, W 2,... are i.i.d. (independent and identically distributed) random variables such that W 1 > a.s. Then { T : is a renewal sequence and is the renewal process. T n : W W n, n 1 N(t) : #{i 1 : T i t}, t In order to study the limit behavior of N we need the Strong Law of Large Numbers (SLLN): Theorem (SLLN). If the random variables X 1, X 2,... are i.i.d. with E X 1 < then X 1 + X X n n EX 1 a.s. n Theorem (SLLN for renewal processes). Assume N(t) is a renewal process. If EW 1 <, then N(t) lim t t 1 EW 1 a.s. Proof: Because of {ω Ω : N(t)(ω) n} {ω Ω : T n (ω) t < T n+1 (ω)}, n N we have for N(t)(ω) > T N(t)(ω) (ω) N(t)(ω) t N(t)(ω) < T N(t)(ω)+1(ω) N(t)(ω) T N(t)(ω)+1(ω) N(t)(ω) + 1 N(t)(ω) + 1. (3) N(t)(ω)
16 16 CHAPTER 2. CLAIM NUMBER PROCESS MODELS Note that Ω {ω Ω : T 1 (ω) < } {ω Ω : sup N(t) > }. t Theorem implies that T n n EW 1 (4) holds on a set Ω with P(Ω ) 1. Hence lim n T n on Ω and by definition of N also lim t N(t) on Ω. From (4) we get Finally 3 implies lim t lim t T N(t)(ω) N(t)(ω) EW 1 for ω Ω. t N(t)(ω) EW 1 for ω Ω. In the following we will investigate the behavior of EN(t) as t. Theorem (Elementary renewal theorem). Assume the above setting, i.e. N(t) is a renewal process. If EW 1 <, then EN(t) lim t t 1 EW 1. (5) Remark If the W i s are exponentially distributed with parameter λ >, W i Exp(λ), i 1, 2,..., then N(t) is a Poisson process. Consequently, EN(t) λt. Since EW i 1, it follows that for all t > λ EN(t) t 1 EW 1. (6) If the W i s are not exponentially distributed, then the equation (6) holds only for the limit t.
17 2.2. THE RENEWAL PROCESS 17 In order to prove Theorem we formulate the following Lemma of Fatou type: Lemma Let Z (Z t ) t [, ) be a stochastic process such that Then Z t : Ω [, ) for all t. E lim inf t Z t lim inf t EZ t. Proof. By monotone convergence, since t inf s t Z s is nondecreasing, we have E lim t inf s t Z s lim t E inf s t Z s. Obviously, E inf s t Z s EZ u for all u t which allows us to write This implies the assertion. Proof of Theorem 2.2.4: E inf s t Z s inf u t EZ u. Let c 1 EW 1. From Theorem we conclude c lim t N(t) t N(s) lim inf t s t s a.s. Since Z t : inf s t N(s) s N(s) c E lim inf t s t s fulfills the requirements of Lemma we have E lim inf t inf N(s) s t s lim inf t E N(t).and t We only have to show that lim sup t E N(t) t W c i : W i c lim inf t E inf N(s) s t s c. For c > we define and get T c i : W c W c i T i, i 1, 2,...
18 18 CHAPTER 2. CLAIM NUMBER PROCESS MODELS Since N (c) (t) : #{i 1 : T c i lim sup t Assume we could show that t} N(t) we obtain EN(t) t lim sup t lim sup t EN (c) (t) t Then EW c 1 EW 1, for c implies We start showing (7). Let and lim sup t EN(t) t EN (c) (t). t 1. (7) EW1 c c. τ(ω) : N (c) (t)(ω) + 1 F n : σ(w 1,..., W n ), n 1, F : {, Ω}. The random variable τ is a stopping time w.r.t. (F n ) i.e. {τ n} {N (c) (t) + 1 n} F n Hence it follows by Wald s identity that This implies lim sup t EN (c) (t) t ET c N (c) (t)+1 E τ lim sup t lim sup t i1 W c i EτEW c 1. EN (c) (t) + 1 t Eτ t lim sup t lim sup t lim sup t ET c N (c) (t)+1 t EW1 c E ( W1 c W c N (t)) + EW c (c) N (c) (t)+1 t EW1 c t + c 1. t EW1 c EW1 c
19 2.3. THE INHOMOGENEOUS POISSON PROCESS The inhomogeneous Poisson process and the mixed Poisson process Definition Let µ : [, ) [, ) be a function such that 1. µ() 2. µ is nondecreasing, i.e. s t µ(s) µ(t) 3. µ is càdlàg. Then the function µ is called a meanvalue function. µ(t)t y t µ(t) continuous y t
20 2 CHAPTER 2. CLAIM NUMBER PROCESS MODELS µ(t) càdlàg y t Definition (Inhomogeneous Poisson process). A stochastic process N N(t) t [, ) is an inhomogeneous Poisson process if and only if it has the following properties: (P1) N() a.s. (P2) N has independent increments, i.e. if t < t 1 <... < t n, (n 1), it holds that N(t n ) N(t n 1 ), N(t n 1 ) N(t n 2 ),..., N(t 1 ) N(t ) are independent. (P inh. 3) There exists a meanvalue function µ such that for s < t (µ(t) µ(s)) (µ(t) µ(s))m P(N(t) N(s) m) e, m! where m, 1, 2,..., and t >. (P4) The paths of N are càdlàg a.s. Theorem (Time change for the Poisson process). If µ denotes the meanvalue function of an inhomogeneous Poisson process N and Ñ is a homogeneous Poisson process with λ 1, then (1) (N(t)) t [, ) d (Ñ(µ(t))) t [, ) (2) If µ is continuous, increasing and lim t µ(t), then N(µ 1 (t)) t [, ) d (Ñ(t)) t [, ).
21 2.3. THE INHOMOGENEOUS POISSON PROCESS Here µ 1 (t) denotes the inverse function of µ and f d g means that the two random variables f and g have the same distribution (but one can not conclude that f(ω) g(ω) for ω Ω). Definition (Mixed Poisson process). Let ˆN be a homogeneous Poisson process with intensity λ 1 and µ be a meanvalue function. Let θ : Ω R be a random variable such that θ > a.s., and θ is independent of ˆN. Then N(t) : ˆN(θµ(t)), t is a mixed Poisson process with mixing variable θ. Proposition It holds ( var( ˆN(θµ(t))) E ˆN(θµ(t)) 1 + var(θ) ) Eθ µ(t). Proof: We recall that E ˆN(t) var( ˆN(t)) t and therefore E ˆN(t) 2 t + t 2. We conclude var( ˆN(θµ(t))) E ˆN(θµ(t)) 2 [ E ˆN(θµ(t)) ] 2 E ( θµ(t) + θ 2 µ(t) 2) (Eθµ(t)) 2 µ(t) (Eθ + varθµ(t)). The property var(n(t)) > EN(t) is called overdispersion. inhomogeneous Poisson process, then If N is an var(n(t)) EN(t).
22 22 CHAPTER 2. CLAIM NUMBER PROCESS MODELS
23 3. The total claim amount process S(t) 3.1 The CramérLundbergmodel Definition The CramérLundbergmodel considers the following setting: 1. Claims happen at the claim arrival times < T 1 < T 2 <... of a Poisson process N(t) #{i 1 : T i t}, t. 2. At time T i the claim size X i happens and it holds that the sequence (X i ) i1 is i.i.d., X i. 3. The processes (T i ) i1 and (X i ) i1 are independent. Remark: Are N and (X i ) i1 independent? 3.2 The renewal model Definition The renewal model (or SparreAndersonmodel) considers the following setting: 1. Claims happen at the claim arrival times T 1 T 2... of a renewal process N(t) #{i 1 : T i t}, t. 23
24 24 CHAPTER 3. THE TOTAL CLAIM AMOUNT PROCESS S(T ) 2. At time T i the claim size X i happens and it holds that the sequence (X i ) i1 is i.i.d., X i. 3. The processes (T i ) i1 and (X i ) i1 are independent. 3.3 Properties of the total claim amount process S(t) Definition The total claim amount process is defined as N(t) S(t) : X i, t. i1 The insurance company needs information about S(t) in order to determine a premium which covers the losses represented by S(t). In general, the distribution of S(t), i.e. P({ω Ω : S(t, ω) x}), x, can only be approximated by numerical methods or simulations while ES(t) and var(s(t)) are easy to compute exactly. One can establish principles which use only ES(t) and var(s(t)) to calculate the premium. This will be done in chapter 4. Proposition (a) For the CramérLundbergmodel it holds (i) ES(t) λtex 1, (ii) var(s(t)) λtex 2 1. (b) Assume the renewal model. Let EW 1 1 λ (, ) and EX 1 <. (i) Then lim t ES(t) t λex 1. (ii) If var(w 1 ) < and var(x 1 ) <, then var(s(t)) lim t t λ ( var(x 1 ) + var(w 1 )λ 2 (EX 1 ) 2).
25 3.3. PROPERTIES OF S(T) 25 Proof: Since by direct computation, 1 1I Ω (ω) 1I {N(t)k}, k N(t) ES(t) E E i1 k X i ( ( k ) X i )1I {N(t)k} i1 E(X X k ) E1I } {{ } {N(t)k} } {{ } k kex 1 P(N(t)k) EX 1 kp(n(t) k) k EX 1 EN(t). In the CLmodel we have EN(t) λt. For the general case we use the Elementary Renewal Theorem (Thereom 2.2.4) to get the assertion. We continue with ES(t) 2 N(t) E E i1 X i 2 ( ( k ) ) 2 E X i 1I {N(t)k} k ( k ) 2 X i 1I {N(t)k} k k i,j1 EX 2 1 i1 k E ( ) X i X j 1I {N(t)k} i1 kp(n(t) k) + (EX 1 ) 2 k k1 EX 2 1 EN(t) + (EX 1 ) 2 (EN(t) 2 EN(t)) var(x 1 )EN(t) + (EX 1 ) 2 EN(t) 2. k(k 1)P(N(t) k)
26 26 CHAPTER 3. THE TOTAL CLAIM AMOUNT PROCESS S(T ) It follows that var(s(t)) ES(t) 2 (ES(t)) 2 ES(t) 2 (EX 1 ) 2 (EN(t)) 2 var(x 1 )EN(t) + (EX 1 ) 2 var(n(t)). For the CramérLundbergmodel it holds EN(t) var(n(t)) λt, hence we have var(s(t)) λt(var(x 1 ) + (EX 1 ) 2 ) λtex1. 2 For the renewal model we get var(x 1 )EN(t) lim var(x 1 )λ. t t The relation is shown in [5, Theorem 2.5.2]. var(n(t)) lim t t var(w 1) (EW 1 ) 3. Theorem The Strong Law of Large Numbers (SLLN) and the Central Limit Theorem (CLT) for (S(t)) in the renewal model can be stated as follows: (i) SLLN for (S(t)): If EW 1 1 λ < and EX 1 <, then S(t) lim t t λex 1 a.s. (ii) CLT for (S(t)): If var(w 1 ) <, and var(x 1 ) <, then ( ) P S(t) ES(t) x Φ(x) t, var(s(t)) sup x R where Φ is the distribution function of the standard normal distribution, Φ(x) 1 2π x e y2 2 dy. Proof: (i) We follow the proof of [6, Theorem ]. We have shown that N(t) lim t t λ a.s.
27 3.3. PROPERTIES OF S(T) 27 and it holds lim N(t) t a.s. Because of S(t) X 1 + X X N(t) and, by the SSLN we get S(t) lim t t (ii) See [4, Theorem ] X X n lim n n lim t N(t) t lim t EX 1 a.s., S(t) N(t) λex 1 a.s.
28 28 CHAPTER 3. THE TOTAL CLAIM AMOUNT PROCESS S(T )
29 4. Classical premium calculation principles The standard problem for the insurance companies is to determine that amount of premium such that the losses S(t) are covered. On the over hand the price should be low enough to be competitive and attract customers. A first approximation of S(t) is given by ES(t). For the premium income p(t) this implies p(t) < ES(t) insurance company loses on average p(t) > ES(t) insurance company gains on average A reasonable solution would be p(t) (1 + ρ)es(t) where ρ > is the safety loading. Proposition tells us that in the renewal model with EW 1 1 λ it holds ES(t) λt EX 1 for large t. 4.1 Used principles (1) The net principle, p NET (t) ES(t) defines the premium to be a fair market premium. This however, can be very risky for the company, which one can conclude from the Central Limit Theorem for S(t). (2) The expected value principle, p EV (t) (1 + ρ)es(t), which is motivated by the Strong Law of Large Numbers. 29
30 3 CHAPTER 4. PREMIUM CALCULATION PRINCIPLES (3) The variance principle, p V AR (t) ES(t) + αvar(s(t)), α >. This principle is in the renewal model asymptotically the same as p EV (t), since by Proposition we have that lim t p EV (t) p V AR (t) is a constant. This means that α plays the role of a safety loading ρ. (4) The standard deviation principle, p SD (t) ES(t) + α var(s(t)), α >.
31 5. Claim size distributions What distributions one should choose to model the claim sizes (X i )? If one analyzes data of claim sizes that have happened in the past, for example by a histogram or a QQplot, it turns out that the distribution is often heavytailed. Definition Let F (x) be the distribution function of X 1, i.e. F is called lighttailed for some λ >. F is called heavytailed for all λ >. 5.2 Examples F (x) P({ω Ω : X 1 (ω) x}). lim sup 1 F (x) n x n e λx lim n < inf 1 F (x) > x n e λx (1) The exponential distribution Exp(α) is lighttailed for all α >, since the distribution function is F (x) 1 e αx, x >, and 1 F (x) e λx and by choosing < λ < α, e αx e λx e(λ α)x, sup e (λ α)x e (λ α)n, as n. x n 31
32 32 CHAPTER 5. CLAIM SIZE DISTRIBUTIONS (2) The Pareto distribution is heavytailed. The distribution function is F (x) 1 or κ α, x, α >, κ >, (κ + x) α F (x) 1 ba, x b >, a >. xa 5.3 The QQplot A quantile is the inverse of the distribution function. We take the left inverse if the distribution function is not strictly increasing and continuous which is is defined by F (t) : inf{x R, F (x) t}, < t < 1, and the empirical distribution function of the data X 1,...X n as F n (x) : 1 n n 1I (,x] (X i ), x R. i1 It can be shown that if X 1 F, (X i ) i1 i.i.d., then lim F n (t) F (t), n almost surely for all continuity points t of F. Hence, if X 1 F, then the plot of (F n (t), F (t)) should give almost the straight line y x.
33 5.3. THE QQPLOT 33 y left inverse of F(x) F(x) x
34 34 CHAPTER 5. CLAIM SIZE DISTRIBUTIONS
35 6. About modern premium calculation principles 6.1 The exponential principle The exponential principle is defined as p exp (t) : 1 δ log EeδS(t), for some δ >, where δ is the risk aversion constant. The function p exp (t) is defined via the socalled utility theory. 6.2 The quantile principle Suppose F (x) P({ω : S(t) x}), x R, is the distribution function of S(t). In Section 5.3 we defined the left inverse of the distribution function F by F (y) : inf{x R : F (x) y}, < y < 1. Then the (1 ε) quantile principle is defined as p quant (t) F (1 ε), where the expression F (1 ε) converges for ε to the probable maximal loss. This setting is related to the theory of Value at Risk. 6.3 The Esscher principle The Esscher principle is defined as p Ess (t) ES(t)eδS(t) Ee δ(s(t)), δ >. 35
36 36 CHAPTER 6. MODERN PREMIUM CALCULATION PRINCIPLES In all the above principles the expected value E(g(S(t)) needs to be computed for a certain function g(x) to compute p(t). This means it is not enough to know ES(t) and var(s(t)), the distribution of S(t) is needed as well.
37 7. The distribution of the total claim amount S(t) Theorem Let (Ω, F, P) be a probability space. (a) The distribution of a random variable f : Ω R can be uniquely described by its distribution function F : R [, 1], F (x) : P({ω Ω : f(ω) x}), x R. (b) Especially, it holds for g : R R, such that g 1 (B) B(R), for all B B(R), that Eg(f) g(x)df (x) R (in the sense that, if either side of this expression exists, so does the other, and then they are equal, see [7], pp ). (c) The distribution of f can also be determined by its characteristic function, (see [9]) ϕ f (u) : Ee iuf, u R, or by its momentgenerating function m f (h) : Ee hf, h ( h, h ) provided that Ee h f < for some h >. Remember: for independent random variables f and g it holds ϕ f+g (u) ϕ f (u)ϕ g (u). 37
38 38 CHAPTER 7. THE DISTRIBUTION OF S(T ) 7.2 Mixture distributions Definition (Mixture distributions). Let F i, i 1,..., n be distribution functions and p i [, 1] such that n i1 p i 1. Then G(x) p 1 F 1 (x) p n F n (x), x R, is called the mixture distribution of F 1,..., F n. Lemma Let f 1,..., f n be random variables with distribution function F 1,..., F n, respectively. Assume that J : Ω {1,..., n} is independent from f 1,..., f n and P(J i) p i. Then the random variable Z 1I {J1} f I {Jn} f n has the mixture distribution function G. Definition (Compound Poisson random variable). Let N λ P ois(λ) and (X i ) i1 i.i.d. random variables, independent from N λ. Then N λ Z : i1 is called a compound Poisson random variable. Proposition The sum of independent compound Poisson random variables is a compound Poisson random variable: Let S 1,..., S n given by j1 X i N k S k X (k) j, k 1,..., n be independent compound Poisson random variables such that N k P ois(λ k ), λ k >, (X (k) j ) j 1 i.i.d., and N k is independent from (X (k) j ) j 1 for all k 1,..., n. Then S : S S n is a compound Poisson random variable with representation N λ S d Y l, N λ P ois(λ), λ λ λ n l1
39 7.2. MIXTURE DISTRIBUTIONS 39 and (Y l ) l 1 is an i.i.d. sequence, independent from N λ and Y 1 d n k1 1I {Jk} X (k) 1, with P(J k) λ k λ, and J is independent of (X (k) 1 ) k. Proof: From Theorem we know that it is sufficient to show that S and N λ l1 Y l have the same characteristic function. We start with the characteristic function of S k : Then ϕ Sk (u) Ee ius k Ee iu N k j1 X(k) j E e iu m j1 X(k) j m 1I {Nk m} E e iux(k) 1... e iux(k) m 1I {Nk m} } {{ } m all of these are independent ) m (Ee iux(k) 1 P(Nk m) m ( m ( m ϕ X (k) 1 ϕ X (k) 1 (u)) m P(Nk m) ) m λ m k (u) ϕ S (u) Ee iu(s S n) Ee ius 1... Ee iusn ϕ S1 (u)... ϕ Sn (u) e λ 1(1 ϕ (1) X 1 ( ( exp λ 1 m! e λ k e λk(1 ϕ X (k)(u)) 1. (u)) λ n(1 ϕ (n) (u))... e X 1 n k1 λ k λ ϕ X (k) 1 (u)) ).
40 4 CHAPTER 7. THE DISTRIBUTION OF S(T ) Let ξ N λ l1 Y l. Then by the same computation as we have done for ϕ Sk (u) we get Finally, ϕ ξ (u) Ee iuξ e λ(1 ϕ Y 1 (u)). ϕ Y1 (u) Ee iu n k1 1I {Jk}X (k) 1 n ( E e iu ) n k1 1I {Jk}X (k) 1 1I {Jl} l1 l1 n ) E (e iux(l) 1 1I{Jl} n l1 ϕ X (l) 1 (u)) λ l λ. 7.3 Applications in insurance First application Assume that the claims arrive according to an inhomogeneous Poisson process, i.e. N(t) N(s) P ois(µ(t) µ(s)). The total claim amount in year l is S l Now, it can be seen, that S l d N(l) jn(l 1)+1 N(l) N(l 1) j1 X (l) j, l 1,..., n. X (l) j, l 1,..., n and S l is compound Poisson distributed. Proposition implies that the total claim amount of the first n years is again compound Poisson distributed,
41 7.3. APPLICATIONS IN INSURANCE 41 where N λ S(n) : S S d n N λ P ois(µ(n)) i1 Y d i 1I {J1} X (1) I {Jn} X (n) 1 µ(i) µ(i 1) P(J i). µ(n) Hence the total claim amount S(n) in the first n years (with possibly different claim size distributions in each year) has a representation as a compound Poisson random variable. Second application We can interprete the random variables N i S i X (i) j, N i P ois(λ i ), i 1,..., n, j1 as the total claim amounts of n independent portfolios for the same fixed period of time. The (X (i) j ) j 1 in the ith portfolio are i.i.d, but the distributions may differ from portfolio to portfolio (one particular type of car insurance, for example). Then N λ S(n) S S d n is again compound Poisson distributed with N λ P ois(λ λ n ) i1 Y i Y i Y d i 1I {J1} X (1) I {Jn} X (n) 1 and P(J l) λ l λ.
42 42 CHAPTER 7. THE DISTRIBUTION OF S(T ) 7.4 The Panjer recursion: an exact numerical procedure to calculate F S(t) Let S N X i, i1 N : Ω {, 1,...} and (X i ) i 1 i.i.d, N and (X i ) independent. Then, setting S :, S n : X X n, n 1 yields P(S x) P(S x, N n) n P(S x N n)p(n n) n P(S n x)p(n n) n n F n X 1 (x)p(n n), where F n X 1 (x) is the nth convolution of F X1, i.e. F 2 X 1 (x) P(X 1 + X 2 x) E1I {X1 +X 2 x} X 1,X 2 independent 1I {x1 +x 2 x}(x 1, x 2 )df X1 (x 1 )df X2 (x 2 ) R R 1I {x1 x x 2 }(x 1, x 2 )df X1 (x 1 )df X2 (x 2 ) R R F X1 (x x 2 )df X2 (x 2 ) and by recursion using F X1 F X2, F (n+1) X 1 (x) : R R F n X 1 (x y)df X1 (y). But the computation of F n X 1 (x) is numerically difficult. However, there is a recursion formula for P(S x) that holds under certain conditions:
43 7.4. THE PANJER RECURSION 43 Theorem (Panjer recursion scheme). Assume the following conditions: (C1) X i : Ω {, 1,...} (C2) for N it holds that Then for Proof: for some a, b R. q n P(N n) ( a + b ) q n 1, n 1, 2,... n p n : P(S n), n, 1, 2,... { q p, if P(X 1 ) EP(X 1 ) N (1), otherwise 1 n ( p n a + bi ) P(X 1 i)p n i, n 1. (2) 1 ap(x 1 ) n i1 This implies (1). For p n, n 1, p P(S ) P(S, N ) + P(S, N > ) P(S ) P(N ) + P(S, N > ) } {{ } 1 P(N ) +P(S, N > ) } {{ } q q }{{} + P(X 1 ) P(N) EP(X 1 ) N. P(X X k, N k) } {{ } P(X 1 ) k P(N k) } {{ } q k k1 p n P(S n) P(S k n)q k k1
44 44 CHAPTER 7. THE DISTRIBUTION OF S(T ) (C2) P(S k n)(a + b k )q k 1. (3) k1 Assume P(S k n) >. Now, because Q P( S k n) is a probability measure the following holds. n ( a + bl ) P(X 1 l S k n) n } {{ } l Q(X 1 l) a + b n E QX 1 a + b nk E Q(X X k ) a + b nk E } QS {{ } k n a + b k, (4) where the last equation yields from the fact that Q(S k n) 1. On the other hand, we can express the term a + b also by k n ( a + bl ) P(X 1 l S k n) n l n (a + bl n )P(X 1 l, S k X 1 n l) P(S k n) l n (a + bl n )P(X 1 l)p(s k 1 n l). P(S k n) (5) l Thanks to (4) we can now replace the term a + b in (3) by the RHS of (5) k which yields p n k1 n l n l ( a + bl n ( a + bl n ) P(X 1 l)p(s k 1 n l)q k 1 ) P(X 1 l) P(S k 1 n l)q k 1 k1 } {{ } P(Sn l)
45 7.5. APPROXIMATION OF F S(T ) 45 ap(x 1 )P(S n) + ap(x 1 )p n + n l1 which will give the equation (2) p n Remark ap(x 1 ) n l1 ( a + bl n n l1 ( a + bl n ) P(X 1 l)p(s n l) ) P(X 1 l)p n l, ( a + bl ) P(X 1 l)p n l n The Panjer recursion only works for distributions of X i on {, 1, 2,...} i.e. k P X i (k) 1 (or, by scaling, on a lattice {, d, 2d,...} for d > fixed). Traditionally, the distributions used to model X i have a density, and {,1,2,...} h x i (x)dx. But on the other hand, claim sizes are expressed in terms of prices, so they take values on a lattice. The density h Xi (x) could be approximated to have a distribution on a lattice, but how large would the approximation error then be? N can only be Poisson, binomially or negative binomially distributed. 7.5 Approximation of F S(t) using the Central Limit Theorem Assume, that the renewal model is used, and that N(t) S(t) X i, t. i1 In Theorem the Central Limit Theorem is used to state that if var(w 1 ) < and var(x 1 ) <, then ( ) P S(t) ES(t) x Φ(x) t. var(s(t)) sup x R
General theory of stochastic processes
CHAPTER 1 General theory of stochastic processes 1.1. Definition of stochastic process First let us recall the definition of a random variable. A random variable is a random number appearing as a result
More informationRenewal Theory. (iv) For s < t, N(t) N(s) equals the number of events in (s, t].
Renewal Theory Def. A stochastic process {N(t), t 0} is said to be a counting process if N(t) represents the total number of events that have occurred up to time t. X 1, X 2,... times between the events
More informationTHE CENTRAL LIMIT THEOREM TORONTO
THE CENTRAL LIMIT THEOREM DANIEL RÜDT UNIVERSITY OF TORONTO MARCH, 2010 Contents 1 Introduction 1 2 Mathematical Background 3 3 The Central Limit Theorem 4 4 Examples 4 4.1 Roulette......................................
More informationInsurance models and riskfunction premium principle
Insurance models and riskfunction premium principle Aditya Challa Supervisor : Prof. Vassili Kolokoltsov July 2, 212 Abstract Insurance sector was developed based on the idea to protect people from random
More informationLecture 4: Random Variables
Lecture 4: Random Variables 1. Definition of Random variables 1.1 Measurable functions and random variables 1.2 Reduction of the measurability condition 1.3 Transformation of random variables 1.4 σalgebra
More informationProbability Generating Functions
page 39 Chapter 3 Probability Generating Functions 3 Preamble: Generating Functions Generating functions are widely used in mathematics, and play an important role in probability theory Consider a sequence
More informationAsymptotics of discounted aggregate claims for renewal risk model with risky investment
Appl. Math. J. Chinese Univ. 21, 25(2: 29216 Asymptotics of discounted aggregate claims for renewal risk model with risky investment JIANG Tao Abstract. Under the assumption that the claim size is subexponentially
More information3. Renewal Theory. Definition 3 of the Poisson process can be generalized: Let X 1, X 2,..., iidf(x) be nonnegative interarrival times.
3. Renewal Theory Definition 3 of the Poisson process can be generalized: Let X 1, X 2,..., iidf(x) be nonnegative interarrival times. Set S n = n i=1 X i and N(t) = max {n : S n t}. Then {N(t)} is a
More informationExponential Distribution
Exponential Distribution Definition: Exponential distribution with parameter λ: { λe λx x 0 f(x) = 0 x < 0 The cdf: F(x) = x Mean E(X) = 1/λ. f(x)dx = Moment generating function: φ(t) = E[e tx ] = { 1
More informatione.g. arrival of a customer to a service station or breakdown of a component in some system.
Poisson process Events occur at random instants of time at an average rate of λ events per second. e.g. arrival of a customer to a service station or breakdown of a component in some system. Let N(t) be
More informationThe Exponential Distribution
21 The Exponential Distribution From DiscreteTime to ContinuousTime: In Chapter 6 of the text we will be considering Markov processes in continuous time. In a sense, we already have a very good understanding
More information3. Continuous Random Variables
3. Continuous Random Variables A continuous random variable is one which can take any value in an interval (or union of intervals) The values that can be taken by such a variable cannot be listed. Such
More informationSTAT 571 Assignment 1 solutions
STAT 571 Assignment 1 solutions 1. If Ω is a set and C a collection of subsets of Ω, let A be the intersection of all σalgebras that contain C. rove that A is the σalgebra generated by C. Solution: Let
More informationWhat is Statistics? Lecture 1. Introduction and probability review. Idea of parametric inference
0. 1. Introduction and probability review 1.1. What is Statistics? What is Statistics? Lecture 1. Introduction and probability review There are many definitions: I will use A set of principle and procedures
More informationNonLife Insurance Mathematics
Thomas Mikosch NonLife Insurance Mathematics An Introduction with the Poisson Process Second Edition 4y Springer Contents Part I Collective Risk Models 1 The Basic Model 3 2 Models for the Claim Number
More informationA Uniform Asymptotic Estimate for Discounted Aggregate Claims with Subexponential Tails
12th International Congress on Insurance: Mathematics and Economics July 1618, 2008 A Uniform Asymptotic Estimate for Discounted Aggregate Claims with Subexponential Tails XUEMIAO HAO (Based on a joint
More informationIntroduction to Probability
Introduction to Probability EE 179, Lecture 15, Handout #24 Probability theory gives a mathematical characterization for experiments with random outcomes. coin toss life of lightbulb binary data sequence
More informationAggregate Loss Models
Aggregate Loss Models Chapter 9 Stat 477  Loss Models Chapter 9 (Stat 477) Aggregate Loss Models Brian Hartman  BYU 1 / 22 Objectives Objectives Individual risk model Collective risk model Computing
More information6. Metric spaces. In this section we review the basic facts about metric spaces. d : X X [0, )
6. Metric spaces In this section we review the basic facts about metric spaces. Definitions. A metric on a nonempty set X is a map with the following properties: d : X X [0, ) (i) If x, y X are points
More informationImportant Probability Distributions OPRE 6301
Important Probability Distributions OPRE 6301 Important Distributions... Certain probability distributions occur with such regularity in reallife applications that they have been given their own names.
More informationLecture 8: Random Walk vs. Brownian Motion, Binomial Model vs. LogNormal Distribution
Lecture 8: Random Walk vs. Brownian Motion, Binomial Model vs. Logormal Distribution October 4, 200 Limiting Distribution of the Scaled Random Walk Recall that we defined a scaled simple random walk last
More informationLecture 13: Martingales
Lecture 13: Martingales 1. Definition of a Martingale 1.1 Filtrations 1.2 Definition of a martingale and its basic properties 1.3 Sums of independent random variables and related models 1.4 Products of
More informationAsymptotics for a discretetime risk model with Gammalike insurance risks. Pokfulam Road, Hong Kong
Asymptotics for a discretetime risk model with Gammalike insurance risks Yang Yang 1,2 and Kam C. Yuen 3 1 Department of Statistics, Nanjing Audit University, Nanjing, 211815, China 2 School of Economics
More informationPast and present trends in aggregate claims analysis
Past and present trends in aggregate claims analysis Gordon E. Willmot Munich Re Professor of Insurance Department of Statistics and Actuarial Science University of Waterloo 1st QuebecOntario Workshop
More information6.4 The Infinitely Many Alleles Model
6.4. THE INFINITELY MANY ALLELES MODEL 93 NE µ [F (X N 1 ) F (µ)] = + 1 i
More informationSequences of Functions
Sequences of Functions Uniform convergence 9. Assume that f n f uniformly on S and that each f n is bounded on S. Prove that {f n } is uniformly bounded on S. Proof: Since f n f uniformly on S, then given
More informationExercises with solutions (1)
Exercises with solutions (). Investigate the relationship between independence and correlation. (a) Two random variables X and Y are said to be correlated if and only if their covariance C XY is not equal
More informationSection 5.1 Continuous Random Variables: Introduction
Section 5. Continuous Random Variables: Introduction Not all random variables are discrete. For example:. Waiting times for anything (train, arrival of customer, production of mrna molecule from gene,
More informationIEOR 6711: Stochastic Models, I Fall 2012, Professor Whitt, Final Exam SOLUTIONS
IEOR 6711: Stochastic Models, I Fall 2012, Professor Whitt, Final Exam SOLUTIONS There are four questions, each with several parts. 1. Customers Coming to an Automatic Teller Machine (ATM) (30 points)
More informationOptimal reinsurance with ruin probability target
Optimal reinsurance with ruin probability target Arthur Charpentier 7th International Workshop on Rare Event Simulation, Sept. 2008 http ://blogperso.univrennes1.fr/arthur.charpentier/ 1 Ruin, solvency
More information3. The Economics of Insurance
3. The Economics of Insurance Insurance is designed to protect against serious financial reversals that result from random evens intruding on the plan of individuals. Limitations on Insurance Protection
More informationMathematical Finance
Mathematical Finance Option Pricing under the RiskNeutral Measure Cory Barnes Department of Mathematics University of Washington June 11, 2013 Outline 1 Probability Background 2 Black Scholes for European
More informationProblem Set. Problem Set #2. Math 5322, Fall December 3, 2001 ANSWERS
Problem Set Problem Set #2 Math 5322, Fall 2001 December 3, 2001 ANSWERS i Problem 1. [Problem 18, page 32] Let A P(X) be an algebra, A σ the collection of countable unions of sets in A, and A σδ the collection
More informationCHAPTER II THE LIMIT OF A SEQUENCE OF NUMBERS DEFINITION OF THE NUMBER e.
CHAPTER II THE LIMIT OF A SEQUENCE OF NUMBERS DEFINITION OF THE NUMBER e. This chapter contains the beginnings of the most important, and probably the most subtle, notion in mathematical analysis, i.e.,
More informationPUTNAM TRAINING POLYNOMIALS. Exercises 1. Find a polynomial with integral coefficients whose zeros include 2 + 5.
PUTNAM TRAINING POLYNOMIALS (Last updated: November 17, 2015) Remark. This is a list of exercises on polynomials. Miguel A. Lerma Exercises 1. Find a polynomial with integral coefficients whose zeros include
More informationContinuous Random Variables
Continuous Random Variables COMP 245 STATISTICS Dr N A Heard Contents 1 Continuous Random Variables 2 11 Introduction 2 12 Probability Density Functions 3 13 Transformations 5 2 Mean, Variance and Quantiles
More informationWorked examples Random Processes
Worked examples Random Processes Example 1 Consider patients coming to a doctor s office at random points in time. Let X n denote the time (in hrs) that the n th patient has to wait before being admitted
More informationStochastic Inventory Control
Chapter 3 Stochastic Inventory Control 1 In this chapter, we consider in much greater details certain dynamic inventory control problems of the type already encountered in section 1.3. In addition to the
More informationMathematics Course 111: Algebra I Part IV: Vector Spaces
Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 19967 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are
More informationContinuous random variables
Continuous random variables So far we have been concentrating on discrete random variables, whose distributions are not continuous. Now we deal with the socalled continuous random variables. A random
More informationINSURANCE RISK THEORY (Problems)
INSURANCE RISK THEORY (Problems) 1 Counting random variables 1. (Lack of memory property) Let X be a geometric distributed random variable with parameter p (, 1), (X Ge (p)). Show that for all n, m =,
More informationConcentration inequalities for order statistics Using the entropy method and Rényi s representation
Concentration inequalities for order statistics Using the entropy method and Rényi s representation Maud Thomas 1 in collaboration with Stéphane Boucheron 1 1 LPMA Université ParisDiderot High Dimensional
More informationsheng@mail.ncyu.edu.tw 1 Content Introduction Expectation and variance of continuous random variables Normal random variables Exponential random variables Other continuous distributions The distribution
More informationA spot price model feasible for electricity forward pricing Part II
A spot price model feasible for electricity forward pricing Part II Fred Espen Benth Centre of Mathematics for Applications (CMA) University of Oslo, Norway Wolfgang Pauli Institute, Wien January 1718
More informationPricing catastrophe options in incomplete market
Pricing catastrophe options in incomplete market Arthur Charpentier arthur.charpentier@univrennes1.fr Actuarial and Financial Mathematics Conference Interplay between finance and insurance, February 2008
More informationCourse 221: Analysis Academic year , First Semester
Course 221: Analysis Academic year 200708, First Semester David R. Wilkins Copyright c David R. Wilkins 1989 2007 Contents 1 Basic Theorems of Real Analysis 1 1.1 The Least Upper Bound Principle................
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES Contents 1. Random variables and measurable functions 2. Cumulative distribution functions 3. Discrete
More informationOverview of Monte Carlo Simulation, Probability Review and Introduction to Matlab
Monte Carlo Simulation: IEOR E4703 Fall 2004 c 2004 by Martin Haugh Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab 1 Overview of Monte Carlo Simulation 1.1 Why use simulation?
More informationMarshallOlkin distributions and portfolio credit risk
MarshallOlkin distributions and portfolio credit risk Moderne Finanzmathematik und ihre Anwendungen für Banken und Versicherungen, Fraunhofer ITWM, Kaiserslautern, in Kooperation mit der TU München und
More information5. Convergence of sequences of random variables
5. Convergence of sequences of random variables Throughout this chapter we assume that {X, X 2,...} is a sequence of r.v. and X is a r.v., and all of them are defined on the same probability space (Ω,
More informationInfinite product spaces
Chapter 7 Infinite product spaces So far we have talked a lot about infinite sequences of independent random variables but we have never shown how to construct such sequences within the framework of probability/measure
More informationMathematics for Econometrics, Fourth Edition
Mathematics for Econometrics, Fourth Edition Phoebus J. Dhrymes 1 July 2012 1 c Phoebus J. Dhrymes, 2012. Preliminary material; not to be cited or disseminated without the author s permission. 2 Contents
More informationPOISSON PROCESS AND INSURANCE : AN INTRODUCTION 1
POISSON PROCESS AND INSURANCE : AN INTRODUCTION S.RAMASUBRAMANIAN Statistics and Mathematics Unit Indian Statistical Institute 8th Mile, Mysore Road Bangalore  560059. Abstract: Basic aspects of the classical
More informationChapter 4 Lecture Notes
Chapter 4 Lecture Notes Random Variables October 27, 2015 1 Section 4.1 Random Variables A random variable is typically a realvalued function defined on the sample space of some experiment. For instance,
More informationDefinition: Suppose that two random variables, either continuous or discrete, X and Y have joint density
HW MATH 461/561 Lecture Notes 15 1 Definition: Suppose that two random variables, either continuous or discrete, X and Y have joint density and marginal densities f(x, y), (x, y) Λ X,Y f X (x), x Λ X,
More informationChapter 7. Induction and Recursion. Part 1. Mathematical Induction
Chapter 7. Induction and Recursion Part 1. Mathematical Induction The principle of mathematical induction is this: to establish an infinite sequence of propositions P 1, P 2, P 3,..., P n,... (or, simply
More informationNonparametric adaptive age replacement with a onecycle criterion
Nonparametric adaptive age replacement with a onecycle criterion P. CoolenSchrijner, F.P.A. Coolen Department of Mathematical Sciences University of Durham, Durham, DH1 3LE, UK email: Pauline.Schrijner@durham.ac.uk
More informationSolution Using the geometric series a/(1 r) = x=1. x=1. Problem For each of the following distributions, compute
Math 472 Homework Assignment 1 Problem 1.9.2. Let p(x) 1/2 x, x 1, 2, 3,..., zero elsewhere, be the pmf of the random variable X. Find the mgf, the mean, and the variance of X. Solution 1.9.2. Using the
More informationUNIFORM ASYMPTOTICS FOR DISCOUNTED AGGREGATE CLAIMS IN DEPENDENT RISK MODELS
Applied Probability Trust 2 October 2013 UNIFORM ASYMPTOTICS FOR DISCOUNTED AGGREGATE CLAIMS IN DEPENDENT RISK MODELS YANG YANG, Nanjing Audit University, and Southeast University KAIYONG WANG, Southeast
More informationOptimisation Problems in NonLife Insurance
Frankfurt, 6. Juli 2007 1 The de Finetti Problem The Optimal Strategy De Finetti s Example 2 Minimal Ruin Probabilities The HamiltonJacobiBellman Equation Two Examples 3 Optimal Dividends Dividends in
More informationRecitation 4. 24xy for 0 < x < 1, 0 < y < 1, x + y < 1 0 elsewhere
Recitation. Exercise 3.5: If the joint probability density of X and Y is given by xy for < x
More informationPoisson processes (and mixture distributions)
Poisson processes (and mixture distributions) James W. Daniel Austin Actuarial Seminars www.actuarialseminars.com June 26, 2008 c Copyright 2007 by James W. Daniel; reproduction in whole or in part without
More informationAdvanced Probability
Advanced Probability Perla Sousi October 13, 2013 Contents 1 Conditional expectation 3 1.1 Discrete case................................... 4 1.2 Existence and uniqueness............................. 5
More informationA Note on the Ruin Probability in the Delayed Renewal Risk Model
Southeast Asian Bulletin of Mathematics 2004 28: 1 5 Southeast Asian Bulletin of Mathematics c SEAMS. 2004 A Note on the Ruin Probability in the Delayed Renewal Risk Model Chun Su Department of Statistics
More informationRieszFredhölm Theory
RieszFredhölm Theory T. Muthukumar tmk@iitk.ac.in Contents 1 Introduction 1 2 Integral Operators 1 3 Compact Operators 7 4 Fredhölm Alternative 14 Appendices 18 A AscoliArzelá Result 18 B Normed Spaces
More informationMeasure Theory and Probability
Chapter 2 Measure Theory and Probability 2.1 Introduction In advanced probability courses, a random variable (r.v.) is defined rigorously through measure theory, which is an indepth mathematics discipline
More informationSECONDORDER LINEAR HOMOGENEOUS DIFFERENTIAL EQUATIONS
L SECONDORDER LINEAR HOOGENEOUS DIFFERENTIAL EQUATIONS SECONDORDER LINEAR HOOGENEOUS DIFFERENTIAL EQUATIONS WITH CONSTANT COEFFICIENTS A secondorder linear differential equation is one of the form d
More informationCHAPTER 5. Product Measures
54 CHAPTER 5 Product Measures Given two measure spaces, we may construct a natural measure on their Cartesian product; the prototype is the construction of Lebesgue measure on R 2 as the product of Lebesgue
More informationStatistics 100A Homework 8 Solutions
Part : Chapter 7 Statistics A Homework 8 Solutions Ryan Rosario. A player throws a fair die and simultaneously flips a fair coin. If the coin lands heads, then she wins twice, and if tails, the onehalf
More informationIntroduction to MonteCarlo Methods
Introduction to MonteCarlo Methods Bernard Lapeyre Halmstad January 2007 MonteCarlo methods are extensively used in financial institutions to compute European options prices to evaluate sensitivities
More informationChapter 4 Expected Values
Chapter 4 Expected Values 4. The Expected Value of a Random Variables Definition. Let X be a random variable having a pdf f(x). Also, suppose the the following conditions are satisfied: x f(x) converges
More informationThe Basics of Financial Mathematics. Spring Richard F. Bass Department of Mathematics University of Connecticut
The Basics of Financial Mathematics Spring 23 Richard F. Bass Department of Mathematics University of Connecticut These notes are c 23 by Richard Bass. They may be used for personal use or class use, but
More informationSums of Independent Random Variables
Chapter 7 Sums of Independent Random Variables 7.1 Sums of Discrete Random Variables In this chapter we turn to the important question of determining the distribution of a sum of independent random variables
More informationConditional Tail Expectations for Multivariate Phase Type Distributions
Conditional Tail Expectations for Multivariate Phase Type Distributions Jun Cai Department of Statistics and Actuarial Science University of Waterloo Waterloo, ON N2L 3G1, Canada jcai@math.uwaterloo.ca
More informationModule MA1S11 (Calculus) Michaelmas Term 2016 Section 3: Functions
Module MA1S11 (Calculus) Michaelmas Term 2016 Section 3: Functions D. R. Wilkins Copyright c David R. Wilkins 2016 Contents 3 Functions 43 3.1 Functions between Sets...................... 43 3.2 Injective
More informationEXIT TIME PROBLEMS AND ESCAPE FROM A POTENTIAL WELL
EXIT TIME PROBLEMS AND ESCAPE FROM A POTENTIAL WELL Exit Time problems and Escape from a Potential Well Escape From a Potential Well There are many systems in physics, chemistry and biology that exist
More information1 if 1 x 0 1 if 0 x 1
Chapter 3 Continuity In this chapter we begin by defining the fundamental notion of continuity for real valued functions of a single real variable. When trying to decide whether a given function is or
More informationModern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem
More informationFunction Sequences and Series
Chapter 7 Function Sequences and Series 7.1 Complex Numbers This is not meant to be a thourogh and completely rigorous construction of complex numbers but rather a quick brush up of things assumed to be
More informationk=1 k2, and therefore f(m + 1) = f(m) + (m + 1) 2 =
Math 104: Introduction to Analysis SOLUTIONS Alexander Givental HOMEWORK 1 1.1. Prove that 1 2 +2 2 + +n 2 = 1 n(n+1)(2n+1) for all n N. 6 Put f(n) = n(n + 1)(2n + 1)/6. Then f(1) = 1, i.e the theorem
More informationJoint Exam 1/P Sample Exam 1
Joint Exam 1/P Sample Exam 1 Take this practice exam under strict exam conditions: Set a timer for 3 hours; Do not stop the timer for restroom breaks; Do not look at your notes. If you believe a question
More informationTHE PRIME NUMBER THEOREM
THE PRIME NUMBER THEOREM NIKOLAOS PATTAKOS. introduction In number theory, this Theorem describes the asymptotic distribution of the prime numbers. The Prime Number Theorem gives a general description
More informationFrom Ruin Theory to Solvency in NonLife Insurance
From Ruin Theory to Solvency in NonLife Insurance Mario V. Wüthrich RiskLab ETH Zurich & Swiss Finance Institute SFI January 23, 2014 LUH Colloquium Versicherungs und Finanzmathematik Leibniz Universität
More informationMATH 425, PRACTICE FINAL EXAM SOLUTIONS.
MATH 45, PRACTICE FINAL EXAM SOLUTIONS. Exercise. a Is the operator L defined on smooth functions of x, y by L u := u xx + cosu linear? b Does the answer change if we replace the operator L by the operator
More informationChapter 2: Binomial Methods and the BlackScholes Formula
Chapter 2: Binomial Methods and the BlackScholes Formula 2.1 Binomial Trees We consider a financial market consisting of a bond B t = B(t), a stock S t = S(t), and a calloption C t = C(t), where the
More informationDefinition 6.1.1. A r.v. X has a normal distribution with mean µ and variance σ 2, where µ R, and σ > 0, if its density is f(x) = 1. 2σ 2.
Chapter 6 Brownian Motion 6. Normal Distribution Definition 6... A r.v. X has a normal distribution with mean µ and variance σ, where µ R, and σ > 0, if its density is fx = πσ e x µ σ. The previous definition
More informationVERTICES OF GIVEN DEGREE IN SERIESPARALLEL GRAPHS
VERTICES OF GIVEN DEGREE IN SERIESPARALLEL GRAPHS MICHAEL DRMOTA, OMER GIMENEZ, AND MARC NOY Abstract. We show that the number of vertices of a given degree k in several kinds of seriesparallel labelled
More informationThe Delta Method and Applications
Chapter 5 The Delta Method and Applications 5.1 Linear approximations of functions In the simplest form of the central limit theorem, Theorem 4.18, we consider a sequence X 1, X,... of independent and
More informationLecture Notes 1. Brief Review of Basic Probability
Probability Review Lecture Notes Brief Review of Basic Probability I assume you know basic probability. Chapters 3 are a review. I will assume you have read and understood Chapters 3. Here is a very
More information1. Prove that the empty set is a subset of every set.
1. Prove that the empty set is a subset of every set. Basic Topology Written by MenGen Tsai email: b89902089@ntu.edu.tw Proof: For any element x of the empty set, x is also an element of every set since
More informationSensitivity analysis of European options in jumpdiffusion models via the Malliavin calculus on the Wiener space
Sensitivity analysis of European options in jumpdiffusion models via the Malliavin calculus on the Wiener space Virginie Debelley and Nicolas Privault Département de Mathématiques Université de La Rochelle
More informationLINEAR ALGEBRA W W L CHEN
LINEAR ALGEBRA W W L CHEN c W W L Chen, 1997, 2008 This chapter is available free to all individuals, on understanding that it is not to be used for financial gain, and may be downloaded and/or photocopied,
More informationSensitivity analysis of utility based prices and risktolerance wealth processes
Sensitivity analysis of utility based prices and risktolerance wealth processes Dmitry Kramkov, Carnegie Mellon University Based on a paper with Mihai Sirbu from Columbia University Math Finance Seminar,
More informationPoisson Processes. Chapter 5. 5.1 Exponential Distribution. The gamma function is defined by. Γ(α) = t α 1 e t dt, α > 0.
Chapter 5 Poisson Processes 5.1 Exponential Distribution The gamma function is defined by Γ(α) = t α 1 e t dt, α >. Theorem 5.1. The gamma function satisfies the following properties: (a) For each α >
More informationA PRIORI ESTIMATES FOR SEMISTABLE SOLUTIONS OF SEMILINEAR ELLIPTIC EQUATIONS. In memory of RouHuai Wang
A PRIORI ESTIMATES FOR SEMISTABLE SOLUTIONS OF SEMILINEAR ELLIPTIC EQUATIONS XAVIER CABRÉ, MANEL SANCHÓN, AND JOEL SPRUCK In memory of RouHuai Wang 1. Introduction In this note we consider semistable
More informationThe Discrete Binomial Model for Option Pricing
The Discrete Binomial Model for Option Pricing Rebecca Stockbridge Program in Applied Mathematics University of Arizona May 4, 2008 Abstract This paper introduces the notion of option pricing in the context
More informationThe BlackScholesMerton Approach to Pricing Options
he BlackScholesMerton Approach to Pricing Options Paul J Atzberger Comments should be sent to: atzberg@mathucsbedu Introduction In this article we shall discuss the BlackScholesMerton approach to determining
More information4: SINGLEPERIOD MARKET MODELS
4: SINGLEPERIOD MARKET MODELS Ben Goldys and Marek Rutkowski School of Mathematics and Statistics University of Sydney Semester 2, 2015 B. Goldys and M. Rutkowski (USydney) Slides 4: SinglePeriod Market
More informationON LIMIT LAWS FOR CENTRAL ORDER STATISTICS UNDER POWER NORMALIZATION. E. I. Pancheva, A. GacovskaBarandovska
Pliska Stud. Math. Bulgar. 22 (2015), STUDIA MATHEMATICA BULGARICA ON LIMIT LAWS FOR CENTRAL ORDER STATISTICS UNDER POWER NORMALIZATION E. I. Pancheva, A. GacovskaBarandovska Smirnov (1949) derived four
More informationMATH 201. Final ANSWERS August 12, 2016
MATH 01 Final ANSWERS August 1, 016 Part A 1. 17 points) A bag contains three different types of dice: four 6sided dice, five 8sided dice, and six 0sided dice. A die is drawn from the bag and then rolled.
More information