# Exclusive OR (XOR) and hardware random number generators

Save this PDF as:

Size: px
Start display at page:

## Transcription

1 Exclusive OR (XOR) and hardware random number generators Robert B Davies February 28, Introduction The exclusive or (XOR) operation is commonly used to reduce the bias from the bits generated by a hardware random number generator. Typically,the uncorrected bits generated by a hardware random number generator will have expectation (long run average value) different from the ideal value of 1 2. Also, adjacent bits may be correlated. In this note, I find the expectations and correlations of various combinations of random bits using the XOR operator under a variety of assumptions about the means and correlations of the original variables. Specifically, I am interested in the effectiveness of the XOR operator for reducing bias (the deviation of the expectation from 1 2 ), and what happens when the successive bits are correlated. The symbols X, Y, Z etc. will denote random bits (taking the values 0 or 1 in probability theory they would be known as the outcomes of Bernoulli trials). The results are in section 3 and the derivations of these results are in section 4. 2 Notation and basics 2.1 Exclusive OR I use the symbol to denote the exclusive-or operation. So X Y =1if just one of X and Y is equal to 1; otherwise X Y =0. Y =0 Y =1 X =0 X Y =0 X Y =1 X =1 X Y =1 X Y =0 The XOR operation is commutative X Y = Y X 1

2 and associative X (Y Z) =(X Y ) Z. 2.2 Expectation, variance and covariance E(X) denotes the expected value or mean value of X, that is, the average value of a large number of repeated trials. In the case of random bits E(X) = Pr(X = 1)where Pr denotes probability. var(x) =E[{X E(X)} 2 ] denotes the variance of X. Variance is a measure of the variability of a random variable. In the case of random bits var(x) =E(X){1 E(X)}. This reaches a maximum value of 1 4 when E(X) = 1 2. cov(x, Y ) =E[{X E(X)}{Y E(Y )}] denotes the covariance of X and Y and cov(x, Y ) corr(x, Y )= p (1) var(x)var(y ) denotes the correlation of X and Y. Correlations are always between 1 and 1. When two variables are identical their correlation is equal to 1. When the two variables are statistically independent the correlation is zero. In the case of random bits cov(x, Y )=Pr(X =1,Y =1) Pr(X =1)Pr(Y =1). When the expected value, E(X), ofabit,x, iscloseto 1 2, the variance, var(x), isverycloseto 1 4 and so corr(x, Y ) is close to 4 cov(x, Y ). This sometimesprovidesconvenient approximate simplification. 2.3 Correlation and independence A sequence of random variables X 1,X 2,... are statistically independent if the following is satisfied for each of variables X i : the variables apart from X i do not provide any useful information for predicting the value of X i. In the case of pairs of random bits correlation = zero and independence are equivalent. However when more than two random bits are involved, independence implies zero correlation but not vice versa. Here is a simple example. Consider X, Y and X Y where X and Y are random bits with expected value 1 2. Each pair of X, Y and X Y are uncorrelated with each other, but X, Y and X Y are not independent since given any two of these variables one can calculate the third. 3 Main results 3.1 Independent biased pairs If X and Y are independent random bits with E(X) =µ and E(Y )=ν then E(X Y )=µ + ν 2µν = 1 2 2(µ 1 2 )(ν 1 2 ). (2) 2

3 Thus if µ and ν are close to 1 2 then E(X Y ) is very close to 1 2. For example, if µ = ν =0.6 then E(X Y )=0.48. In all cases E(X Y ) 1 2 min( µ 1 2, ν 1 2 ) and so the XOR operation always reduces bias (except when one of the variables is fixed) when the component bits are independent. Presenting the same result when µ = ν: ifx and Y are independent and E(X) =E(Y )=µ then 3.2 Independent sequences E(X Y )=2µ(1 µ) = 1 2 2(µ 1 2 )2. (3) Suppose we have n independent random bits each with expected value µ. Then the expected value of the result of XORing all of these variables is 1 2 +( 2)n 1 (µ 1 2 )n. (4) The following table shows the bias (expected value 0.5) for various values of the mean, µ, and number bits, n. µ n =2 n =3 n =4 n =5 n = Correlated pairs If E(X) =µ, E(Y )=ν and X and Y have correlation ρ then E(X Y )= 1 2 2(µ 1 2 )(ν 1 2 ) 2ρp µ(1 µ)ν(1 ν) (5) 1 2 2(µ 1 2 )(ν 1 2 ) 1 2 ρ (assuming µ, ν are near 1 2 ). Thus a small amount of correlation can add significant bias to the result. 3.4 Independent correlated pairs Suppose E(X 1 ) = E(X 2 ) = E(Y 1 ) = E(Y 2 ) = µ, the pair (X 1,X 2 ) is independent of the pair (Y 1,Y 2 ) and cov(x 1,X 2 )=cov(y 1,Y 2 )=c. For example, X 1 and X 2 might be successive observations from a random number generator and so are slightly correlated with each other. Y 1 3

4 and Y 2 are from a second identical generator that operates completely independently of the first one. Then cov(x 1 Y 1,X 2 Y 2 )=4{c 2 +2c(µ 1 2 )2 }. (6) In terms of correlation (assuming µ near 1 2 ) corr(x 1 Y 1,X 2 Y 2 ) ρ 2 +8ρ(µ 1 2 )2 (7) where ρ is the correlation. Thus, if ρ is small and µ is close to 1 2 then corr(x 1 Y 1,X 2 Y 2 ) is very small. If µ is not close to 1 2 then some of the correlation will persist. 3.5 Words Suppose we have a sequence of bits W 1,W 2,...,W n,x 1,X 2,...,X n,y 1,Y 2,...,Y n,z 1,Z 2,...,Z n from a random number generator. We are going to amalgamate then into words W, X, Y, Z, eachwithn>2 bits, and then bit-wise XOR words W and X and bit-wise XOR words Y and Z. Here, I am letting W etc. denote the sequences W 1,W 2,...,W n, etc. W 1 W 2... W n Y 1 Y 2... Y n X 1 X 2... X n Z 1 Z 2... Z n W 1 X 1 W 2 X 2... W n X n Y 1 Z 1 Y 2 Z 2... Y n Z n Suppose the correlation structure is such that only adjacent bits are significantly correlated. Then formulae (6) and (7) give us the expected values and correlation structure of the combined word except for the correlation between W 1 X 1 and W n X n ; Y 1 Z 1 and Y n Z n ; W n X n and Y 1 Z 1. Consider W 1 X 1 and W n X n (or equivalently Y 1 Z 1 and Y n Z n ). Suppose E(W 1 )=E(W n )=E(X 1 )=E(X n )=µ and W 1, (W n,x 1 ),X n are independent but W n and X 1 have covariance c and correlation ρ. Then cov(w 1 X 1,W n X n )=4c(µ 1 2 )2. (8) In terms of correlation (assuming µ near 1 2 ) corr(w 1 X 1,W n X n ) 4ρ(µ 1 2 )2. (9) This will be small under the same kind of conditions that make (7) small. Now consider W n X n and Y 1 Z 1. The correlation structure is the same as above so that corr(w n X n,y 1 Z 1 ) 4ρ(µ 1 2 )2. (10) 4

5 So arranging the bits into bytes, for example, and XORing the bytes gives a better performance than XORing adjacent bits as in section 3.3. I have not proved that there is no hidden dependence of the type found in section While I think it is unlikely that there is hidden dependence, this does need to be proved. 3.6 Flawed correctors Suppose we have a sequence of independent random bits X 1,...,X n with expected value µ 6= 1 2. Following section 3.1 we can reduce the bias by forming the sequence X 1 X 2,X 3 X 4,X 5 X 6,... Of course, we have only half as many corrected bits as raw bits. This section considers two flawed methods of attempting to produce about as many corrected bits as raw bits Accumulated randomness Let Y 1,...,Y n be defined by Y 1 = X 1 and Y i = X i Y i 1 for i>1. We are attempting to produce a corrected sequence of random bits by taking the exclusive-or of all the bits so far observed. Y n 1 and Y n,arethelasttwo bits in our corrected series. They will have expected values very close to 1 2 if n is large. However, they will have correlation corr(y n 1,Y n )= 2(µ 1 2 ) (11) so these corrected numbers have just replaced bias by correlation unless we use only every second Y i Pair-wise exclusive-or Let Y 2,...,Y n be defined by Then Y i = X i 1 X i. E(Y i )= 1 2 2(µ 1 2 )2 (12) and corr(y i,y i+1 )= 2(µ 1 2 )2 1 2µ +2µ 2 4(µ 1 2 )2 (13) (assuming near µ near 1 2 ). If i j > 1 then corr(y i,y j )=0. Apparently this method of correction has taken n raw bits and provided n 1 corrected bits with a substantial improvement in the bias with just a 5

6 minor increase in pair-wise correlation (assuming the bias of the raw bits is not too large). This is illusory. The correlation is not picking up the full extent of the dependence. In section I define a random bit, S, which canbecalculatedfromy 2,...,Y n 1 such that, for large n corr(y n,s) 2 µ 1 2 (14) so that the deviation from unbiasedness and independence is really similar to that in the original series, which, of course, is what one would expect. Of course, there is no problem if we take only every second Y i. On the other hand, suppose we used the series X 1 X 2, X 2 X 3, X 4 X 5, X 5 X 6, X 7 X 8,... That is, we are omitting every third Y i. Then the bias and correlations are of order (µ 1 2 )2 and there can be no hidden dependence. So this might be a reasonable way of increasing the yield of corrected bits. 3.7 Correlated triple Suppose E(X) = E(Y ) = E(Z) = µ, corr(x, Y ) = corr(y,z) = ρ and the process X, Y, Z is Markov so that, for example, the conditional expectation E(Z X, Y )=E(Z Y ). Then E(X Y Z) = (µ 1 2 )3 +4ρ(µ 1 2 )µ(1 µ)(2 ρ) (15) (µ 1 2 )3 +2ρ(µ 1 2 ) if µ is near 1 2 and ρ is near 0. This formula is important if we are going to correct by XORing three bits and auto-correlation is present. If ρ is very small it is not going to present a problem but if it is of a similar size to µ 1 2 then it may have a significant effect on the bias. In general, there is no reason to suppose the process of random bits is exactly Markov. However, if ρ is small, and the dependence between nonadjacent bits very small, the Markov assumption is probably close enough. 4 Derivation of formulae 4.1 Calculation trick Suppose X can take values 0 or 1. Let a(x) =1 2X. SoX = {1 a(x)}/2 and a(x) takes the values 1 and 1 corresponding to X s 0 and 1. Then a(x Y )=a(x)a(y ) where denotes the XOR operation. The usefulness of this is that we know how to do manipulate multiplication in probability calculations but doing 6

7 XOR calculations directly is awkward and unfamiliar. Also E{a(X)} =1 2E(X) var{a(x)} =4var(X) cov{a(x),a(y )} =4cov(X, Y ) corr{a(x),a(y )} =corr(x, Y ) so we can transform expectations and variances between X and a(x). Also note that a(x) 2 = The derivations Formula (2) Suppose X and Y are independent, E(X) =µ and E(Y )=ν then E(X Y )= E{a(X Y )} = E{a(X)a(Y )} = E{a(X)}E{a(Y )} = (1 2µ)(1 2ν) = 1 2 2(µ 1 2 )(ν 1 2 ) Formula (5) Suppose X and Y have covariance c and E(X) =µ and E(Y )=ν. E(X Y )= E{a(X Y )} = E{a(X)a(Y )} = [E{a(X)}E{a(Y )} +4c] = (1 2µ)(1 2ν) 2c = 1 2 2(µ 1 2 )(ν 1 2 ) 2c. 7

8 4.2.3 Formula (6) Suppose E(X 1 ) = E(X 2 ) = E(Y 1 ) = E(Y 2 ) = µ, the pair (X 1,X 2 ) is independent of the pair (Y 1,Y 2 ) and cov(x 1,X 2 )=cov(y 1,Y 2 )=c then cov(x 1 Y 1,X 2 Y 2 ) = 1 4 cov{a(x 1 Y 1 ),a(x 2 Y 2 )} = 1 4 cov{a(x 1)a(Y 1 ),a(x 2 )a(y 2 )} = 1 4 [E{a(X 1)a(Y 1 )a(x 2 )a(y 2 )} E{a(X 1 )a(y 1 )}E{a(X 2 )a(y 2 )}] = 1 4 [E{a(X 1)a(X 2 )}E{a(Y 1 )a(y 2 )} E{a(X 1 )}E{a(Y 1 )}E{a(X 2 )}E{a(Y 2 )}] = 1 4 [{4c +(1 2µ)2 } 2 (1 2µ) 4 ] =4{c 2 +2c(µ 1 2 )2 } Formula (8) Suppose E(W 1 )=E(W n )=E(X 1 )=E(X n )=µ and W 1, (W n,x 1 ),X n are independent but W n and X 1 have covariance c. Then cov(w 1 X 1,W n X n ) = 1 4 cov{a(w 1 X 1 ),a(w n X n )} = 1 4 cov{a(w 1)a(X 1 ),a(w n )a(x n )} = 1 4 [E{a(W 1)a(X 1 )a(w n )a(x n )} E{a(W 1 )a(x 1 )}E{a(W n )a(x n )}] = 1 4 [E{a(W 1)}E{a(W n )a(x 1 )}E{a(X n )} E{a(W 1 )}E{a(X 1 )}E{a(W n )}E{a(X n )}] = 1 4 (1 2µ)2 [{4c +(1 2µ) 2 } (1 2µ) 2 ] =4c(µ 1 2 ) Formula (11) Suppose X n and Y n 1 are independent, E(X n )=µ and E(Y n 1 )= 1 2.Then corr(y n 1,X n Y n 1 ) =corr{a(y n 1 ),a(x n Y n 1 )} =cov{a(y n 1 ),a(x n Y n 1 )} =cov{a(y n 1 ),a(x n )a(y n 1 )} = E{a(Y n 1 )a(x n )a(y n 1 )} E{a(Y n 1 )}E{a(X n )}E{a(Y n 1 )} = E{a(X n )} =1 2µ. 8

9 4.2.6 Formula (13) Suppose X 1,X 2,X 3 are independent with expected value µ. Then cov(x 1 X 2,X 2 X 3 ) = 1 4 cov{a(x 1)a(X 2 ),a(x 2 )a(x 3 )} = 1 4 [E{a(X 1)a(X 2 )a(x 2 )a(x 3 )} E{a(X 1 )a(x 2 )}E{a(X 2 )a(x 3 )}] = 1 4 {(1 2µ)2 (1 2µ) 4 } =4(µ 1 2 )2 µ(1 µ). Divide by the variance of X 1 X 2 to get formula (13) Formula (14) Suppose X 1,...,X n are a sequence of independent random bits with expected value µ. LetY 2,...,Y n be defined by Let and define Y i = X i 1 X i. Z n 1 = Y n 1 = X n 1 X n 2, (16) Z n 2 = Z n 1 Y n 2 = X n 1 X n 3, Z n 3 = Z n 2 Y n 3 = X n 1 X n 4,..., Z 2 = Z 3 Y 2 = X n 1 X 1 S =1if n 1 1 X Z i > 1 n 2 2 i=2 and S =0otherwise. S can be calculated from Y 2,...,Y n. From formula (16) and If n is large n 1 1 X Z i = 1 n 2 X X i if X n 1 =0 n 2 n 2 i=2 i=1 n 1 1 X Z i =1 1 n 2 X X i if X n 1 =1. n 2 n 2 i=2 i=1 n 2 1 X X i > 1 n 2 2 with a high probability if µ> 1 2 i=1 9

10 and n 2 1 X X i < 1 n 2 2 with a high probability if µ< 1 2. i=1 Putting all this together, if n is large, we have (with a high probability) and S =1if X n 1 =0and µ> 1 2 or if X n 1 =1and µ< 1 2 S =0if X n 1 =1and µ> 1 2 or if X n 1 =0and µ< 1 2. Hence (assuming µ is close to but not exactly 1 2 ) since corr(y n,s) corr(x n X n 1,X n 1 ) sign(µ 1 2 ) 2 µ 1 2 corr(x n X n 1,X n 1 ) cov{a(x n )a(x n 1 ),a(x n 1 )} = E{a(X n )a(x n 1 )a(x n 1 )} E{a(X n )a(x n 1 )}E{a(X n 1 )} =(1 2µ) (1 2µ) 3 2(µ 1 2 ) Formula (15) Suppose E(X) = E(Y ) = E(Z) = µ, corr(x, Y ) = corr(y,z) = ρ and the process is Markov so that the conditional expectation E(Z X, Y )=E(Z Y ). Now and E(Z Y =1)=Pr(Z =1,Y =1)/ Pr(Y =1) = E(YZ)/µ = {ρµ(1 µ)+µ 2 }/µ = ρ ρµ + µ E(Z Y =0)=Pr(Z =1,Y =0)/ Pr(Y =0) = {Pr(Z =1) Pr(Z =1,Y =1)}/ Pr(Y =0) = {µ E(YZ)}/(1 µ) = {µ ρµ(1 µ) µ 2 }/(1 µ) = ρµ + µ so that E(Z Y )=ρy ρµ + µ. Hence E{a(Z) Y } =1 2(ρY ρµ + µ) =ρa(y )+(1 2µ)(1 ρ). (17) 10

11 Then E(X Y Z) = E{a(X Y Z)} = E{a(X)a(Y )a(z)} = E[a(X)a(Y )E{a(Z) X, Y }] = E[a(X)a(Y )E{a(Z) Y }] = E[a(X)a(Y ){ρa(y )+(1 2µ)(1 ρ)}] = E{ρa(X)+a(X)a(Y )(1 2µ)(1 ρ)} = (1 2µ)[ρ + {(1 2µ)2 +4ρµ(1 µ)}(1 ρ)] = (1 ρ)(1 2µ)3 1 2ρ(1 2µ){1+4µ(1 µ)(1 ρ)} = (1 ρ)(µ 1 2 )3 + ρ(µ 1 2 ){1+4µ(1 µ)(1 ρ)} = (µ 1 2 )3 + ρ(µ 1 2 ){1+4µ(1 µ)(1 ρ) 4µ2 +4µ 1} = (µ 1 2 )3 +4ρ(µ 1 2 )µ(1 µ)(2 ρ). 11

### Lecture 16: Expected value, variance, independence and Chebyshev inequality

Lecture 16: Expected value, variance, independence and Chebyshev inequality Expected value, variance, and Chebyshev inequality. If X is a random variable recall that the expected value of X, E[X] is the

### Joint Distribution and Correlation

Joint Distribution and Correlation Michael Ash Lecture 3 Reminder: Start working on the Problem Set Mean and Variance of Linear Functions of an R.V. Linear Function of an R.V. Y = a + bx What are the properties

### Probability & Statistics Primer Gregory J. Hakim University of Washington 2 January 2009 v2.0

Probability & Statistics Primer Gregory J. Hakim University of Washington 2 January 2009 v2.0 This primer provides an overview of basic concepts and definitions in probability and statistics. We shall

### Chapters 5. Multivariate Probability Distributions

Chapters 5. Multivariate Probability Distributions Random vectors are collection of random variables defined on the same sample space. Whenever a collection of random variables are mentioned, they are

### Expectation Discrete RV - weighted average Continuous RV - use integral to take the weighted average

PHP 2510 Expectation, variance, covariance, correlation Expectation Discrete RV - weighted average Continuous RV - use integral to take the weighted average Variance Variance is the average of (X µ) 2

### Covariance and Correlation

Covariance and Correlation ( c Robert J. Serfling Not for reproduction or distribution) We have seen how to summarize a data-based relative frequency distribution by measures of location and spread, such

### Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)

Summary of Formulas and Concepts Descriptive Statistics (Ch. 1-4) Definitions Population: The complete set of numerical information on a particular quantity in which an investigator is interested. We assume

### Joint Probability Distributions and Random Samples. Week 5, 2011 Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage

5 Joint Probability Distributions and Random Samples Week 5, 2011 Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage Two Discrete Random Variables The probability mass function (pmf) of a single

### Calculate the holding period return for this investment. It is approximately

1. An investor purchases 100 shares of XYZ at the beginning of the year for \$35. The stock pays a cash dividend of \$3 per share. The price of the stock at the time of the dividend is \$30. The dividend

### We have discussed the notion of probabilistic dependence above and indicated that dependence is

1 CHAPTER 7 Online Supplement Covariance and Correlation for Measuring Dependence We have discussed the notion of probabilistic dependence above and indicated that dependence is defined in terms of conditional

### P (x) 0. Discrete random variables Expected value. The expected value, mean or average of a random variable x is: xp (x) = v i P (v i )

Discrete random variables Probability mass function Given a discrete random variable X taking values in X = {v 1,..., v m }, its probability mass function P : X [0, 1] is defined as: P (v i ) = Pr[X =

### Econometrics Regression Analysis with Time Series Data

Econometrics Regression Analysis with Time Series Data João Valle e Azevedo Faculdade de Economia Universidade Nova de Lisboa Spring Semester João Valle e Azevedo (FEUNL) Econometrics Lisbon, May 2011

### Solution to HW - 1. Problem 1. [Points = 3] In September, Chapel Hill s daily high temperature has a mean

Problem 1. [Points = 3] In September, Chapel Hill s daily high temperature has a mean of 81 degree F and a standard deviation of 10 degree F. What is the mean, standard deviation and variance in terms

### ST 371 (VIII): Theory of Joint Distributions

ST 371 (VIII): Theory of Joint Distributions So far we have focused on probability distributions for single random variables. However, we are often interested in probability statements concerning two or

### 15.062 Data Mining: Algorithms and Applications Matrix Math Review

.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop

### Exercises with solutions (1)

Exercises with solutions (). Investigate the relationship between independence and correlation. (a) Two random variables X and Y are said to be correlated if and only if their covariance C XY is not equal

### The Scalar Algebra of Means, Covariances, and Correlations

3 The Scalar Algebra of Means, Covariances, and Correlations In this chapter, we review the definitions of some key statistical concepts: means, covariances, and correlations. We show how the means, variances,

### Variances and covariances

Chapter 4 Variances and covariances 4.1 Overview The expected value of a random variable gives a crude measure for the center of location of the distribution of that random variable. For instance, if the

### Random Variables, Expectation, Distributions

Random Variables, Expectation, Distributions CS 5960/6960: Nonparametric Methods Tom Fletcher January 21, 2009 Review Random Variables Definition A random variable is a function defined on a probability

### Notes for STA 437/1005 Methods for Multivariate Data

Notes for STA 437/1005 Methods for Multivariate Data Radford M. Neal, 26 November 2010 Random Vectors Notation: Let X be a random vector with p elements, so that X = [X 1,..., X p ], where denotes transpose.

### Sections 2.11 and 5.8

Sections 211 and 58 Timothy Hanson Department of Statistics, University of South Carolina Stat 704: Data Analysis I 1/25 Gesell data Let X be the age in in months a child speaks his/her first word and

### Definition The covariance of X and Y, denoted by cov(x, Y ) is defined by. cov(x, Y ) = E(X µ 1 )(Y µ 2 ).

Correlation Regression Bivariate Normal Suppose that X and Y are r.v. s with joint density f(x y) and suppose that the means of X and Y are respectively µ 1 µ 2 and the variances are 1 2. Definition The

### ECE302 Spring 2006 HW7 Solutions March 11, 2006 1

ECE32 Spring 26 HW7 Solutions March, 26 Solutions to HW7 Note: Most of these solutions were generated by R. D. Yates and D. J. Goodman, the authors of our textbook. I have added comments in italics where

### Worked examples Multiple Random Variables

Worked eamples Multiple Random Variables Eample Let X and Y be random variables that take on values from the set,, } (a) Find a joint probability mass assignment for which X and Y are independent, and

### Chapter 4 Lecture Notes

Chapter 4 Lecture Notes Random Variables October 27, 2015 1 Section 4.1 Random Variables A random variable is typically a real-valued function defined on the sample space of some experiment. For instance,

### 1 OLS under Measurement Error

ECON 370: Measurement Error 1 Violations of Gauss Markov Assumptions: Measurement Error Econometric Methods, ECON 370 1 OLS under Measurement Error We have found out that another source of endogeneity

### Math 370/408, Spring 2008 Prof. A.J. Hildebrand. Actuarial Exam Practice Problem Set 2 Solutions

Math 70/408, Spring 2008 Prof. A.J. Hildebrand Actuarial Exam Practice Problem Set 2 Solutions About this problem set: These are problems from Course /P actuarial exams that I have collected over the years,

### Correlation in Random Variables

Correlation in Random Variables Lecture 11 Spring 2002 Correlation in Random Variables Suppose that an experiment produces two random variables, X and Y. What can we say about the relationship between

### Statistics 100A Homework 4 Solutions

Problem 1 For a discrete random variable X, Statistics 100A Homework 4 Solutions Ryan Rosario Note that all of the problems below as you to prove the statement. We are proving the properties of epectation

### Queuing Formulas. 1 Notation 2. 2 Basic Queueing Formulas 2

Queuing Formulas Contents 1 Notation 2 2 Basic Queueing Formulas 2 3 Single-Server Queues 3 3.1 Formulas...................................... 3 3.2 Some additional useful facts...........................

### Notes 11 Autumn 2005

MAS 08 Probabilit I Notes Autumn 005 Two discrete random variables If X and Y are discrete random variables defined on the same sample space, then events such as X = and Y = are well defined. The joint

### Homework Zero. Max H. Farrell Chicago Booth BUS41100 Applied Regression Analysis. Complete before the first class, do not turn in

Homework Zero Max H. Farrell Chicago Booth BUS41100 Applied Regression Analysis Complete before the first class, do not turn in This homework is intended as a self test of your knowledge of the basic statistical

### Math 151. Rumbos Spring 2014 1. Solutions to Assignment #22

Math 151. Rumbos Spring 2014 1 Solutions to Assignment #22 1. An experiment consists of rolling a die 81 times and computing the average of the numbers on the top face of the die. Estimate the probability

### Topic 8 The Expected Value

Topic 8 The Expected Value Functions of Random Variables 1 / 12 Outline Names for Eg(X ) Variance and Standard Deviation Independence Covariance and Correlation 2 / 12 Names for Eg(X ) If g(x) = x, then

### The Bivariate Normal Distribution

The Bivariate Normal Distribution This is Section 4.7 of the st edition (2002) of the book Introduction to Probability, by D. P. Bertsekas and J. N. Tsitsiklis. The material in this section was not included

### Joint Exam 1/P Sample Exam 1

Joint Exam 1/P Sample Exam 1 Take this practice exam under strict exam conditions: Set a timer for 3 hours; Do not stop the timer for restroom breaks; Do not look at your notes. If you believe a question

### Lesson 5 Chapter 4: Jointly Distributed Random Variables

Lesson 5 Chapter 4: Jointly Distributed Random Variables Department of Statistics The Pennsylvania State University 1 Marginal and Conditional Probability Mass Functions The Regression Function Independence

### Stat 704 Data Analysis I Probability Review

1 / 30 Stat 704 Data Analysis I Probability Review Timothy Hanson Department of Statistics, University of South Carolina Course information 2 / 30 Logistics: Tuesday/Thursday 11:40am to 12:55pm in LeConte

### Worked examples Random Processes

Worked examples Random Processes Example 1 Consider patients coming to a doctor s office at random points in time. Let X n denote the time (in hrs) that the n th patient has to wait before being admitted

### 4. Joint Distributions of Two Random Variables

4. Joint Distributions of Two Random Variables 4.1 Joint Distributions of Two Discrete Random Variables Suppose the discrete random variables X and Y have supports S X and S Y, respectively. The joint

### Limitations of regression analysis

Limitations of regression analysis Ragnar Nymoen Department of Economics, UiO 8 February 2009 Overview What are the limitations to regression? Simultaneous equations bias Measurement errors in explanatory

### University of California, Los Angeles Department of Statistics. Random variables

University of California, Los Angeles Department of Statistics Statistics Instructor: Nicolas Christou Random variables Discrete random variables. Continuous random variables. Discrete random variables.

### Topic 4: Multivariate random variables. Multiple random variables

Topic 4: Multivariate random variables Joint, marginal, and conditional pmf Joint, marginal, and conditional pdf and cdf Independence Expectation, covariance, correlation Conditional expectation Two jointly

### 1 Short Introduction to Time Series

ECONOMICS 7344, Spring 202 Bent E. Sørensen January 24, 202 Short Introduction to Time Series A time series is a collection of stochastic variables x,.., x t,.., x T indexed by an integer value t. The

### Math 431 An Introduction to Probability. Final Exam Solutions

Math 43 An Introduction to Probability Final Eam Solutions. A continuous random variable X has cdf a for 0, F () = for 0 <

### Discrete Math in Computer Science Homework 7 Solutions (Max Points: 80)

Discrete Math in Computer Science Homework 7 Solutions (Max Points: 80) CS 30, Winter 2016 by Prasad Jayanti 1. (10 points) Here is the famous Monty Hall Puzzle. Suppose you are on a game show, and you

### 4.3 Moving Average Process MA(q)

66 CHAPTER 4. STATIONARY TS MODELS 4.3 Moving Average Process MA(q) Definition 4.5. {X t } is a moving-average process of order q if X t = Z t + θ 1 Z t 1 +... + θ q Z t q, (4.9) where and θ 1,...,θ q

### Aggregate Loss Models

Aggregate Loss Models Chapter 9 Stat 477 - Loss Models Chapter 9 (Stat 477) Aggregate Loss Models Brian Hartman - BYU 1 / 22 Objectives Objectives Individual risk model Collective risk model Computing

### Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written

### Math 141. Lecture 7: Variance, Covariance, and Sums. Albyn Jones 1. 1 Library 304. jones/courses/141

Math 141 Lecture 7: Variance, Covariance, and Sums Albyn Jones 1 1 Library 304 jones@reed.edu www.people.reed.edu/ jones/courses/141 Last Time Variance: expected squared deviation from the mean: Standard

### Lecture 7: Approximation via Randomized Rounding

Lecture 7: Approximation via Randomized Rounding Often LPs return a fractional solution where the solution x, which is supposed to be in {0, } n, is in [0, ] n instead. There is a generic way of obtaining

### Random variables P(X = 3) = P(X = 3) = 1 8, P(X = 1) = P(X = 1) = 3 8.

Random variables Remark on Notations 1. When X is a number chosen uniformly from a data set, What I call P(X = k) is called Freq[k, X] in the courseware. 2. When X is a random variable, what I call F ()

### BOOLEAN ALGEBRA & LOGIC GATES

BOOLEAN ALGEBRA & LOGIC GATES Logic gates are electronic circuits that can be used to implement the most elementary logic expressions, also known as Boolean expressions. The logic gate is the most basic

### Random Vectors and the Variance Covariance Matrix

Random Vectors and the Variance Covariance Matrix Definition 1. A random vector X is a vector (X 1, X 2,..., X p ) of jointly distributed random variables. As is customary in linear algebra, we will write

### On Correlating Performance Metrics

On Correlating Performance Metrics Yiping Ding and Chris Thornley BMC Software, Inc. Kenneth Newman BMC Software, Inc. University of Massachusetts, Boston Performance metrics and their measurements are

### Solutions for the exam for Matematisk statistik och diskret matematik (MVE050/MSG810). Statistik för fysiker (MSG820). December 15, 2012.

Solutions for the exam for Matematisk statistik och diskret matematik (MVE050/MSG810). Statistik för fysiker (MSG8). December 15, 12. 1. (3p) The joint distribution of the discrete random variables X and

### Joint Probability Distributions and Random Samples (Devore Chapter Five)

Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01 Probability and Statistics for Engineers Winter 2010-2011 Contents 1 Joint Probability Distributions 1 1.1 Two Discrete

### For a partition B 1,..., B n, where B i B j = for i. A = (A B 1 ) (A B 2 ),..., (A B n ) and thus. P (A) = P (A B i ) = P (A B i )P (B i )

Probability Review 15.075 Cynthia Rudin A probability space, defined by Kolmogorov (1903-1987) consists of: A set of outcomes S, e.g., for the roll of a die, S = {1, 2, 3, 4, 5, 6}, 1 1 2 1 6 for the roll

### Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS

Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS Part 4: Geometric Distribution Negative Binomial Distribution Hypergeometric Distribution Sections 3-7, 3-8 The remaining discrete random

### Practice Exam 1: Long List 18.05, Spring 2014

Practice Eam : Long List 8.05, Spring 204 Counting and Probability. A full house in poker is a hand where three cards share one rank and two cards share another rank. How many ways are there to get a full-house?

### The sample space for a pair of die rolls is the set. The sample space for a random number between 0 and 1 is the interval [0, 1].

Probability Theory Probability Spaces and Events Consider a random experiment with several possible outcomes. For example, we might roll a pair of dice, flip a coin three times, or choose a random real

### Lecture 14: Correlation and Autocorrelation Steven Skiena. skiena

Lecture 14: Correlation and Autocorrelation Steven Skiena Department of Computer Science State University of New York Stony Brook, NY 11794 4400 http://www.cs.sunysb.edu/ skiena Overuse of Color, Dimensionality,

### Homework 4 - KEY. Jeff Brenion. June 16, 2004. Note: Many problems can be solved in more than one way; we present only a single solution here.

Homework 4 - KEY Jeff Brenion June 16, 2004 Note: Many problems can be solved in more than one way; we present only a single solution here. 1 Problem 2-1 Since there can be anywhere from 0 to 4 aces, the

### MULTIVARIATE PROBABILITY DISTRIBUTIONS

MULTIVARIATE PROBABILITY DISTRIBUTIONS. PRELIMINARIES.. Example. Consider an experiment that consists of tossing a die and a coin at the same time. We can consider a number of random variables defined

### Unit 4 The Bernoulli and Binomial Distributions

PubHlth 540 4. Bernoulli and Binomial Page 1 of 19 Unit 4 The Bernoulli and Binomial Distributions Topic 1. Review What is a Discrete Probability Distribution... 2. Statistical Expectation.. 3. The Population

### Math/Stats 342: Solutions to Homework

Math/Stats 342: Solutions to Homework Steven Miller (sjm1@williams.edu) November 17, 2011 Abstract Below are solutions / sketches of solutions to the homework problems from Math/Stats 342: Probability

### 1 Portfolio mean and variance

Copyright c 2005 by Karl Sigman Portfolio mean and variance Here we study the performance of a one-period investment X 0 > 0 (dollars) shared among several different assets. Our criterion for measuring

### Advanced Fixed Income Analytics Lecture 1

Advanced Fixed Income Analytics Lecture 1 Backus & Zin/April 1, 1999 Vasicek: The Fixed Income Benchmark 1. Prospectus 2. Models and their uses 3. Spot rates and their properties 4. Fundamental theorem

### An axiomatic approach to capital allocation

An axiomatic approach to capital allocation Michael Kalkbrener Deutsche Bank AG Abstract Capital allocation techniques are of central importance in portfolio management and risk-based performance measurement.

### Example. A casino offers the following bets (the fairest bets in the casino!) 1 You get \$0 (i.e., you can walk away)

: Three bets Math 45 Introduction to Probability Lecture 5 Kenneth Harris aharri@umich.edu Department of Mathematics University of Michigan February, 009. A casino offers the following bets (the fairest

### Chapter 5. Random variables

Random variables random variable numerical variable whose value is the outcome of some probabilistic experiment; we use uppercase letters, like X, to denote such a variable and lowercase letters, like

### Mathematical Expectation

Mathematical Expectation Properties of Mathematical Expectation I The concept of mathematical expectation arose in connection with games of chance. In its simplest form, mathematical expectation is the

### Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 13. Random Variables: Distribution and Expectation

CS 70 Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 3 Random Variables: Distribution and Expectation Random Variables Question: The homeworks of 20 students are collected

### Autocovariance and Autocorrelation

Chapter 3 Autocovariance and Autocorrelation If the {X n } process is weakly stationary, the covariance of X n and X n+k depends only on the lag k. This leads to the following definition of the autocovariance

### Sampling Central Limit Theorem Proportions. Outline. 1 Sampling. 2 Central Limit Theorem. 3 Proportions

Outline 1 Sampling 2 Central Limit Theorem 3 Proportions Outline 1 Sampling 2 Central Limit Theorem 3 Proportions Populations and samples When we use statistics, we are trying to find out information about

### 5.3 The Cross Product in R 3

53 The Cross Product in R 3 Definition 531 Let u = [u 1, u 2, u 3 ] and v = [v 1, v 2, v 3 ] Then the vector given by [u 2 v 3 u 3 v 2, u 3 v 1 u 1 v 3, u 1 v 2 u 2 v 1 ] is called the cross product (or

### Covariance and Correlation. Consider the joint probability distribution f XY (x, y).

Chapter 5: JOINT PROBABILITY DISTRIBUTIONS Part 2: Section 5-2 Covariance and Correlation Consider the joint probability distribution f XY (x, y). Is there a relationship between X and Y? If so, what kind?

### 3 Multiple Discrete Random Variables

3 Multiple Discrete Random Variables 3.1 Joint densities Suppose we have a probability space (Ω, F,P) and now we have two discrete random variables X and Y on it. They have probability mass functions f

### Continuous random variables

Continuous random variables So far we have been concentrating on discrete random variables, whose distributions are not continuous. Now we deal with the so-called continuous random variables. A random

### Econ 132 C. Health Insurance: U.S., Risk Pooling, Risk Aversion, Moral Hazard, Rand Study 7

Econ 132 C. Health Insurance: U.S., Risk Pooling, Risk Aversion, Moral Hazard, Rand Study 7 C2. Health Insurance: Risk Pooling Health insurance works by pooling individuals together to reduce the variability

### WHERE DOES THE 10% CONDITION COME FROM?

1 WHERE DOES THE 10% CONDITION COME FROM? The text has mentioned The 10% Condition (at least) twice so far: p. 407 Bernoulli trials must be independent. If that assumption is violated, it is still okay

### 5. Linear Regression

5. Linear Regression Outline.................................................................... 2 Simple linear regression 3 Linear model............................................................. 4

### 2. Discrete random variables

2. Discrete random variables Statistics and probability: 2-1 If the chance outcome of the experiment is a number, it is called a random variable. Discrete random variable: the possible outcomes can be

### Conditional expectation

A Conditional expectation A.1 Review of conditional densities, expectations We start with the continuous case. This is sections 6.6 and 6.8 in the book. Let X, Y be continuous random variables. We defined

### CS229 Lecture notes. Andrew Ng

CS229 Lecture notes Andrew Ng Part X Factor analysis Whenwehavedatax (i) R n thatcomesfromamixtureofseveral Gaussians, the EM algorithm can be applied to fit a mixture model. In this setting, we usually

### Math 461 Fall 2006 Test 2 Solutions

Math 461 Fall 2006 Test 2 Solutions Total points: 100. Do all questions. Explain all answers. No notes, books, or electronic devices. 1. [105+5 points] Assume X Exponential(λ). Justify the following two

### Time Series Analysis

Time Series Analysis Time series and stochastic processes Andrés M. Alonso Carolina García-Martos Universidad Carlos III de Madrid Universidad Politécnica de Madrid June July, 2012 Alonso and García-Martos

### RANDOM VARIABLES MATH CIRCLE (ADVANCED) 3/3/2013. 3 k) ( 52 3 )

RANDOM VARIABLES MATH CIRCLE (ADVANCED) //0 0) a) Suppose you flip a fair coin times. i) What is the probability you get 0 heads? ii) head? iii) heads? iv) heads? For = 0,,,, P ( Heads) = ( ) b) Suppose

### Department of Mathematics, Indian Institute of Technology, Kharagpur Assignment 2-3, Probability and Statistics, March 2015. Due:-March 25, 2015.

Department of Mathematics, Indian Institute of Technology, Kharagpur Assignment -3, Probability and Statistics, March 05. Due:-March 5, 05.. Show that the function 0 for x < x+ F (x) = 4 for x < for x

### Lecture 5: Mathematical Expectation

Lecture 5: Mathematical Expectation Assist. Prof. Dr. Emel YAVUZ DUMAN MCB1007 Introduction to Probability and Statistics İstanbul Kültür University Outline 1 Introduction 2 The Expected Value of a Random

### Math 370/408, Spring 2008 Prof. A.J. Hildebrand. Actuarial Exam Practice Problem Set 5 Solutions

Math 370/408, Spring 2008 Prof. A.J. Hildebrand Actuarial Exam Practice Problem Set 5 Solutions About this problem set: These are problems from Course 1/P actuarial exams that I have collected over the

### ECON1003: Analysis of Economic Data Fall 2003 Answers to Quiz #2 11:40a.m. 12:25p.m. (45 minutes) Tuesday, October 28, 2003

ECON1003: Analysis of Economic Data Fall 2003 Answers to Quiz #2 11:40a.m. 12:25p.m. (45 minutes) Tuesday, October 28, 2003 1. (4 points) The number of claims for missing baggage for a well-known airline

### Definition: Suppose that two random variables, either continuous or discrete, X and Y have joint density

HW MATH 461/561 Lecture Notes 15 1 Definition: Suppose that two random variables, either continuous or discrete, X and Y have joint density and marginal densities f(x, y), (x, y) Λ X,Y f X (x), x Λ X,

### Chapter 6: Point Estimation. Fall 2011. - Probability & Statistics

STAT355 Chapter 6: Point Estimation Fall 2011 Chapter Fall 2011 6: Point1 Estimat / 18 Chap 6 - Point Estimation 1 6.1 Some general Concepts of Point Estimation Point Estimate Unbiasedness Principle of

### 3.1 Least squares in matrix form

118 3 Multiple Regression 3.1 Least squares in matrix form E Uses Appendix A.2 A.4, A.6, A.7. 3.1.1 Introduction More than one explanatory variable In the foregoing chapter we considered the simple regression

### Section 5.1 Continuous Random Variables: Introduction

Section 5. Continuous Random Variables: Introduction Not all random variables are discrete. For example:. Waiting times for anything (train, arrival of customer, production of mrna molecule from gene,

### Power & Effect Size power Effect Size

Power & Effect Size Until recently, researchers were primarily concerned with controlling Type I errors (i.e. finding a difference when one does not truly exist). Although it is important to make sure

### DISCRETE RANDOM VARIABLES

DISCRETE RANDOM VARIABLES DISCRETE RANDOM VARIABLES Documents prepared for use in course B01.1305, New York University, Stern School of Business Definitions page 3 Discrete random variables are introduced