Statistics 100A Homework 8 Solutions

Similar documents
Statistics 100A Homework 7 Solutions

Joint Exam 1/P Sample Exam 1

Chapter 4 Lecture Notes

Statistics 100A Homework 4 Solutions

For a partition B 1,..., B n, where B i B j = for i. A = (A B 1 ) (A B 2 ),..., (A B n ) and thus. P (A) = P (A B i ) = P (A B i )P (B i )

Statistics 100A Homework 3 Solutions

UNIT I: RANDOM VARIABLES PART- A -TWO MARKS

Department of Mathematics, Indian Institute of Technology, Kharagpur Assignment 2-3, Probability and Statistics, March Due:-March 25, 2015.

Lecture 6: Discrete & Continuous Probability and Random Variables

( ) = 1 x. ! 2x = 2. The region where that joint density is positive is indicated with dotted lines in the graph below. y = x

Final Mathematics 5010, Section 1, Fall 2004 Instructor: D.A. Levin

Practice problems for Homework 11 - Point Estimation

Math 461 Fall 2006 Test 2 Solutions

Probability Calculator

e.g. arrival of a customer to a service station or breakdown of a component in some system.

Math/Stats 342: Solutions to Homework

Probability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur

Section 5.1 Continuous Random Variables: Introduction

Lecture Notes 1. Brief Review of Basic Probability

Errata and updates for ASM Exam C/Exam 4 Manual (Sixteenth Edition) sorted by page

ISyE 6761 Fall 2012 Homework #2 Solutions

Feb 7 Homework Solutions Math 151, Winter Chapter 4 Problems (pages )

5. Continuous Random Variables

( ) is proportional to ( 10 + x)!2. Calculate the

Mathematical Expectation

The Exponential Distribution

Aggregate Loss Models

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 13. Random Variables: Distribution and Expectation

Math 425 (Fall 08) Solutions Midterm 2 November 6, 2008

Question: What is the probability that a five-card poker hand contains a flush, that is, five cards of the same suit?

6.041/6.431 Spring 2008 Quiz 2 Wednesday, April 16, 7:30-9:30 PM. SOLUTIONS

Statistics 100A Homework 2 Solutions

MATH4427 Notebook 2 Spring MATH4427 Notebook Definitions and Examples Performance Measures for Estimators...

SF2940: Probability theory Lecture 8: Multivariate Normal Distribution

INSURANCE RISK THEORY (Problems)

Sums of Independent Random Variables

RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS

Math 370, Spring 2008 Prof. A.J. Hildebrand. Practice Test 2 Solutions

Notes on Continuous Random Variables

Binomial random variables

Recursive Estimation

Math 431 An Introduction to Probability. Final Exam Solutions

Math 55: Discrete Mathematics

WHERE DOES THE 10% CONDITION COME FROM?

Lecture 8. Confidence intervals and the central limit theorem

Maximum Likelihood Estimation

1 Prior Probability and Posterior Probability

Homework 4 - KEY. Jeff Brenion. June 16, Note: Many problems can be solved in more than one way; we present only a single solution here.

E3: PROBABILITY AND STATISTICS lecture notes

Principle of Data Reduction

Section 5-3 Binomial Probability Distributions

What is Statistics? Lecture 1. Introduction and probability review. Idea of parametric inference

Master s Theory Exam Spring 2006

The problems of first TA class of Math144

Definition: Suppose that two random variables, either continuous or discrete, X and Y have joint density

Probability density function : An arbitrary continuous random variable X is similarly described by its probability density function f x = f X

MULTIVARIATE PROBABILITY DISTRIBUTIONS

Lecture 13: Martingales

Section 6.1 Joint Distribution Functions

Second Order Linear Nonhomogeneous Differential Equations; Method of Undetermined Coefficients. y + p(t) y + q(t) y = g(t), g(t) 0.

Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab

Statistics 100A Homework 4 Solutions

Binomial random variables (Review)

2. Discrete random variables

Generating Functions

The sample space for a pair of die rolls is the set. The sample space for a random number between 0 and 1 is the interval [0, 1].

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 18. A Brief Introduction to Continuous Probability

Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS. Part 3: Discrete Uniform Distribution Binomial Distribution

Problem sets for BUEC 333 Part 1: Probability and Statistics

1 Sufficient statistics

AP Statistics 7!3! 6!

SF2940: Probability theory Lecture 8: Multivariate Normal Distribution

Exponential Distribution

Hypothesis Testing for Beginners

6 PROBABILITY GENERATING FUNCTIONS

Chapter 3 RANDOM VARIATE GENERATION

STAT 830 Convergence in Distribution

2 Binomial, Poisson, Normal Distribution

CHAPTER 6: Continuous Uniform Distribution: 6.1. Definition: The density function of the continuous random variable X on the interval [A, B] is.

Section 5 Part 2. Probability Distributions for Discrete Random Variables

WORKED EXAMPLES 1 TOTAL PROBABILITY AND BAYES THEOREM

STAT 35A HW2 Solutions

Data Mining: Algorithms and Applications Matrix Math Review

Discrete Math in Computer Science Homework 7 Solutions (Max Points: 80)

Nonparametric adaptive age replacement with a one-cycle criterion

Part 2: One-parameter models

Chapter 5: Normal Probability Distributions - Solutions

Econometrics Simple Linear Regression

STA 256: Statistics and Probability I

ECE302 Spring 2006 HW4 Solutions February 6,

Feb 28 Homework Solutions Math 151, Winter Chapter 6 Problems (pages )

A Uniform Asymptotic Estimate for Discounted Aggregate Claims with Subexponential Tails

The Bivariate Normal Distribution

You flip a fair coin four times, what is the probability that you obtain three heads.

Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)

University of California, Los Angeles Department of Statistics. Random variables

Poisson processes (and mixture distributions)

Exact Confidence Intervals

STAT 315: HOW TO CHOOSE A DISTRIBUTION FOR A RANDOM VARIABLE

Transcription:

Part : Chapter 7 Statistics A Homework 8 Solutions Ryan Rosario. A player throws a fair die and simultaneously flips a fair coin. If the coin lands heads, then she wins twice, and if tails, the one-half of the value that appears on the die. Explain her expected winnings. First, this problem is poorly worded. If the coin lands heads, then she wins twice the value on the die, otherwise she wins the value that appears on the die. Let X represent the number of pips that show on the die. Let Y be the result of the toss of the fair coin, such that Y if the coin lands heads, and Y otherwise. Then, { X, Y E(XY ), Y Each value of Y occurs with probability. So, E(XY ) (X) + ( 4.375 ( ) X [ 6 + 6 +... + 6 6 6 ]) + ( [ 6 + 6 +... + 6 6 ]) 6 5. The county hospital is located at the center of a square whose sides are 3 miles wide. If an accident occurs within this square, then the hospital sends out an ambulance. The road network is rectangular so the travel distance from the hospital, whose coordinates are (,), to the point (x, y) is x + y. If an accident occurs at a point that is uniformly distributed in the square, find the expected travel distance of the ambulance. First, let Z X + Y for simplicity. We want to find E(Z). Note that 3 < X < 3 and 3 < Y < 3 since the sides of the square have length 3. Thus, X < 3 and Y < 3. E(Z) E ( X + Y ) E X + E Y 3 3 3 3 3 x ( 3 x 3 dx + 3 )dx + 3 3 y 3 dy y 3 ( 3 )dy

6. A fair die is rolled times. Calculate the expected sum of the rolls. This is a sum X of independent random variables X i. E(X) E ( i i i E(X i ) 6 i ( ) 6 35 X i ) 6 + 6 +... + 6 6 7. Suppose that A and B each randomly, and independently, choose 3 of objects. Find the expected number of objects (a) chosen by both A and B. Let X be the number of objects that are selected by both A and B. To further simplify the problem we use indicator variables X i. Let X i if object i is selected by both A and B, and X i otherwise, where i. Then, E(X) E ( ) X i E(X i ) i Now we must find E(X i ). We know that X i only takes on one of two values, X i or X i. So, for the case of a sum of independent random indicator variables, E(X i ) P (X i ). Each person can choose 3 of the items. There are 3 ways to choose the item of interest, since a person can draw 3 objects. Since person A and B draw independently, Then, P (X i ) i ( ) 3 ( ) 3 ( ) 3 E(X) E(X i ).9 i i

(b) not chosen by either A or B. The principle is similar to part (a). Let X i if object i is not chosen by A and is not chosen by B. P (X i ) ( 7, ) because the probability that an arbitrary person does not choose object i is 7 and person A and person B draw independently. Then, ( ) 7 ( ) 7 E(X) E(X i ) 4.9 i (c) Chosen by exactly one of A or B i In this case, either person A draws object i and person B does not, or person B draws object i and person A does not. Again, let X i if exactly one of A or B draws object i, X i otherwise. The person that eventually draws object i had probability 3 of of drawing the object and the person that does not draw object i had probability 7 not drawing object i. But there are two ways to arrange this situation. A can draw the object, and B does not, or B draws the object and A does not. Thus, and E(X i ) P (X i ) 3 7 ( E(X) 3 7 ) 4. 9. A total of n balls, numbered through n, are put into n urns, also numbered through n in such a way that ball i is equally likely to go into any of the urns,,..., i. Find (a) the expected number of urns that are empty. Let X be the number of empty urns. Define an indicator variable X i if urn i is empty, X i otherwise. We must find E(X i ) P (X i ). The ith urn remains empty as the first i balls are deposited into the other urns. On the ith drop, urn i must remain empty. Since a ball can land in any of the i urns with equal probability, the probability that the ith ball will not land in urn i is i. To remain empty, the urn must not receive a ball on the i + st drop etc. so the probability that the i + st ball will not land in urn i, is i+. So, ( E(X i ) ) ( ) ( ) (... ) i i + i + n ( ) ( ) ( ) i i n... i i + n i n 3

So, E(X) n E(X i ) i n i n i n i n i n j n n j n n(n ) (b) the probability that none of the urns is empty. For all of the urns to have at least one ball in them, the nth ball must be dropped into the nth urn, which has probability n. The n st ball must be placed in the n st urn which has probability n and so on. So, the probability that none of the urns will be empty is n n n... n! 7. A deck of n cards, numbered through n, is thoroughly shuffled so that all possible n! orderings can be assumed to be equally likely. Suppose you are to make n guesses sequentially, where the ith one is a guess of the card in position i. Let N denote the number of correct guesses. (a) If you are not given any information about your earlier guesses show that, for any strategy, E(N). Proof. Let N i be an indicator variable such that N i if the ith guess is correct. The probability that the guess is correct is simply n. Then, E(N) n E(N i ) i n P (N i ) i n i n n n (b) Suppose that after each guess you are shown the card that was in the position in question. What do you think is the best strategy? Show that under this strategy E(N) n + n n +... + dx log n x Once the card is shown, and assuming the person is not a complete idiot, the player has n cards remaining to guess from. Therefore, the best strategy (or is it just common sense?) is to not guess any cards that have already been shown. 4

Proof. After a card is shown, there is one less card to guess from given our elegant strategy, so the denominator in the probability dereases by one at each guess. So, E(N) n i E(N i ) n + n + n +... + which is a discrete sum. For large n, the sum can be approximated as an integral. n dx log n x (c) Suppose that you are told after each guess whether you are right or wrong. In this case it can be shown that the strategy that maximizes E(N) is one which keeps on guessing the same card until you are told you are correct and then changes to a new card. For this strategy show that E(N) +! + 3! +... + n! e Proof. Let N i be an indicator variable such that N i if the guess about card i is eventually correct. This occurs only if of cards through i, the first card is card, the second card is card and so on, and the last card is card i. There is only one possible way this arrangement can occur, and the total number of arrangements of cards is n!. Therefore, E(N) n E(N i ) i n j j! Note that this is the power series representation of e.. For a group of people, compute (a) the expected number of days of the year that are birthdays of exactly 3 people. Let X be the number of days of the year that are birthdays for exactly 3 people. Let X i if day i is a birthday for 3 people. This is a binomial problem with n and p ( 365) 3. Then, E(X i ) P (X i ) ( 3 ) ( ) 3 ( ) 364 97 365 365 Recall that i represents a day of the year, so in the following calculation, we sum i to i 365. 365 ( E(X) E(X i ) 365 3 i ) ( 365 ) 3 ( 364 ) 97 365 5

(b) the expected number of distinct birthdays. You may be tempted to think of this problem as the number of people that have different birthdays, but that is incorrect. Let X be the number of distinct birthdays; that is, the number of days of the year that are occupied by a birthday. Let X i if someone has a birthday on day i. We use the complement. We find the probability that all people have different birthdays and subtract from. Then, 365 E(X) E(X i ) i ( ) 364 365 [ E(X i ) 365 ( 364 365 ) ] 38. The random variables X and Y have a joint density function given by f(x, y) Compute Cov (X, Y ). { e x x By definition, Cov (X, Y ) E(XY ) E(X)E(Y ). x <, y < x otherwise f X (x) x e x x dy e x E(X) x xe x dx let u x, dv e x so du dx, v e x E(Y ) y e x x dxdy y x ye x x y e x x xe x dx x dx dydx take t x then dt dx. t e t dt 4 Γ() 4 6

Finally, E(XY ) x x xy e x x dydx ye x dydx y e x x dx x e x dx let t x then dt dx. ( ) t e t dt 8 Γ(3) So at last, 4 Cov (X, Y ) E(XY ) E(X)E(Y ) 4 4 8 4. The joint density function of X and Y is given by f(x, y) y e y+ x y Find E(X), E(Y ) and show that Cov (X, Y ). Proof. By definition, Cov (X, Y ) E(XY ) E(X)E(Y ) This problem involves a trick, and we must solve E(Y ) first. f Y (y) e y y y e e y [ e x y e y y+ x y dx e x y dx ] 7

Then, E(Y ) ye y dy Using E(Y ), we can find E(X). Using properties of conditional expectation, E(X) E(E(X Y )) We can find the distribution of X Y by taking the joint density function and treating y as a constant. To make it clearer, let y c, some constant. Then, f X Y (x y) c e c x c c e c e c x which looks exponential with parameter λ c. random variable is λ. But, λ c Y, so The expected value of an exponential E(X) E(E(X Y )) E (Y ) Now we use another property of conditional expectation to find E(XY ). E(XY ) E(E(XY Y )) E(Y E(X Y )) but E(X Y ) Y, so E(XY ) E(Y E(X Y )) E ( Y ). We know that the moment-generating function of Y is M Y (t) λ λ t To find the second moment, take the second derivative and set t. M Y (t) λ (λ t) because E(Y ) λ so λ. So, E(XY ). Then, E(Y ) M Y () λ Cov (X, Y ) E(XY ) E(X)E(Y ) 8

45. If X, X, X 3, X 4 are (pairwise) uncorrelated random variables each having mean and variance, compute the correlations of Recall that pairwise uncorrelated means that any pair of these variables are uncorrelated. (a) X + X and X + X 3. By definition, Corr (X + X, X + X 3) Cov (X + X, X + X3) σ X +X σ X +X 3 Cov (X, X ) + Cov (X, X 3) + Cov (X, X ) + Cov (X, X 3) p Var (X) + Var (X ) Cov (X, X p ) Var (X ) + Var (X 3) Cov (X, X 3) Cov (X i, X j) for i j due to pairwise uncorrelatedness. Cov (X, X ) p Var (X) + Var (X p ) Var (X ) + Var (X 3) Cov (X i, X i) Var (X i) Var (X ) p Var (X) + Var (X p ) Var (X ) + Var (X 3) (b) X + X and X 3 + X 4. Corr (X + X, X 3 + X 4) Cov (X + X, X3 + X4) σ X +X σ X3 +X 4 Cov (X, X 3) + Cov (X, X 4) + Cov (X, X 3) + Cov (X, X 4) p Var (X) + Var (X ) Cov (X, X p ) Var (X 3) + Var (X 4) Cov (X 3, X 4) Cov (X i, X j) for i j due to pairwise uncorrelatedness. 9

Part. Suppose that the random variable X is described by the pdf f X (x) 5x 6, x >. What is the highest moment of X that exists? First, there was a bad typo in this problem that may have confused students. There should not be a y in the problem. Everything is in terms of x. Second, the key to solving this problem is to not use the moment-generating function because we end up with an integral that cannot easily be solved. Instead, we compute the moments using E(X r ). We integrate over x >. E(X r ) x r 5x 6 dx 5 x r 6 dx [ ] x r 5 5 r 5 [ ] We want to find the largest value of r such that 5 x r 5 r 5 exists. First, r 5 because then the quantity is undetermined. Now consider the numerator. If r > 5, then we have divergence. If r < 5, we get a negative exponent for x and it may as well be in the denominator, which will make the whole quantity go to as x. Thus, r < 5, so the fourth moment, E(X 4 ) is the largest moment that exists.. Two chips are drawn at random without replacement from a box that contains five chips numbered through 5. If the sum of chips drawn is even, the random variable X equals 5; if the sum of chips drawn is odd, X 3. First, let s define Y, which is the sum of the two chips X and X. Y X + X. Then, X { 5, Y even 3, Y odd (a) Find the moment-generating function (mgf) for X. Since E(X) n i x ip (X i x i ), we need to find the probabilities P (X 5) and P (X 3). The easiest way to find these is to just enumerate all of the possible drawings. P (X 3) 3 5 P (X 5) 8 5 Then, M X (t) E ( e tx) e 3t ( ) 3 5 + e 5t ( ) 5

(b) Use the mgf to find the first and second non-central moments. To find the first moment, E(X), take the derivative of the moment-generating function and evaluate at t. M X(t) 9 5 e 3t + e 5t M X() 9 5 + 5 Thus, E(X) 5. So, E ( X ) 5.4. (c) Find the expected value and variance of X. M X(t) 7 5 e 3t + e 5t M X() 7 + 5.4 5 The expected value is just the first moment from part (a). E(X) 5. Var (X) E ( X ) E(X) where E ( X ) is the second moment from part (a). 3. Let X have pdf Var (X) 5.4 ( ) 5.36 5 x x f X (x) x x elsewhere (a) Find the moment-generating function (mgf) M X (t) for X. E ( e tx) xe tx dx + xe tx dx + ( x)e tx dx e tx dx xe tx dx take u x, dv e tx so du dx, v t etx.... ( t e t )

(b) Use mgf to find the first and second non-central moments. This part is anticlimactic. We note that the M X (t) has a singularity at t. If we take the first and second derivatives, we are still going to have that obnoxious singularity. The first and second moments do not exist. (c) Find the expected value and variance of X. Since the first and second moments of X do not exist, we have to compute E(X) and Var (X) using the definition. Var (X) E(X ) E(X) E(X) x3 3 x dx + + (x x3 3 x( x)dx ) E(X ) x4 4 4 3 The variance is then x 3 dx + [ + 3 x3 x4 4 x ( x)dx ] Var (X) E(X ) E(X) 4 3 3 4. Find the moment generating function (mgf) for X, if X Uniform(a, b). Since X U(a, b), we know that f X (x) b a, a x b. E ( e tx) b a e tx b a dx b e tx dx b a a [ e tx ] b t(b a) a etb e ta t(b a)

5. Find the moment-generating function for X, if X is a Bernoulli random variable with the pdf f X (x) { p x ( p) x x, x otherwise Use the mgf to find the expected value and variance of X. Note that X can only take on one of two possible values. So, E ( e tx) e t p ( p) + e t p( p) p + pe t M X (t) + p ( e t ) E(X) M X() pe p Var (X) E(X ) E(X) M X() p p p p( p) 6. Find the moment-generating function for X, if X is a Poisson random variable with the pdf f X (x) λx x! e λ. M X (t) E ( e tx) e tx λ λx e x! x tx λ λx e x! x e λ λ + e t λ t λ λ λ + e! +... ( λe e λ t ) x x! x which resembles the power series for e y where y λe t. e λ e λet e λ(et ) 3

7. Find the moment-generating function for Z, if Z is a standard normal random variable with the pdf f Z (z) π e z Use your result to find the moment-generating function for a normally distributed random variable X N(µ, σ). This problem is very tricky, and it is a classic example of how to work with the normal distribution in advanced statistics. If you continue in Statistics, or ever take Bayesian Statistics, you will encounter working with the normal distribution in this way. The goal is to make the computation of M Z (t) look like the original normal distribution. By definition, M Z (t) E ( e tz) π e tz z dz which is simple enough so far. In the next step, I factor out the negative sign in the exponential because the normal distribution has that pesky negative sign in front of the quadratic. π e z dz tz This integral would be a nightmare to integrate because of the quadratic in the exponent, and it does not have the form of the normal distribution. To remedy this, we complete the square. z tz (z t) t I can now break up that exponential and factor part of it out of the integral. π «e (z t) t Believe it or not, we can integrate this! Recall that dz e t π e (z t) dz π e u So take u z t, then du dz and we have e u du π e t π e u e t du π π e t Thus, M Z (t) e t. 4

Recall that Z X µ σ so, X Zσ + µ. To find the moment-generating function of X, we use the hint. M X (t) M Zσ+µ (t) e µt M Z (σt) e µt e (σt) (σt) µt+ e If you continue with Statistics, you will see that derivations involving the normal distribution always follow a similar recipe: ) factor out a negative sign, ) complete the square in the exponent, 3) compare to the standard normal distribution to get to the result. 8. Let X and Y be independent random variables, and let α and β be scalars. Find an expression for the mgf of Z αx + βy in terms of mgfs of X and Y. We just use the definition of moment generating function and use independence. M Z (t) E ( e tz) (e t(αx+βy)) E E (e tαx e tβy) E ( e tαx) E (e tβy) by independence. M X (αt)m Y (βt) 9. Use the mgf to show that if X follows the exponential distribution, cx (c > ) does also. To show that two random variables follow the same distribution, show that the momentgenerating functions are equal. Proof. From a previous problem, we know that M X (t) λ λ t. M cx (t) E ( e ctx) λ λ e ctx cλx dx e xc(t λ) d(cx) cλ c(t λ) λ t λ ( ) [e cx(t λ)] e ctx λe cλx dx λ λ t λ t λ Therefore, since M X (t) M cx (t), cx Exp(λ). [e cx(λ t)] 5

. Show that (a) If X i follows a binomial distribution with n i trials and probability of success p i p where i,,..., n and the X i are independent then n i X i follows a binomial distribution. The probability of success p is the same for all i. Proof. We know that the moment generating function for a Bernoulli random variable Y i is M Yi (t) p + pe t. We also know that a binomial random variable is just a sum of Bernoulli random variables, Y Y +... + Y ni. Then by the properties of moment-generating functions, So analogously, M Y (t) M Y +...+Y ni (t) M Y (t)... M Yni (t) M Xi (t) E ( e tx i ) ( p + pe t ) n i All of the x i s are independent, then similarly, M P n i X i n M Xi (t) i ( p + pe t ) n... ( p + pe t ) nn ( p + pe t ) n +...+n n ( p + pe t ) P n i n i Therefore, n i X i is binomial with parameters N n i n i and p. (b) Referring to part a, show that if p i are unequal, the sum n i X i does not follow a binomial distribution. Now the probability of success is not the same for each i. Replace p in the previous problem with p i. Proof. M P n i X i n M Xi (t) i ( p + p e t ) n... ( p n + p n e t ) nn which is the best we can do, and does not resemble the binomial. Since we do not have a constant probability of success p for all i, the sum is not a binomial random variable. 6