Definition of Random Variable A random variable is a function from a sample space S into the real numbers.

Similar documents
Probability: Terminology and Examples Class 2, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES

Random variables, probability distributions, binomial random variable

E3: PROBABILITY AND STATISTICS lecture notes

Random variables P(X = 3) = P(X = 3) = 1 8, P(X = 1) = P(X = 1) = 3 8.

ST 371 (IV): Discrete Random Variables

Lesson 1. Basics of Probability. Principles of Mathematics 12: Explained! 314

Math/Stats 425 Introduction to Probability. 1. Uncertainty and the axioms of probability

Cardinality. The set of all finite strings over the alphabet of lowercase letters is countable. The set of real numbers R is an uncountable set.

6.2 Permutations continued

People have thought about, and defined, probability in different ways. important to note the consequences of the definition:

The sample space for a pair of die rolls is the set. The sample space for a random number between 0 and 1 is the interval [0, 1].

Probabilistic Strategies: Solutions

For two disjoint subsets A and B of Ω, say that A and B are disjoint events. For disjoint events A and B we take an axiom P(A B) = P(A) + P(B)

M2S1 Lecture Notes. G. A. Young ayoung

Chapter What is the probability that a card chosen from an ordinary deck of 52 cards is an ace? Ans: 4/52.

Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay

You flip a fair coin four times, what is the probability that you obtain three heads.

3. Mathematical Induction

Chapter 3. Cartesian Products and Relations. 3.1 Cartesian Products

Pattern matching probabilities and paradoxes A new variation on Penney s coin game

Bayesian Tutorial (Sheet Updated 20 March)

Chapter 4 Lecture Notes

STAT 315: HOW TO CHOOSE A DISTRIBUTION FOR A RANDOM VARIABLE

Lecture 7: Continuous Random Variables

Handout #1: Mathematical Reasoning

Today s Topics. Primes & Greatest Common Divisors

Probability Theory. Florian Herzog. A random variable is neither random nor variable. Gian-Carlo Rota, M.I.T..

Lecture 8. Confidence intervals and the central limit theorem

1. Prove that the empty set is a subset of every set.

Conditional Probability, Independence and Bayes Theorem Class 3, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom

10.2 Series and Convergence

Sample Space and Probability

Elements of probability theory

Principle of Data Reduction

Chapter 4 - Practice Problems 1

Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS. Part 3: Discrete Uniform Distribution Binomial Distribution

Introduction to Probability

Notes on metric spaces

Practice with Proofs

sin(x) < x sin(x) x < tan(x) sin(x) x cos(x) 1 < sin(x) sin(x) 1 < 1 cos(x) 1 cos(x) = 1 cos2 (x) 1 + cos(x) = sin2 (x) 1 < x 2

STATISTICS HIGHER SECONDARY - SECOND YEAR. Untouchability is a sin Untouchability is a crime Untouchability is inhuman

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 10

Sums of Independent Random Variables

Lecture 1 Introduction Properties of Probability Methods of Enumeration Asrat Temesgen Stockholm University

Section 1.3 P 1 = 1 2. = P n = 1 P 3 = Continuing in this fashion, it should seem reasonable that, for any n = 1, 2, 3,..., =

Lemma 5.2. Let S be a set. (1) Let f and g be two permutations of S. Then the composition of f and g is a permutation of S.

x a x 2 (1 + x 2 ) n.

Extension of measure

Basic Probability Concepts

2. How many ways can the letters in PHOENIX be rearranged? 7! = 5,040 ways.

Chapter 4 & 5 practice set. The actual exam is not multiple choice nor does it contain like questions.

REPEATED TRIALS. The probability of winning those k chosen times and losing the other times is then p k q n k.

Math 4310 Handout - Quotient Vector Spaces

IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem

Question: What is the probability that a five-card poker hand contains a flush, that is, five cards of the same suit?

Introduction to Probability

God created the integers and the rest is the work of man. (Leopold Kronecker, in an after-dinner speech at a conference, Berlin, 1886)

How To Find Out How To Calculate A Premeasure On A Set Of Two-Dimensional Algebra

4.5 Linear Dependence and Linear Independence

Arkansas Tech University MATH 4033: Elementary Modern Algebra Dr. Marcel B. Finan

PROBABILITY. The theory of probabilities is simply the Science of logic quantitatively treated. C.S. PEIRCE

6.3 Conditional Probability and Independence

A Tutorial on Probability Theory

Chapter 4. Probability and Probability Distributions

5 Numerical Differentiation

Lecture 16 : Relations and Functions DRAFT

V. RANDOM VARIABLES, PROBABILITY DISTRIBUTIONS, EXPECTED VALUE

26 Integers: Multiplication, Division, and Order

1 Norms and Vector Spaces

Introduction to Probability

Undergraduate Notes in Mathematics. Arkansas Tech University Department of Mathematics

Metric Spaces Joseph Muscat 2003 (Last revised May 2009)

Math 55: Discrete Mathematics

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 13. Random Variables: Distribution and Expectation

Triangle deletion. Ernie Croot. February 3, 2010

So let us begin our quest to find the holy grail of real analysis.

Discrete Structures for Computer Science

Normal distribution. ) 2 /2σ. 2π σ

Real Roots of Univariate Polynomials with Real Coefficients

MAS108 Probability I

Homework Assignment #2: Answer Key

Probability and Statistics Vocabulary List (Definitions for Middle School Teachers)

Taylor and Maclaurin Series

1 if 1 x 0 1 if 0 x 1

1. Give the 16 bit signed (twos complement) representation of the following decimal numbers, and convert to hexadecimal:

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 2

Lecture 6: Discrete & Continuous Probability and Random Variables

MULTIVARIATE PROBABILITY DISTRIBUTIONS

Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab

Vector and Matrix Norms

For a partition B 1,..., B n, where B i B j = for i. A = (A B 1 ) (A B 2 ),..., (A B n ) and thus. P (A) = P (A B i ) = P (A B i )P (B i )

Scalar Valued Functions of Several Variables; the Gradient Vector

Mathematics Course 111: Algebra I Part IV: Vector Spaces

Metric Spaces. Chapter Metrics

Definition and Calculus of Probability

Chapter 3. Distribution Problems. 3.1 The idea of a distribution The twenty-fold way

Inner Product Spaces

1 The Brownian bridge construction

How To Prove The Dirichlet Unit Theorem

Transcription:

.4 Random Variable Motivation example In an opinion poll, we might decide to ask 50 people whether they agree or disagree with a certain issue. If we record a for agree and 0 for disagree, the sample space for this experiment has 2 50 elements. If we define a variable X=number of s recorded out of 50, we have captured the essence of the problem. Note that the sample space of X is the set of integers {, 2,..., 50} and is much easier to deal with than the original sample space. In defining the quantity X, we have defined a mapping (a function) from the original sample space to a new sample space, usually a set of real numbers. In general, we have the following definition. Definition of Random Variable A random variable is a function from a sample space S into the real numbers. Example.4.2 (Random variables) In some experiments random variables are implicitly used; some examples are these. Experiment Toss two dice Toss a coin 25 times Apply different amounts of fertilizer to corn plants Random variable X =sum of the numbers X =number of heads in 25 tosses X =yield/acre Suppose we have a sample space S = {s,..., s n } with a probability function P and we define a random variable X with range X = {x,..., x m }. We can define a probability function P X on X in the following way. We will observe X = x i if and only if the outcome of the random experiment is an s j S such that X(s j ) = x i. Thus, P X (X = x i ) = P ({s j S : X(s j ) = x i }). ()

Note P X is an induced probability function on X, defined in terms of the original function P. Later, we will simply write P X (X = x i ) = P (X = x i ). Fact The induced probability function defined in () defines a legitimate probability function in that it satisfies the Kolmogorov Axioms. Proof: CX is finite. Therefore B is the set of all subsets of X. We must verify each of the three properties of the axioms. () If A B then P X (A) = P ( xi A{s j S : X(s j ) = x i }) 0 since P is a probability function. (2) P X (X ) = P ( m i={s j S : X(s j ) = x i }) = P (S) =. (3) If A, A 2,... B and pairwise disjoint then P X ( k=a k ) = P ( k={ xi A k {s j S : X(s j ) = x i }}) = P ( xi A k {s j S : X(s j ) = x i } = P X (A k ), k= k= where the second inequality follows from the fact P is a probability function. A note on notation: Random variables will always be denoted with uppercase letters and the realized values of the variable will be denoted by the corresponding lowercase letters. Thus, the random variable X can take the value x. Example.4.3 (Three coin tosses-ii) Consider again the experiment of tossing a fair coin three times independently. Define the random variable X to be the number of heads obtained in the three tosses. A complete enumeration of the value of X for each point in the sample space is s HHH HHT HTH THH TTH THT HTT TTT X(s) 3 2 2 2 0 The range for the random variable X is X = {0,, 2, 3}. Assuming that all eight points in S have probability, by simply counting in the above display we see that the induced probability function on X is given by 2

x 0 2 3 P X (X = x) 3 3 The previous illustrations had both a finite S and finite X, and the definition of P X was straightforward. Such is also the case if X is countable. If X is uncountable, we define the induced probability function, P X, in a manner similar to (). For any set A X, P X (X A) = P ({s S : X(s) A}). (2) This does define a legitimate probability function for which the Kolmogorov Axioms can be verified. Distribution Functions Definition of Distribution The cumulative distribution function (cdf) of a random variable X, denoted by F X (x), is defined by F X (x) = P X (X x), for all x. Example.5.2 (Tossing three coins) Consider the experiment of tossing three fair coins, and let X =number of heads observed. The cdf of X is 0 if < x < 0 if 0 x < F X (x) = if x < 2 2 7 if 2 x < 3 if 3 x <. 3

Remark:. F X is defined for all values of x, not just those in X = {0,, 2, 3}. Thus, for example, F X (2.5) = P (X 2.5) = P (X = 0,, 2) = 7. 2. F X has jumps at the values of x i X and the size of the jump at x i is equal to P (X = x i ). 3. F X = 0 for x < 0 since X cannot be negative, and F X (x) = for x 3 since x is certain to be less than or equal to such a value. F X is right-continuous, namely, the function is continuous when a point is approached from the right. The property of right-continuity is a consequence of the definition of the cdf. In contrast, if we had defined F X (x) = P X (X < x), F X would then be left-continuous. Theorem.5.3 The function F X (x) is a cdf if and only of the following three conditions hold: a. lim x F (x) = 0 and lim x F (x) =. b. F (x) is a nondecreasing function of x. c. F (x) is right-continuous; that is, for every number x 0, lim x x0 F (x) = F (x 0 ). Example.5.4 (Tossing for a head) Suppose we do an experiment that consists of tossing a coin until a head appears. Let p =probability of a head on any given toss, and define X =number of tosses required to get a head. Then, for any x =, 2,..., The cdf is F X (x) = P (X x) = P (X = x) = ( p) x p. x P (X = i) = i= x ( p) i p = ( p) x. It is easy to show that if 0 < p <, then F X (x) satisfies the conditions of Theorem.5.3. First, i= lim F X(x) = 0 x 4

since F X (x) = 0 for all x < 0, and lim F X(x) = lim ( ( p) x ) =, x x where x goes through only integer values when this limit is taken. To verify property (b), we simply note that the sum contains more positive terms as x increases. Finally, to verify (c), note that, for any x, F X (x + ɛ) = F X (x) if ɛ > 0 is sufficiently small. Hence, so F X (x) is right-continuous. lim ɛ 0 F X (x + ɛ) = F X (x), Example.5.5 (Continuous cdf) An example of a continuous cdf (logistic distribution) is the function It is easy to verify that F X (x) = + e x. Differentiating F X (x) gives lim F X(x) = 0 and lim F X (x) =. x x d dx F X(x) = e x ( + e x ) 2 > 0, showing that F X (x) is increasing. F X is not only right-continuous, but also continuous. Definition of Continuous Random Variable A random variable X is continuous if F X (x) is a continuous function of x. A random variable X is discrete if F X (x) is a step function of x. We close this section with a theorem formally stating that F X completely determines the probability distribution of a random variable X. This is true if P (X A) is defined only for events A in B, the smallest sigma algebra containing all the intervals of real numbers of the form (a, b), [a, b), (a, b], and [a, b]. If probabilities are defined for a larger class of events, it is possible for two random variables to have the same distribution function but not the same probability for every event (see Chung 974, page 27). Definition of Identical Random Variables The random variables X and Y are identically distributed if, for every set A B, P (X A) = P (Y A). 5

Note that two random variables that are identically distributed are not necessarily equal. That is, the above definition does not say that X = Y. Example.5.9 (identically distributed random variables) Consider the experiment of tossing a fair coin three times. Define the random variables X and Y by X =number of heads observed and Y =number of tails observed. For each k = 0,, 2, 3, we have P (X = k) = P (Y = k). So X and Y are identically distributed. However, for no sample point do we have X(s) = Y (s). Theorem.5.0 The following two statements are equivalent: a. The random variables X and Y are identically distributed. b. F X (x) = F Y (x) for every x. Proof: To show equivalence we must show that each statement implies the other. We first show that (a) (b). Because X and Y are identically distributed, for any set A B, P (X A) = P (Y A). In particular, for every x, the set (, x] is in B, and F X (x) = P (X (, x]) = P (Y (, x]) = F Y (x). The above argument showed that if the X and Y probabilities agreed in all sets, then agreed on intervals. To show (b) (a), we must prove if the X and Y probabilities agree on all intervals, then they agree on all sets. For more details see Chung (974, section 2.2). 6