For a partition B 1,..., B n, where B i B j = for i. A = (A B 1 ) (A B 2 ),..., (A B n ) and thus. P (A) = P (A B i ) = P (A B i )P (B i )
|
|
- Rosalyn Booker
- 1 years ago
- Views:
Transcription
1 Probability Review Cynthia Rudin A probability space, defined by Kolmogorov ( ) consists of: A set of outcomes S, e.g., for the roll of a die, S = {1, 2, 3, 4, 5, 6}, for the roll of two dice, S =,,,,..., temperature on Monday, S = [ 50, 50]. A set of events, where an event is a subset of S, e.g., roll at least one 3 :,..., temperature above 80 : (80, 150] The union, intersection and complement of events are also events (an algebra). A probability measure, which gives a number between 0 and 1 to each event, where P (S) = 1 and A B = P (A B) = P (A) + P (B). Think of P as measuring the size of a set, or an area of the Venn diagram. The conditional probability of event A given event B is: P (A B) P (A B) :=. P (B) 1
2 Event A is independent of B if P (A B) = P (A). That is, knowing B occurred doesn t impact whether A occurred (e.g. A and B each represent an event where a coin returned heads). In that case, P (A B) P (A) = P (A B) := so P (A B) = P (A)P (B). P (B) Do not confuse independence with disjointness. Disjoint events A and B have P (A B) = 0. For a partition B 1,..., B n, where B i B j = for i and thus A = (A B 1 ) (A B 2 ),..., (A B n ) = j and B 1 B 2 B n = S then n P (A) = P (A B i ) = P (A B i )P (B i ) i=1 i from the definition of conditional probability. Bayes Theorem P (A B)P (B) P (B A) =. P (A) Derivation of Bayes Rule Monty Hall Problem A random variable (r.v.) assigns a number to each outcome in S. Example 1: toss 2 dice: random var. X is the sum of the numbers on the dice. Example 2: collections of transistors: random var X is the min of the lifetimes of the transistors. 2
3 A probability density function (pdf... or probability mass function pmf) for random variable X is defined via: f(x) = P (X = x) for discrete distributions (Note f(x) 0, x f(x) = 1.) b P (a X b) = a f(x)dx for continuous distributions. (Note f(x) 0, f(x) = 1.) The cumulative distribution function (cdf) for r.v. X is: F (x) = P (X x) = f(k) (discrete) k x x = f(y)dy (continuous) 3
4 The expected value (mean) of an r.v. X is: E(X) = µ = xf(x) (discrete) x Expectation is linear, meaning Roulette The variance of an r.v. X is: E(X) = µ = xf(x)dx (continuous). x E(aX + by ) = ae(x) + be(y ). V ar(x) = σ 2 = E(X µ) 2. Variance measures dispersion around the mean. Variance is not linear: The standard deviation (SD) is: V ar(ax + b) = a 2 V ar(x). SD(x) = σ = V ar(x). Note that people sometimes use another definition for variance that is equivalent: V ar(x) = σ 2 = E(X 2 ) (E(X)) 2. Sum of Two Dice Quantiles/Percentiles The p th quantile (or 100p th percentile), denoted θ p, of r.v. X obeys: P (X θ p ) = p. Note: 50 th percentile, or.5 th quantile, is the median θ.5. Exponential Distribution 4
5 Section 2.5. Jointly Distributed R.V. s Jointly distributed r.v. s have joint pdf s: f(x, y) = P (X = x, Y = y). Their covariance is defined by Cov(X, Y ) := σ x,y := E[(X µ x )(Y µ y )]. The first term considers dispersion from the mean of X, the second term is dispersion from the mean of Y. If X and Y are independent, then E(XY ) = E(X)E(Y ) = µ x µ y (see expected value on Wikipedia). So, multiplying the terms out and passing the expectation through, we get: Cov(X, Y ) = E[(X µ x )(Y µ y )] = E(XY ) µ x E(Y ) µ y E(X) + µ x µ y = 0. Useful relationships: 1. Cov(X, X) = E[(X µ x ) 2 ] = V ar(x) 2. Cov(aX + c, by + d) = ab Cov(X, Y ) 3. V ar(x±y ) = V ar(x)+v ar(y )±2Cov(X, Y ) where Cov(X, Y ) is 0 if X and Y are indep. The correlation coefficient is a normalized version of covariance so that its range is [-1,1]: Cov(X, Y ) σ XY ρ XY := Corr(X, Y ) := =. V ar(x)v ar(y ) σ X σ Y Section 2.6. Chebyshev s Inequality (we don t use it much in this class) σ 2 For c > 0, P ( X EX c). c 2 Chebyshev s inequality is useful when you want a upper bound on how often an r.v. is far from its mean. You can t use it for directly calculating P ( X EX c), only for bounding it. Weak Law of Large Numbers Let s say we want to know the average number of pizza slices we will sell in a day. We ll measure pizza sales over the next couple weeks. Each day gets a random variable X i which represents the amount of pizza we sell on day i. Each X i has the same distribution, because each day basically has the same kind of randomness as the others. Let s say we take the average of over the couple weeks we measured. There s a random variable for that, it s X = 1 n i X i. Does X have anything to do with the average pizza sales per day? In other words, does measuring X tell us anything about the X i s? For instance, (on average) is X close to the average sales per day, E(X i )? 5
6 Does it matter what the distribution of pizza sales is? For instance, what if we usually have 1-4 customers but sometimes we have a conference where we have people; that is kind of a weird distribution. Does that mean that the X i needs to be adjusted in some way to help us understand the X i s? It turns out the first answer is yes, in fact pretty often, the average X is very similar to the average value of X i. Especially when n is large. The second answer is also yes, but as long as n is large, X is very similar to the average value of X i. This means that no matter what the distribution is, as long as we measure enough days, we can get a pretty good sense of the average pizza sales. Weak LLN: Let X be a r.v. for the sample mean X = 1 n i X i of n iid (independent and identically distributed) r.v. s. The distribution for each X i has finite mean µ and variance 2 σ. Then for any c > 0, P ( X µ c) 0 as n. W eak LLN says that X approaches µ when n is large. Weak LLN is nice because it tells us that no matter what the distribution of the X i s is, the sample mean approaches the true mean. Proof Weak LLN using Chebyshev Section 2.7. Selected Discrete Distributions Bernoulli X Bernoulli(p) coin flipping f(x) = P (X = x) = { p if x = 1 heads 1 p if x = 0 tails Binomial X Bin(n, p) n coins flipping, balls in a bag drawn with replacement, balls in an infinite bag ( ) n f(x ) = P (X = x) = p x (1 p) n x for x = 0, 1,..., n, x where n is the number of ways to distribute x successes and n x failures, x ( ) n n! =. x x!(n x)! (If you have n coins flipping, f(x) is the probability to get x heads, p is the probability of heads.) X Bin(n, p) has E(X) = np, V ar(x) = np(1 p). You ll end up needing these facts in Chapter 9. 6
7 Hypergeometric X HyGE(N, M, n) balls in a bag drawn without replacement, where: N is the size of the total population (the number of balls in the bag), M is the number of items that have a specific attribute (perhaps M balls are red), n is our sample size, f(x) is the probability that x items from the sample have the attribute. ways to draw x balls ways to draw n x balls ( ) ( ) M N M f(x) = with attribute without attribute x n x = ( ) ways to draw n balls N. n Multinomial Distribution generalization of binomial Think of customers choosing backpacks of different colors. A random group of n customers each choose their favorite color backpack. There is a multinomial distribution governing how many backpacks of each color were chosen by the group. Below, x k is the number of people who ordered the k th color backpack. Also, f(x) is the probability that x 1 customers chose color 1, x 2 customers chose color 2, and so on. f(x 1, x 2, x 3,..., x k ) = P (X 1 = x 1, X 2 = x 2, X 3 = x 3,..., X k = x k ) n! x 1 x 2 x 3 x = p p p... p k k x 1!x 2!x 3!... x k! where x i 0 for all i and i x i = n (there are n total customers), and p i is the probability that outcome i occurs. (p i is the probability to choose the i th color backpack). 2.8 Selected Continuous Distributions Uniform Distribution X U[a, b] f(x) = { 1 a x b b a 0 otherwise Poisson Distribution X P ois(λ) binomial when np λ, rare events X P ois(λ) has e λ λ x f(x) = x = 0,1,2,... x! E(X) = λ, V ar(x) = λ. 7
8 Exponential Distribution X Exp(λ) waiting times for Poisson events f(x) = λe λx for x 0 Gamma Distribution X Gamma(λ, r) sums of r iid exponential r.v. s, sums of waiting times for Poisson events r 1 λx λ r x e f(x) = for x 0, Γ(r) where Γ(r) is the Gamma function, which is a generalization of factorial. Customers on line Normal (Gaussian) Distribution X N(µ, σ 2 ) We have that: 1 (x µ) 2 /2σ 2 f(x) = e for < x <. σ 2π E(X) = µ, V ar(x) = σ 2. We often standardize a normal distribution by shifting its mean to 0 and scaling its variance to 1: X µ If X N(µ, σ 2 ) then Z = N(0, 1) standard normal σ Since we can standardize any gaussian, let s work mostly with the standard normal Z N(0, 1): pdf: 1 φ(z) = e z2 /2 2π z cdf: Φ(z) = P (Z z) = φ(y)dy We can t integrate the cdf in closed form. This means that in order to get from z to Φ(z) (or the reverse) we need to get the answer from a book or computer where they integrated 8
9 it numerically, by computing an approximation to the integral. MATLAB uses the command normcdf. Table A.3 in your book has values for Φ(z) for many possible z s. Read the table left to right, and up to down. So if the entries in the table look like this: z This means that for z = 2.43, then Φ(z) = P (Z z) = So the table relates z to Φ(z). You can either be given z and need Φ(z) or vice versa. 75 th percentile calculation Denote z α as the solution to 1 Φ(z) = α. z α is called the upper α critical point or the 100(1 α) th percentile. Linear Combinations of Normal r.v. s are also normal. For n iid observations from N(µ, σ 2 ), that is. the sample mean X obeys: X i N(µ, σ 2 ) for i = 1,..., n, ( ) σ X 2 N µ,. n Hmm, where have you seen σ 2 /n before? Of course! You saw it here: V ar( X) = V ar 1 n i X i = 1 n 2 i V ar(x i ) = nσ2 n 2 = σ2 n. This is the variance of the mean of n iid r.v. s who each have V ar(x i ) = σ 2. 9
10 MIT OpenCourseWare J / ESD.07J Statistical Thinking and Data Analysis Fall 2011 For information about citing these materials or our Terms of Use, visit:
Chapters 5. Multivariate Probability Distributions
Chapters 5. Multivariate Probability Distributions Random vectors are collection of random variables defined on the same sample space. Whenever a collection of random variables are mentioned, they are
Department of Mathematics, Indian Institute of Technology, Kharagpur Assignment 2-3, Probability and Statistics, March 2015. Due:-March 25, 2015.
Department of Mathematics, Indian Institute of Technology, Kharagpur Assignment -3, Probability and Statistics, March 05. Due:-March 5, 05.. Show that the function 0 for x < x+ F (x) = 4 for x < for x
P (x) 0. Discrete random variables Expected value. The expected value, mean or average of a random variable x is: xp (x) = v i P (v i )
Discrete random variables Probability mass function Given a discrete random variable X taking values in X = {v 1,..., v m }, its probability mass function P : X [0, 1] is defined as: P (v i ) = Pr[X =
MATH 201. Final ANSWERS August 12, 2016
MATH 01 Final ANSWERS August 1, 016 Part A 1. 17 points) A bag contains three different types of dice: four 6-sided dice, five 8-sided dice, and six 0-sided dice. A die is drawn from the bag and then rolled.
Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)
Summary of Formulas and Concepts Descriptive Statistics (Ch. 1-4) Definitions Population: The complete set of numerical information on a particular quantity in which an investigator is interested. We assume
5. Continuous Random Variables
5. Continuous Random Variables Continuous random variables can take any value in an interval. They are used to model physical characteristics such as time, length, position, etc. Examples (i) Let X be
Math 141. Lecture 7: Variance, Covariance, and Sums. Albyn Jones 1. 1 Library 304. jones/courses/141
Math 141 Lecture 7: Variance, Covariance, and Sums Albyn Jones 1 1 Library 304 jones@reed.edu www.people.reed.edu/ jones/courses/141 Last Time Variance: expected squared deviation from the mean: Standard
ST 371 (VIII): Theory of Joint Distributions
ST 371 (VIII): Theory of Joint Distributions So far we have focused on probability distributions for single random variables. However, we are often interested in probability statements concerning two or
Expectation Discrete RV - weighted average Continuous RV - use integral to take the weighted average
PHP 2510 Expectation, variance, covariance, correlation Expectation Discrete RV - weighted average Continuous RV - use integral to take the weighted average Variance Variance is the average of (X µ) 2
Contents. TTM4155: Teletraffic Theory (Teletrafikkteori) Probability Theory Basics. Yuming Jiang. Basic Concepts Random Variables
TTM4155: Teletraffic Theory (Teletrafikkteori) Probability Theory Basics Yuming Jiang 1 Some figures taken from the web. Contents Basic Concepts Random Variables Discrete Random Variables Continuous Random
E3: PROBABILITY AND STATISTICS lecture notes
E3: PROBABILITY AND STATISTICS lecture notes 2 Contents 1 PROBABILITY THEORY 7 1.1 Experiments and random events............................ 7 1.2 Certain event. Impossible event............................
Math 431 An Introduction to Probability. Final Exam Solutions
Math 43 An Introduction to Probability Final Eam Solutions. A continuous random variable X has cdf a for 0, F () = for 0 <
University of California, Los Angeles Department of Statistics. Random variables
University of California, Los Angeles Department of Statistics Statistics Instructor: Nicolas Christou Random variables Discrete random variables. Continuous random variables. Discrete random variables.
6. Jointly Distributed Random Variables
6. Jointly Distributed Random Variables We are often interested in the relationship between two or more random variables. Example: A randomly chosen person may be a smoker and/or may get cancer. Definition.
Joint Exam 1/P Sample Exam 1
Joint Exam 1/P Sample Exam 1 Take this practice exam under strict exam conditions: Set a timer for 3 hours; Do not stop the timer for restroom breaks; Do not look at your notes. If you believe a question
Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS. Part 3: Discrete Uniform Distribution Binomial Distribution
Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS Part 3: Discrete Uniform Distribution Binomial Distribution Sections 3-5, 3-6 Special discrete random variable distributions we will cover
Joint Probability Distributions and Random Samples (Devore Chapter Five)
Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01 Probability and Statistics for Engineers Winter 2010-2011 Contents 1 Joint Probability Distributions 1 1.1 Two Discrete
Chapter 5. Random variables
Random variables random variable numerical variable whose value is the outcome of some probabilistic experiment; we use uppercase letters, like X, to denote such a variable and lowercase letters, like
Practice Exam 1: Long List 18.05, Spring 2014
Practice Eam : Long List 8.05, Spring 204 Counting and Probability. A full house in poker is a hand where three cards share one rank and two cards share another rank. How many ways are there to get a full-house?
Chapter 4 Lecture Notes
Chapter 4 Lecture Notes Random Variables October 27, 2015 1 Section 4.1 Random Variables A random variable is typically a real-valued function defined on the sample space of some experiment. For instance,
Lecture 7: Continuous Random Variables
Lecture 7: Continuous Random Variables 21 September 2005 1 Our First Continuous Random Variable The back of the lecture hall is roughly 10 meters across. Suppose it were exactly 10 meters, and consider
Practice problems for Homework 11 - Point Estimation
Practice problems for Homework 11 - Point Estimation 1. (10 marks) Suppose we want to select a random sample of size 5 from the current CS 3341 students. Which of the following strategies is the best:
Random Variables, Expectation, Distributions
Random Variables, Expectation, Distributions CS 5960/6960: Nonparametric Methods Tom Fletcher January 21, 2009 Review Random Variables Definition A random variable is a function defined on a probability
3 Multiple Discrete Random Variables
3 Multiple Discrete Random Variables 3.1 Joint densities Suppose we have a probability space (Ω, F,P) and now we have two discrete random variables X and Y on it. They have probability mass functions f
Statistics 100A Homework 8 Solutions
Part : Chapter 7 Statistics A Homework 8 Solutions Ryan Rosario. A player throws a fair die and simultaneously flips a fair coin. If the coin lands heads, then she wins twice, and if tails, the one-half
Probability & Statistics Primer Gregory J. Hakim University of Washington 2 January 2009 v2.0
Probability & Statistics Primer Gregory J. Hakim University of Washington 2 January 2009 v2.0 This primer provides an overview of basic concepts and definitions in probability and statistics. We shall
Covariance and Correlation. Consider the joint probability distribution f XY (x, y).
Chapter 5: JOINT PROBABILITY DISTRIBUTIONS Part 2: Section 5-2 Covariance and Correlation Consider the joint probability distribution f XY (x, y). Is there a relationship between X and Y? If so, what kind?
1.1 Introduction, and Review of Probability Theory... 3. 1.1.1 Random Variable, Range, Types of Random Variables... 3. 1.1.2 CDF, PDF, Quantiles...
MATH4427 Notebook 1 Spring 2016 prepared by Professor Jenny Baglivo c Copyright 2009-2016 by Jenny A. Baglivo. All Rights Reserved. Contents 1 MATH4427 Notebook 1 3 1.1 Introduction, and Review of Probability
Topic 4: Multivariate random variables. Multiple random variables
Topic 4: Multivariate random variables Joint, marginal, and conditional pmf Joint, marginal, and conditional pdf and cdf Independence Expectation, covariance, correlation Conditional expectation Two jointly
MATH4427 Notebook 2 Spring 2016. 2 MATH4427 Notebook 2 3. 2.1 Definitions and Examples... 3. 2.2 Performance Measures for Estimators...
MATH4427 Notebook 2 Spring 2016 prepared by Professor Jenny Baglivo c Copyright 2009-2016 by Jenny A. Baglivo. All Rights Reserved. Contents 2 MATH4427 Notebook 2 3 2.1 Definitions and Examples...................................
Section 5.1 Continuous Random Variables: Introduction
Section 5. Continuous Random Variables: Introduction Not all random variables are discrete. For example:. Waiting times for anything (train, arrival of customer, production of mrna molecule from gene,
DISCRETE RANDOM VARIABLES
DISCRETE RANDOM VARIABLES DISCRETE RANDOM VARIABLES Documents prepared for use in course B01.1305, New York University, Stern School of Business Definitions page 3 Discrete random variables are introduced
ST 371 (IV): Discrete Random Variables
ST 371 (IV): Discrete Random Variables 1 Random Variables A random variable (rv) is a function that is defined on the sample space of the experiment and that assigns a numerical variable to each possible
( ) = P Z > = P( Z > 1) = 1 Φ(1) = 1 0.8413 = 0.1587 P X > 17
4.6 I company that manufactures and bottles of apple juice uses a machine that automatically fills 6 ounce bottles. There is some variation, however, in the amounts of liquid dispensed into the bottles
MULTIVARIATE PROBABILITY DISTRIBUTIONS
MULTIVARIATE PROBABILITY DISTRIBUTIONS. PRELIMINARIES.. Example. Consider an experiment that consists of tossing a die and a coin at the same time. We can consider a number of random variables defined
Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab
Monte Carlo Simulation: IEOR E4703 Fall 2004 c 2004 by Martin Haugh Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab 1 Overview of Monte Carlo Simulation 1.1 Why use simulation?
Examination 110 Probability and Statistics Examination
Examination 0 Probability and Statistics Examination Sample Examination Questions The Probability and Statistics Examination consists of 5 multiple-choice test questions. The test is a three-hour examination
Math 461 Fall 2006 Test 2 Solutions
Math 461 Fall 2006 Test 2 Solutions Total points: 100. Do all questions. Explain all answers. No notes, books, or electronic devices. 1. [105+5 points] Assume X Exponential(λ). Justify the following two
An Introduction to Basic Statistics and Probability
An Introduction to Basic Statistics and Probability Shenek Heyward NCSU An Introduction to Basic Statistics and Probability p. 1/4 Outline Basic probability concepts Conditional probability Discrete Random
Random variables, probability distributions, binomial random variable
Week 4 lecture notes. WEEK 4 page 1 Random variables, probability distributions, binomial random variable Eample 1 : Consider the eperiment of flipping a fair coin three times. The number of tails that
Summary of Probability
Summary of Probability Mathematical Physics I Rules of Probability The probability of an event is called P(A), which is a positive number less than or equal to 1. The total probability for all possible
Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS
Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS Part 4: Geometric Distribution Negative Binomial Distribution Hypergeometric Distribution Sections 3-7, 3-8 The remaining discrete random
7 Hypothesis testing - one sample tests
7 Hypothesis testing - one sample tests 7.1 Introduction Definition 7.1 A hypothesis is a statement about a population parameter. Example A hypothesis might be that the mean age of students taking MAS113X
A (random) experiment is an activity with observable results. The sample space S of an experiment is the set of all outcomes.
Chapter 7 Probability 7.1 Experiments, Sample Spaces, and Events A (random) experiment is an activity with observable results. The sample space S of an experiment is the set of all outcomes. Each outcome
Statistics 100A Homework 7 Solutions
Chapter 6 Statistics A Homework 7 Solutions Ryan Rosario. A television store owner figures that 45 percent of the customers entering his store will purchase an ordinary television set, 5 percent will purchase
SYSM 6304: Risk and Decision Analysis Lecture 3 Monte Carlo Simulation
SYSM 6304: Risk and Decision Analysis Lecture 3 Monte Carlo Simulation M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas Email: M.Vidyasagar@utdallas.edu September 19, 2015 Outline
Exercises with solutions (1)
Exercises with solutions (). Investigate the relationship between independence and correlation. (a) Two random variables X and Y are said to be correlated if and only if their covariance C XY is not equal
Chapter 4 Expected Values
Chapter 4 Expected Values 4. The Expected Value of a Random Variables Definition. Let X be a random variable having a pdf f(x). Also, suppose the the following conditions are satisfied: x f(x) converges
Notes 11 Autumn 2005
MAS 08 Probabilit I Notes Autumn 005 Two discrete random variables If X and Y are discrete random variables defined on the same sample space, then events such as X = and Y = are well defined. The joint
Lecture 8. Confidence intervals and the central limit theorem
Lecture 8. Confidence intervals and the central limit theorem Mathematical Statistics and Discrete Mathematics November 25th, 2015 1 / 15 Central limit theorem Let X 1, X 2,... X n be a random sample of
Chapter 2: Data quantifiers: sample mean, sample variance, sample standard deviation Quartiles, percentiles, median, interquartile range Dot diagrams
Review for Final Chapter 2: Data quantifiers: sample mean, sample variance, sample standard deviation Quartiles, percentiles, median, interquartile range Dot diagrams Histogram Boxplots Chapter 3: Set
Lecture 8: More Continuous Random Variables
Lecture 8: More Continuous Random Variables 26 September 2005 Last time: the eponential. Going from saying the density e λ, to f() λe λ, to the CDF F () e λ. Pictures of the pdf and CDF. Today: the Gaussian
MT426 Notebook 3 Fall 2012 prepared by Professor Jenny Baglivo. 3 MT426 Notebook 3 3. 3.1 Definitions... 3. 3.2 Joint Discrete Distributions...
MT426 Notebook 3 Fall 2012 prepared by Professor Jenny Baglivo c Copyright 2004-2012 by Jenny A. Baglivo. All Rights Reserved. Contents 3 MT426 Notebook 3 3 3.1 Definitions............................................
Probability and Statistics Vocabulary List (Definitions for Middle School Teachers)
Probability and Statistics Vocabulary List (Definitions for Middle School Teachers) B Bar graph a diagram representing the frequency distribution for nominal or discrete data. It consists of a sequence
Chapter 4. Multivariate Distributions
1 Chapter 4. Multivariate Distributions Joint p.m.f. (p.d.f.) Independent Random Variables Covariance and Correlation Coefficient Expectation and Covariance Matrix Multivariate (Normal) Distributions Matlab
Normal distribution Approximating binomial distribution by normal 2.10 Central Limit Theorem
1.1.2 Normal distribution 1.1.3 Approimating binomial distribution by normal 2.1 Central Limit Theorem Prof. Tesler Math 283 October 22, 214 Prof. Tesler 1.1.2-3, 2.1 Normal distribution Math 283 / October
Joint Probability Distributions and Random Samples. Week 5, 2011 Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage
5 Joint Probability Distributions and Random Samples Week 5, 2011 Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage Two Discrete Random Variables The probability mass function (pmf) of a single
Random Variables. Chapter 2. Random Variables 1
Random Variables Chapter 2 Random Variables 1 Roulette and Random Variables A Roulette wheel has 38 pockets. 18 of them are red and 18 are black; these are numbered from 1 to 36. The two remaining pockets
IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem
IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem Time on my hands: Coin tosses. Problem Formulation: Suppose that I have
Definition: Suppose that two random variables, either continuous or discrete, X and Y have joint density
HW MATH 461/561 Lecture Notes 15 1 Definition: Suppose that two random variables, either continuous or discrete, X and Y have joint density and marginal densities f(x, y), (x, y) Λ X,Y f X (x), x Λ X,
MAS108 Probability I
1 QUEEN MARY UNIVERSITY OF LONDON 2:30 pm, Thursday 3 May, 2007 Duration: 2 hours MAS108 Probability I Do not start reading the question paper until you are instructed to by the invigilators. The paper
1 Sufficient statistics
1 Sufficient statistics A statistic is a function T = rx 1, X 2,, X n of the random sample X 1, X 2,, X n. Examples are X n = 1 n s 2 = = X i, 1 n 1 the sample mean X i X n 2, the sample variance T 1 =
STAT 315: HOW TO CHOOSE A DISTRIBUTION FOR A RANDOM VARIABLE
STAT 315: HOW TO CHOOSE A DISTRIBUTION FOR A RANDOM VARIABLE TROY BUTLER 1. Random variables and distributions We are often presented with descriptions of problems involving some level of uncertainty about
CHAPTER 6: Continuous Uniform Distribution: 6.1. Definition: The density function of the continuous random variable X on the interval [A, B] is.
Some Continuous Probability Distributions CHAPTER 6: Continuous Uniform Distribution: 6. Definition: The density function of the continuous random variable X on the interval [A, B] is B A A x B f(x; A,
ECE302 Spring 2006 HW7 Solutions March 11, 2006 1
ECE32 Spring 26 HW7 Solutions March, 26 Solutions to HW7 Note: Most of these solutions were generated by R. D. Yates and D. J. Goodman, the authors of our textbook. I have added comments in italics where
6.041/6.431 Spring 2008 Quiz 2 Wednesday, April 16, 7:30-9:30 PM. SOLUTIONS
6.4/6.43 Spring 28 Quiz 2 Wednesday, April 6, 7:3-9:3 PM. SOLUTIONS Name: Recitation Instructor: TA: 6.4/6.43: Question Part Score Out of 3 all 36 2 a 4 b 5 c 5 d 8 e 5 f 6 3 a 4 b 6 c 6 d 6 e 6 Total
Lecture Notes 1. Brief Review of Basic Probability
Probability Review Lecture Notes Brief Review of Basic Probability I assume you know basic probability. Chapters -3 are a review. I will assume you have read and understood Chapters -3. Here is a very
Math/Stats 342: Solutions to Homework
Math/Stats 342: Solutions to Homework Steven Miller (sjm1@williams.edu) November 17, 2011 Abstract Below are solutions / sketches of solutions to the homework problems from Math/Stats 342: Probability
Math 151. Rumbos Spring 2014 1. Solutions to Assignment #22
Math 151. Rumbos Spring 2014 1 Solutions to Assignment #22 1. An experiment consists of rolling a die 81 times and computing the average of the numbers on the top face of the die. Estimate the probability
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES Contents 1. Random variables and measurable functions 2. Cumulative distribution functions 3. Discrete
Hypothesis Testing COMP 245 STATISTICS. Dr N A Heard. 1 Hypothesis Testing 2 1.1 Introduction... 2 1.2 Error Rates and Power of a Test...
Hypothesis Testing COMP 45 STATISTICS Dr N A Heard Contents 1 Hypothesis Testing 1.1 Introduction........................................ 1. Error Rates and Power of a Test.............................
Worked examples Multiple Random Variables
Worked eamples Multiple Random Variables Eample Let X and Y be random variables that take on values from the set,, } (a) Find a joint probability mass assignment for which X and Y are independent, and
STAT 360 Probability and Statistics. Fall 2012
STAT 360 Probability and Statistics Fall 2012 1) General information: Crosslisted course offered as STAT 360, MATH 360 Semester: Fall 2012, Aug 20--Dec 07 Course name: Probability and Statistics Number
UNIT I: RANDOM VARIABLES PART- A -TWO MARKS
UNIT I: RANDOM VARIABLES PART- A -TWO MARKS 1. Given the probability density function of a continuous random variable X as follows f(x) = 6x (1-x) 0
Introduction to Probability
Introduction to Probability EE 179, Lecture 15, Handout #24 Probability theory gives a mathematical characterization for experiments with random outcomes. coin toss life of lightbulb binary data sequence
Probability & Probability Distributions
Probability & Probability Distributions Carolyn J. Anderson EdPsych 580 Fall 2005 Probability & Probability Distributions p. 1/61 Probability & Probability Distributions Elementary Probability Theory Definitions
Maximum Likelihood Estimation
Math 541: Statistical Theory II Lecturer: Songfeng Zheng Maximum Likelihood Estimation 1 Maximum Likelihood Estimation Maximum likelihood is a relatively simple method of constructing an estimator for
Parametric Models Part I: Maximum Likelihood and Bayesian Density Estimation
Parametric Models Part I: Maximum Likelihood and Bayesian Density Estimation Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Fall 2015 CS 551, Fall 2015
Continuous Random Variables
Continuous Random Variables COMP 245 STATISTICS Dr N A Heard Contents 1 Continuous Random Variables 2 11 Introduction 2 12 Probability Density Functions 3 13 Transformations 5 2 Mean, Variance and Quantiles
Lecture 6: Discrete & Continuous Probability and Random Variables
Lecture 6: Discrete & Continuous Probability and Random Variables D. Alex Hughes Math Camp September 17, 2015 D. Alex Hughes (Math Camp) Lecture 6: Discrete & Continuous Probability and Random September
A Tutorial on Probability Theory
Paola Sebastiani Department of Mathematics and Statistics University of Massachusetts at Amherst Corresponding Author: Paola Sebastiani. Department of Mathematics and Statistics, University of Massachusetts,
Final Mathematics 5010, Section 1, Fall 2004 Instructor: D.A. Levin
Final Mathematics 51, Section 1, Fall 24 Instructor: D.A. Levin Name YOU MUST SHOW YOUR WORK TO RECEIVE CREDIT. A CORRECT ANSWER WITHOUT SHOWING YOUR REASONING WILL NOT RECEIVE CREDIT. Problem Points Possible
Lecture 3: Continuous distributions, expected value & mean, variance, the normal distribution
Lecture 3: Continuous distributions, expected value & mean, variance, the normal distribution 8 October 2007 In this lecture we ll learn the following: 1. how continuous probability distributions differ
Introduction to Hypothesis Testing. Point estimation and confidence intervals are useful statistical inference procedures.
Introduction to Hypothesis Testing Point estimation and confidence intervals are useful statistical inference procedures. Another type of inference is used frequently used concerns tests of hypotheses.
Continuous Random Variables and Probability Distributions. Stat 4570/5570 Material from Devore s book (Ed 8) Chapter 4 - and Cengage
4 Continuous Random Variables and Probability Distributions Stat 4570/5570 Material from Devore s book (Ed 8) Chapter 4 - and Cengage Continuous r.v. A random variable X is continuous if possible values
MAS113 Introduction to Probability and Statistics Exercises
MAS113 Introduction to Probability and Statistics Exercises You will be told which problems to work on each week at the lecture each Monday, and you should try them before your class each Wednesday. Every
You flip a fair coin four times, what is the probability that you obtain three heads.
Handout 4: Binomial Distribution Reading Assignment: Chapter 5 In the previous handout, we looked at continuous random variables and calculating probabilities and percentiles for those type of variables.
INSURANCE RISK THEORY (Problems)
INSURANCE RISK THEORY (Problems) 1 Counting random variables 1. (Lack of memory property) Let X be a geometric distributed random variable with parameter p (, 1), (X Ge (p)). Show that for all n, m =,
Conditional expectation
A Conditional expectation A.1 Review of conditional densities, expectations We start with the continuous case. This is sections 6.6 and 6.8 in the book. Let X, Y be continuous random variables. We defined
WHERE DOES THE 10% CONDITION COME FROM?
1 WHERE DOES THE 10% CONDITION COME FROM? The text has mentioned The 10% Condition (at least) twice so far: p. 407 Bernoulli trials must be independent. If that assumption is violated, it is still okay
What is Statistics? Lecture 1. Introduction and probability review. Idea of parametric inference
0. 1. Introduction and probability review 1.1. What is Statistics? What is Statistics? Lecture 1. Introduction and probability review There are many definitions: I will use A set of principle and procedures
Probability Theory. Florian Herzog. A random variable is neither random nor variable. Gian-Carlo Rota, M.I.T..
Probability Theory A random variable is neither random nor variable. Gian-Carlo Rota, M.I.T.. Florian Herzog 2013 Probability space Probability space A probability space W is a unique triple W = {Ω, F,
3. Continuous Random Variables
3. Continuous Random Variables A continuous random variable is one which can take any value in an interval (or union of intervals) The values that can be taken by such a variable cannot be listed. Such
University of California, Berkeley, Statistics 134: Concepts of Probability
University of California, Berkeley, Statistics 134: Concepts of Probability Michael Lugo, Spring 211 Exam 2 solutions 1. A fair twenty-sided die has its faces labeled 1, 2, 3,..., 2. The die is rolled
Notes on Continuous Random Variables
Notes on Continuous Random Variables Continuous random variables are random quantities that are measured on a continuous scale. They can usually take on any value over some interval, which distinguishes
Probability for Estimation (review)
Probability for Estimation (review) In general, we want to develop an estimator for systems of the form: x = f x, u + η(t); y = h x + ω(t); ggggg y, ffff x We will primarily focus on discrete time linear
Topic 8: The Expected Value
Topic 8: September 27 and 29, 2 Among the simplest summary of quantitative data is the sample mean. Given a random variable, the corresponding concept is given a variety of names, the distributional mean,
Examples: Joint Densities and Joint Mass Functions Example 1: X and Y are jointly continuous with joint pdf
AMS 3 Joe Mitchell Eamples: Joint Densities and Joint Mass Functions Eample : X and Y are jointl continuous with joint pdf f(,) { c 2 + 3 if, 2, otherwise. (a). Find c. (b). Find P(X + Y ). (c). Find marginal
Probabilities and Random Variables
Probabilities and Random Variables This is an elementary overview of the basic concepts of probability theory. 1 The Probability Space The purpose of probability theory is to model random experiments so
4: Probability. What is probability? Random variables (RVs)
4: Probability b binomial µ expected value [parameter] n number of trials [parameter] N normal p probability of success [parameter] pdf probability density function pmf probability mass function RV random