Joint Probability Distributions and Random Samples. Week 5, 2011 Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage


 Dustin Maxwell
 2 years ago
 Views:
Transcription
1 5 Joint Probability Distributions and Random Samples Week 5, 2011 Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage
2 Two Discrete Random Variables The probability mass function (pmf) of a single discrete rv X specifies how much probability mass is placed on each possible X value. The joint pmf of two discrete rv s X and Y describes how much probability mass is placed on each possible pair of values (x, y): p (x, y) = P(X = x and Y = y) 2
3 Two Discrete Random Variables Always: p (x, y) 0 p (x, y) = 1 and So what are events now? A can be any set consisting of pairs of (x, y) values (e.g., A = {(x, y): x + y = 5} or {(x, y): max (x, y) 3}). Then the probability P[(X, Y) A] is obtained by summing the joint pmf over pairs in A: P[(X, Y) A] = p (x, y) 3
4 Example 1 An insurance agency services customers who have both a homeowner s policy and an automobile policy. For each type of policy, a deductible amount must be specified. For an automobile policy, the choices are $100 and $250, whereas for a homeowner s policy, the choices are 0, $100, and $200. Suppose an individual Bob  is selected at random from the agency s files. Let X = his deductible amount on the auto policy and Y = his deductible amount on the homeowner s policy. 4
5 Example 1 cont d What are the possible (X, Y) pairs? Suppose the joint pmf is given by the insurance company in the accompanying joint probability table: What is p(100, 100)? What is P(Y 100)? 5
6 Example 1 What is the probability that X = 100? What is the probability that X = 250? The marginal pmf of X is (focuses only on X, and doesn t care what Y is): What is the marginal distribution of Y? 6
7 Two Continuous Random Variables 7
8 Two Continuous Random Variables Let X and Y be continuous rv s. A joint probability density function f (x, y) for these two variables is a function satisfying f (x, y) 0 and Then for any set A 8
9 Two Continuous Random Variables In particular, if A is the twodimensional rectangle {(x, y): a x b, c y d}, then We can think of A as specifying a surface with height f(x, y) above each point (x, y). 9
10 Two Continuous Random Variables Then P[(X, Y) A] is the volume underneath the joint pdf and above the region A, analogous to the area under a curve in the case of a single rv. P[(X, Y ) A] = volume under density surface above A 10
11 Example 3 A bank operates both a driveup facility and a walkup window. Let X = the proportion of time the driveup facility is in use (at least one customer served or waiting to be served) & Y = the proportion of time the walkup window in use Say the manager has given us the joint pdf based on his experience: 11
12 Example 3 cont d As a good probabilist, you first verify that this is a pdf. 1) f (x, y) 0 (check) 2) Does it integrate to 1? What is the probability that neither facility is busy more than onequarter of the time? 12
13 Marginal Random Variables The marginal probability density functions of X and Y, denoted by f X (x) and f Y (y), respectively, are given by 13
14 Independent Random Variables 14
15 Independent Random Variables In many situations, information about the observed value of one of the two variables X and Y gives information about the value of the other variable. Here, P(X = 250) = 0.5 and P(X = 100) =
16 Independent Random Variables Two random variables X and Y are said to be independent if for every pair (x,y): p (x, y) = p X (x) p Y (y) when X and Y are discrete Or f (x, y) = f X (x) f Y (y) when X and Y are continuous If this is not satisfied for all (x, y), then X and Y are said to be dependent. 16
17 Example 6 What about in the insurance example? Independence of X and Y requires that every entry in the joint probability table be the product of the corresponding row and column marginal probabilities. 17
18 Conditional Distributions 18
19 Conditional Distributions Suppose X = the number of major defects in a randomly selected new automobile and Y = the number of minor defects in that same auto. If we learn that the selected car has one major defect, what now is the probability that the car will have at least one minor defect? That is, what is P(Y >= 1 X = 1)? 19
20 Conditional Distributions Let X and Y be two continuous rv s with joint pdf f (x, y) and marginal X pdf f X (x). Then for any X value x for which f X (x) > 0, the conditional probability density function of Y given that X = x is If X and Y are discrete, replacing pdf s by pmf s in this definition gives the conditional probability mass function of Y when X = x. Notice that the definition of f Y X (y x) parallels that of P(B A), the conditional probability that B will occur, given that A has occurred. 20
21 Example 1, cont Reconsider the insurance example: What is the conditional pmf of Y given that X = 10? What is the expected value of Y given that X=100? This is called the conditional expectation. 21
22 More Than Two Random Variables If X 1, X 2,..., X n are all discrete random variables, the joint pmf of the variables is the function p(x 1, x 2,..., x n ) = P(X 1 = x 1, X 2 = x 2,..., X n = x n ) If the variables are continuous, the joint pdf of X 1,..., X n is the function f (x 1, x 2,..., x n ) such that for any n intervals [a 1, b 1 ],..., [a n, b n ], we have: 22
23 How to find an expected value of a function of two variables If X and Y are jointly distributed rv s with pmf p(x, y) or pdf f (x, y), according to whether the variables are discrete or continuous, then Then the expected value of a function h(x, Y), denoted by E[h(X, Y)] is given by if X and Y are discrete if X and Y are continuous 23
24 Example You have purchased 5 tickets to a concert for you and 4 friends (one is Bob). Tickets are for seats 1 5 in one row. If the tickets are randomly distributed among you, what is the expected number of seats separating you from Bob? 24
25 Covariance 25
26 Covariance When two random variables X and Y are not independent, it is frequently of interest to assess how strongly they are related to one another. The covariance between two rv s X and Y is Cov(X, Y) = E[(X µ X )(Y µ Y )] X, Y discrete X, Y continuous 26
27 Covariance Since X µ X and Y µ Y are the deviations of the two variables from their respective mean values, the covariance is the expected product of deviations. Note that Cov(X, X) = E[(X µ X ) 2 ] = V(X). If both variables tend to deviate in the same direction (both go above their means or below their means at the same time), then the covariance will be positive. If the opposite is true, the covariance will be negative. If X and Y are not strongly related, the covariance will be near 0. 27
28 Covariance The covariance depends on both the set of possible pairs and the probabilities of those pairs. Below are examples of 3 types of covarying : (a) positive covariance; (b) negative covariance; (c) covariance near zero Figure
29 Example Back to the insurance example: are: from which µ X = Σxp X (x) = 175 and µ Y = 125. What is the covariance between X and Y? Why might the covariance not be a useful measure? 29
30 Covariance shortcut The following shortcut formula for Cov(X, Y) simplifies the computations. Proposition Cov(X, Y) = E(XY) µ X µ Y According to this formula, no intermediate subtractions are necessary; only at the end of the computation is µ X µ Y subtracted from E(XY). This is analogous to the shortcut for the variance computation we saw earlier. 30
31 Correlation 31
32 Correlation Definition The correlation coefficient of X and Y, denoted by Corr(X, Y), ρ X,Y, or just ρ, is defined by It represents a scaled covariance correlation ranges between 1 and 1. 32
33 Example In the insurance example, E(X 2 ) = 36,250, so = 36,250 (175) 2 = 5625, so σ X = 75 E(Y 2 ) = 22,500, so σ Y 2 = 6875, and σ Y = This gives 33
34 Correlation Propositions 1. Cov(aX + b, cy + d) = a c Cov (X, Y) 2. Corr(aX + b, cy + d) = sgn(ac) Corr(X, Y) 3. For any two rv s X and Y, 1 Corr(X, Y) 1 4. ρ = 1 or 1 iff Y = ax + b for some numbers a and b with a 0. 34
35 Correlation If X and Y are independent, then ρ = 0, but ρ = 0 does not imply independence. The correlation coefficient ρ is a measure of the linear relationship between X and Y, and only when the two variables are perfectly related in a linear manner will ρ be as positive or negative as it can be. A ρ less than 1 in absolute value indicates only that the relationship is not completely linear, but there may still be a very strong nonlinear relation. 35
36 Correlation Also, ρ = 0 does not imply that X and Y are independent, but only that there is a complete absence of a linear relationship. When ρ = 0, X and Y are said to be uncorrelated. Two variables could be uncorrelated yet highly dependent because there is a strong nonlinear relationship, so be careful not to conclude too much from knowing that ρ = 0. 36
37 Example Let X and Y be discrete rv s with joint pmf The points that receive positive probability mass are identified on the (x, y) coordinate system What is the correlation of X and Y? Are X and Y independent? 37
38 Interpreting Correlation A value of ρ near 1 does not necessarily imply that increasing the value of X causes Y to increase. It implies only that large X values are associated with large Y values. For example, in the population of children, vocabulary size and number of cavities are quite positively correlated, but it is certainly not true that cavities cause vocabulary to grow. Association (a high correlation) is not the same as causation. 38
39 Interpreting Correlation xkcd.com/552/ 39
ST 371 (VIII): Theory of Joint Distributions
ST 371 (VIII): Theory of Joint Distributions So far we have focused on probability distributions for single random variables. However, we are often interested in probability statements concerning two or
More informationBivariate Distributions
Chapter 4 Bivariate Distributions 4.1 Distributions of Two Random Variables In many practical cases it is desirable to take more than one measurement of a random observation: (brief examples) 1. What is
More informationThe Expected Value of X
3.3 Expected Values The Expected Value of X Definition Let X be a discrete rv with set of possible values D and pmf p(x). The expected value or mean value of X, denoted by E(X) or μ X or just μ, is 2 Example
More informationJoint Probability Distributions and Random Samples (Devore Chapter Five)
Joint Probability Distributions and Random Samples (Devore Chapter Five) 101634501 Probability and Statistics for Engineers Winter 20102011 Contents 1 Joint Probability Distributions 1 1.1 Two Discrete
More informationLesson 5 Chapter 4: Jointly Distributed Random Variables
Lesson 5 Chapter 4: Jointly Distributed Random Variables Department of Statistics The Pennsylvania State University 1 Marginal and Conditional Probability Mass Functions The Regression Function Independence
More informationRandom Variables, Expectation, Distributions
Random Variables, Expectation, Distributions CS 5960/6960: Nonparametric Methods Tom Fletcher January 21, 2009 Review Random Variables Definition A random variable is a function defined on a probability
More informationJoint Distributions. Tieming Ji. Fall 2012
Joint Distributions Tieming Ji Fall 2012 1 / 33 X : univariate random variable. (X, Y ): bivariate random variable. In this chapter, we are going to study the distributions of bivariate random variables
More informationJoint Probability Distributions and Random Samples
5 Joint Probabilit Distributions and Random Samples INTRODUCTION In Chapters 3 and 4, we developed probabilit models for a single random variable. Man problems in probabilit and statistics involve several
More informationExpectation Discrete RV  weighted average Continuous RV  use integral to take the weighted average
PHP 2510 Expectation, variance, covariance, correlation Expectation Discrete RV  weighted average Continuous RV  use integral to take the weighted average Variance Variance is the average of (X µ) 2
More information, where f(x,y) is the joint pdf of X and Y. Therefore
INDEPENDENCE, COVARIANCE AND CORRELATION M384G/374G Independence: For random variables and, the intuitive idea behind " is independent of " is that the distribution of shouldn't depend on what is. This
More informationDefinition The covariance of X and Y, denoted by cov(x, Y ) is defined by. cov(x, Y ) = E(X µ 1 )(Y µ 2 ).
Correlation Regression Bivariate Normal Suppose that X and Y are r.v. s with joint density f(x y) and suppose that the means of X and Y are respectively µ 1 µ 2 and the variances are 1 2. Definition The
More informationJointly Distributed Random Variables
Jointly Distributed Random Variables COMP 245 STATISTICS Dr N A Heard Contents 1 Jointly Distributed Random Variables 1 1.1 Definition......................................... 1 1.2 Joint cdfs..........................................
More informationProbability & Statistics Primer Gregory J. Hakim University of Washington 2 January 2009 v2.0
Probability & Statistics Primer Gregory J. Hakim University of Washington 2 January 2009 v2.0 This primer provides an overview of basic concepts and definitions in probability and statistics. We shall
More informationCovariance and Correlation
Covariance and Correlation ( c Robert J. Serfling Not for reproduction or distribution) We have seen how to summarize a databased relative frequency distribution by measures of location and spread, such
More informationJoint Exam 1/P Sample Exam 1
Joint Exam 1/P Sample Exam 1 Take this practice exam under strict exam conditions: Set a timer for 3 hours; Do not stop the timer for restroom breaks; Do not look at your notes. If you believe a question
More informationContinuous Random Variables and Probability Distributions. Stat 4570/5570 Material from Devore s book (Ed 8) Chapter 4  and Cengage
4 Continuous Random Variables and Probability Distributions Stat 4570/5570 Material from Devore s book (Ed 8) Chapter 4  and Cengage Continuous r.v. A random variable X is continuous if possible values
More informationAnNajah National University Faculty of Engineering Industrial Engineering Department. Course : Quantitative Methods (65211)
AnNajah National University Faculty of Engineering Industrial Engineering Department Course : Quantitative Methods (65211) Instructor: Eng. Tamer Haddad 2 nd Semester 2009/2010 Chapter 5 Example: Joint
More informationTopic 4: Multivariate random variables. Multiple random variables
Topic 4: Multivariate random variables Joint, marginal, and conditional pmf Joint, marginal, and conditional pdf and cdf Independence Expectation, covariance, correlation Conditional expectation Two jointly
More information4. Joint Distributions of Two Random Variables
4. Joint Distributions of Two Random Variables 4.1 Joint Distributions of Two Discrete Random Variables Suppose the discrete random variables X and Y have supports S X and S Y, respectively. The joint
More information2.8 Expected values and variance
y 1 b a 0 y 0.0 0.2 0.4 0.6 0.8 1.0 1.2 0 a b (a) Probability density function x 0 a b (b) Cumulative distribution function x Figure 2.3: The probability density function and cumulative distribution function
More informationReview Exam Suppose that number of cars that passes through a certain rural intersection is a Poisson process with an average rate of 3 per day.
Review Exam 2 This is a sample of problems that would be good practice for the exam. This is by no means a guarantee that the problems on the exam will look identical to those on the exam but it should
More informationThis is a good time to refresh your memory on doubleintegration. We will be using this skill in the upcoming lectures.
Chapter 5: JOINT PROBABILITY DISTRIBUTIONS Part 1: Sections 51.1 to 51.4 For both discrete and continuous random variables we will discuss the following... Joint Distributions (for two or more r.v. s)
More informationLecture 18 Chapter 6: Empirical Statistics
Lecture 18 Chapter 6: Empirical Statistics M. George Akritas Definition (Sample Covariance and Pearson s Correlation) Let (X 1, Y 1 ),..., (X n, Y n ) be a sample from a bivariate population. The sample
More informationCovariance and Correlation. Consider the joint probability distribution f XY (x, y).
Chapter 5: JOINT PROBABILITY DISTRIBUTIONS Part 2: Section 52 Covariance and Correlation Consider the joint probability distribution f XY (x, y). Is there a relationship between X and Y? If so, what kind?
More informationExpected Value. Let X be a discrete random variable which takes values in S X = {x 1, x 2,..., x n }
Expected Value Let X be a discrete random variable which takes values in S X = {x 1, x 2,..., x n } Expected Value or Mean of X: E(X) = n x i p(x i ) i=1 Example: Roll one die Let X be outcome of rolling
More informationExamination 110 Probability and Statistics Examination
Examination 0 Probability and Statistics Examination Sample Examination Questions The Probability and Statistics Examination consists of 5 multiplechoice test questions. The test is a threehour examination
More informationSummary of Formulas and Concepts. Descriptive Statistics (Ch. 14)
Summary of Formulas and Concepts Descriptive Statistics (Ch. 14) Definitions Population: The complete set of numerical information on a particular quantity in which an investigator is interested. We assume
More informationProbability and Statistical Methods. Chapter 4 Mathematical Expectation
Math 3 Chapter 4 Mathematical Epectation Mean of a Random Variable Definition. Let be a random variable with probability distribution f( ). The mean or epected value of is, f( ) µ = µ = E =, if is a discrete
More informationSo, using the new notation, P X,Y (0,1) =.08 This is the value which the joint probability function for X and Y takes when X=0 and Y=1.
Joint probabilit is the probabilit that the RVs & Y take values &. like the PDF of the two events, and. We will denote a joint probabilit function as P,Y (,) = P(= Y=) Marginal probabilit of is the probabilit
More informationWe have discussed the notion of probabilistic dependence above and indicated that dependence is
1 CHAPTER 7 Online Supplement Covariance and Correlation for Measuring Dependence We have discussed the notion of probabilistic dependence above and indicated that dependence is defined in terms of conditional
More informationJoint Probability Distributions and Random Samples
STAT5 Sprig 204 Lecture Notes Chapter 5 February, 204 Joit Probability Distributios ad Radom Samples 5. Joitly Distributed Radom Variables Chapter Overview Joitly distributed rv Joit mass fuctio, margial
More information6. Jointly Distributed Random Variables
6. Jointly Distributed Random Variables We are often interested in the relationship between two or more random variables. Example: A randomly chosen person may be a smoker and/or may get cancer. Definition.
More information3 Multiple Discrete Random Variables
3 Multiple Discrete Random Variables 3.1 Joint densities Suppose we have a probability space (Ω, F,P) and now we have two discrete random variables X and Y on it. They have probability mass functions f
More informationChapter 3: Discrete Random Variable and Probability Distribution. January 28, 2014
STAT511 Spring 2014 Lecture Notes 1 Chapter 3: Discrete Random Variable and Probability Distribution January 28, 2014 3 Discrete Random Variables Chapter Overview Random Variable (r.v. Definition Discrete
More information3.6: General Hypothesis Tests
3.6: General Hypothesis Tests The χ 2 goodness of fit tests which we introduced in the previous section were an example of a hypothesis test. In this section we now consider hypothesis tests more generally.
More informationEngineering 323 Beautiful Homework Set 7 1 of 7 Morelli Problem The cdf of checkout duration X as described in Exercise 1 is
Engineering 33 Beautiful Homework Set 7 of 7 Morelli Problem.. The cdf of checkout duration X as described in Eercise is F( )
More informationL10: Probability, statistics, and estimation theory
L10: Probability, statistics, and estimation theory Review of probability theory Bayes theorem Statistics and the Normal distribution Least Squares Error estimation Maximum Likelihood estimation Bayesian
More informationChapter 5: Joint Probability Distributions. Chapter Learning Objectives. The Joint Probability Distribution for a Pair of Discrete Random
Chapter 5: Joint Probability Distributions 51 Two or More Random Variables 51.1 Joint Probability Distributions 51.2 Marginal Probability Distributions 51.3 Conditional Probability Distributions 51.4
More information10 BIVARIATE DISTRIBUTIONS
BIVARIATE DISTRIBUTIONS After some discussion of the Normal distribution, consideration is given to handling two continuous random variables. The Normal Distribution The probability density function f(x)
More informationJoint Distribution and Correlation
Joint Distribution and Correlation Michael Ash Lecture 3 Reminder: Start working on the Problem Set Mean and Variance of Linear Functions of an R.V. Linear Function of an R.V. Y = a + bx What are the properties
More informationSolution to HW  1. Problem 1. [Points = 3] In September, Chapel Hill s daily high temperature has a mean
Problem 1. [Points = 3] In September, Chapel Hill s daily high temperature has a mean of 81 degree F and a standard deviation of 10 degree F. What is the mean, standard deviation and variance in terms
More informationChapters 5. Multivariate Probability Distributions
Chapters 5. Multivariate Probability Distributions Random vectors are collection of random variables defined on the same sample space. Whenever a collection of random variables are mentioned, they are
More informationWorked examples Multiple Random Variables
Worked eamples Multiple Random Variables Eample Let X and Y be random variables that take on values from the set,, } (a) Find a joint probability mass assignment for which X and Y are independent, and
More information5. Conditional Expected Value
Virtual Laboratories > 3. Expected Value > 1 2 3 4 5 6 5. Conditional Expected Value As usual, our starting point is a random experiment with probability measure P on a sample space Ω. Suppose that X is
More informationPOL 571: Expectation and Functions of Random Variables
POL 571: Expectation and Functions of Random Variables Kosuke Imai Department of Politics, Princeton University March 10, 2006 1 Expectation and Independence To gain further insights about the behavior
More informationFor a partition B 1,..., B n, where B i B j = for i. A = (A B 1 ) (A B 2 ),..., (A B n ) and thus. P (A) = P (A B i ) = P (A B i )P (B i )
Probability Review 15.075 Cynthia Rudin A probability space, defined by Kolmogorov (19031987) consists of: A set of outcomes S, e.g., for the roll of a die, S = {1, 2, 3, 4, 5, 6}, 1 1 2 1 6 for the roll
More informationMath 576: Quantitative Risk Management
Math 576: Quantitative Risk Management Haijun Li lih@math.wsu.edu Department of Mathematics Washington State University Week 4 Haijun Li Math 576: Quantitative Risk Management Week 4 1 / 22 Outline 1 Basics
More informationRandom Variables and Their Expected Values
Discrete and Continuous Random Variables The Probability Mass Function The (Cumulative) Distribution Function Discrete and Continuous Random Variables The Probability Mass Function The (Cumulative) Distribution
More informationDiscrete and Continuous Random Variables. Summer 2003
Discrete and Continuous Random Variables Summer 003 Random Variables A random variable is a rule that assigns a numerical value to each possible outcome of a probabilistic experiment. We denote a random
More informationContinuous Random Variables
Continuous Random Variables COMP 245 STATISTICS Dr N A Heard Contents 1 Continuous Random Variables 2 11 Introduction 2 12 Probability Density Functions 3 13 Transformations 5 2 Mean, Variance and Quantiles
More informationCalculate the holding period return for this investment. It is approximately
1. An investor purchases 100 shares of XYZ at the beginning of the year for $35. The stock pays a cash dividend of $3 per share. The price of the stock at the time of the dividend is $30. The dividend
More informationECE302 Spring 2006 HW7 Solutions March 11, 2006 1
ECE32 Spring 26 HW7 Solutions March, 26 Solutions to HW7 Note: Most of these solutions were generated by R. D. Yates and D. J. Goodman, the authors of our textbook. I have added comments in italics where
More informationLecture 14: Correlation and Autocorrelation Steven Skiena. skiena
Lecture 14: Correlation and Autocorrelation Steven Skiena Department of Computer Science State University of New York Stony Brook, NY 11794 4400 http://www.cs.sunysb.edu/ skiena Overuse of Color, Dimensionality,
More informationNotes on the second moment method, Erdős multiplication tables
Notes on the second moment method, Erdős multiplication tables January 25, 20 Erdős multiplication table theorem Suppose we form the N N multiplication table, containing all the N 2 products ab, where
More informationRandom Variables. Consider a probability model (Ω, P ). Discrete Random Variables Chs. 2, 3, 4. Definition. A random variable is a function
Rom Variables Discrete Rom Variables Chs.,, 4 Rom Variables Probability Mass Functions Expectation: The Mean Variance Special Distributions Hypergeometric Binomial Poisson Joint Distributions Independence
More informationEE 322: Probabilistic Methods for Electrical Engineers. Zhengdao Wang Department of Electrical and Computer Engineering Iowa State University
EE 322: Probabilistic Methods for Electrical Engineers Zhengdao Wang Department of Electrical and Computer Engineering Iowa State University Discrete Random Variables 1 Introduction to Random Variables
More informationST 371 (IV): Discrete Random Variables
ST 371 (IV): Discrete Random Variables 1 Random Variables A random variable (rv) is a function that is defined on the sample space of the experiment and that assigns a numerical variable to each possible
More informationSimple Linear Regression Chapter 11
Simple Linear Regression Chapter 11 Rationale Frequently decisionmaking situations require modeling of relationships among business variables. For instance, the amount of sale of a product may be related
More informationThe Bivariate Normal Distribution
The Bivariate Normal Distribution This is Section 4.7 of the st edition (2002) of the book Introduction to Probability, by D. P. Bertsekas and J. N. Tsitsiklis. The material in this section was not included
More informationNotes 11 Autumn 2005
MAS 08 Probabilit I Notes Autumn 005 Two discrete random variables If X and Y are discrete random variables defined on the same sample space, then events such as X = and Y = are well defined. The joint
More informationMATH 201. Final ANSWERS August 12, 2016
MATH 01 Final ANSWERS August 1, 016 Part A 1. 17 points) A bag contains three different types of dice: four 6sided dice, five 8sided dice, and six 0sided dice. A die is drawn from the bag and then rolled.
More informationDepartment of Mathematics, Indian Institute of Technology, Kharagpur Assignment 23, Probability and Statistics, March 2015. Due:March 25, 2015.
Department of Mathematics, Indian Institute of Technology, Kharagpur Assignment 3, Probability and Statistics, March 05. Due:March 5, 05.. Show that the function 0 for x < x+ F (x) = 4 for x < for x
More informationFEGYVERNEKI SÁNDOR, PROBABILITY THEORY AND MATHEmATICAL
FEGYVERNEKI SÁNDOR, PROBABILITY THEORY AND MATHEmATICAL STATIsTICs 4 IV. RANDOm VECTORs 1. JOINTLY DIsTRIBUTED RANDOm VARIABLEs If are two rom variables defined on the same sample space we define the joint
More informationChapter 5: Multivariate Distributions
Chapter 5: Multivariate Distributions Professor Ron Fricker Naval Postgraduate School Monterey, California 3/15/15 Reading Assignment: Sections 5.1 5.12 1 Goals for this Chapter Bivariate and multivariate
More informationTopic 8 The Expected Value
Topic 8 The Expected Value Functions of Random Variables 1 / 12 Outline Names for Eg(X ) Variance and Standard Deviation Independence Covariance and Correlation 2 / 12 Names for Eg(X ) If g(x) = x, then
More informationP (x) 0. Discrete random variables Expected value. The expected value, mean or average of a random variable x is: xp (x) = v i P (v i )
Discrete random variables Probability mass function Given a discrete random variable X taking values in X = {v 1,..., v m }, its probability mass function P : X [0, 1] is defined as: P (v i ) = Pr[X =
More information), 35% use extra unleaded gas ( A
. At a certain gas station, 4% of the customers use regular unleaded gas ( A ), % use extra unleaded gas ( A ), and % use premium unleaded gas ( A ). Of those customers using regular gas, onl % fill their
More informationMT426 Notebook 3 Fall 2012 prepared by Professor Jenny Baglivo. 3 MT426 Notebook 3 3. 3.1 Definitions... 3. 3.2 Joint Discrete Distributions...
MT426 Notebook 3 Fall 2012 prepared by Professor Jenny Baglivo c Copyright 20042012 by Jenny A. Baglivo. All Rights Reserved. Contents 3 MT426 Notebook 3 3 3.1 Definitions............................................
More informationMULTIVARIATE PROBABILITY DISTRIBUTIONS
MULTIVARIATE PROBABILITY DISTRIBUTIONS. PRELIMINARIES.. Example. Consider an experiment that consists of tossing a die and a coin at the same time. We can consider a number of random variables defined
More informationContinuous Random Variables. and Probability Distributions. Continuous Random Variables and Probability Distributions ( ) ( ) Chapter 4 4.
UCLA STAT 11 A Applied Probability & Statistics for Engineers Instructor: Ivo Dinov, Asst. Prof. In Statistics and Neurology Teaching Assistant: Neda Farzinnia, UCLA Statistics University of California,
More informationAdvanced High School Statistics. Preliminary Edition
Chapter 2 Summarizing Data After collecting data, the next stage in the investigative process is to summarize the data. Graphical displays allow us to visualize and better understand the important features
More informationConditional expectation
A Conditional expectation A.1 Review of conditional densities, expectations We start with the continuous case. This is sections 6.6 and 6.8 in the book. Let X, Y be continuous random variables. We defined
More informationSome notes on the Poisson distribution
Some notes on the Poisson distribution Ernie Croot October 2, 2008 1 Introduction The Poisson distribution is one of the most important that we will encounter in this course it is right up there with the
More informationM 384G/374G/CAM384T CONDITIONAL AND MARGINAL MEANS AND VARIANCES. I. Question: How are conditional means E(Y X) and marginal means E(Y) related?
1 M 3G/37G/CAM3T CONDITIONAL AND MARGINAL MEANS AND VARIANCES I. Question: How are conditional means E(Y X) and marginal means E(Y) related? Simple example: Population consisting of men, n women. Y = height
More informationDefinition: Suppose that two random variables, either continuous or discrete, X and Y have joint density
HW MATH 461/561 Lecture Notes 15 1 Definition: Suppose that two random variables, either continuous or discrete, X and Y have joint density and marginal densities f(x, y), (x, y) Λ X,Y f X (x), x Λ X,
More informationChapter 3 Joint Distributions
Chapter 3 Joint Distributions 3.6 Functions of Jointly Distributed Random Variables Discrete Random Variables: Let f(x, y) denote the joint pdf of random variables X and Y with A denoting the twodimensional
More informationwhere b is the slope of the line and a is the intercept i.e. where the line cuts the y axis.
Least Squares Introduction We have mentioned that one should not always conclude that because two variables are correlated that one variable is causing the other to behave a certain way. However, sometimes
More informationSections 2.11 and 5.8
Sections 211 and 58 Timothy Hanson Department of Statistics, University of South Carolina Stat 704: Data Analysis I 1/25 Gesell data Let X be the age in in months a child speaks his/her first word and
More informationRandom variables, probability distributions, binomial random variable
Week 4 lecture notes. WEEK 4 page 1 Random variables, probability distributions, binomial random variable Eample 1 : Consider the eperiment of flipping a fair coin three times. The number of tails that
More information3. Continuous Random Variables
3. Continuous Random Variables A continuous random variable is one which can take any value in an interval (or union of intervals) The values that can be taken by such a variable cannot be listed. Such
More informationData. ECON 251 Research Methods. 1. Data and Descriptive Statistics (Review) CrossSectional and TimeSeries Data. Population vs.
ECO 51 Research Methods 1. Data and Descriptive Statistics (Review) Data A variable  a characteristic of population or sample that is of interest for us. Data  the actual values of variables Quantitative
More informationStatistical Significance and Bivariate Tests
Statistical Significance and Bivariate Tests BUS 735: Business Decision Making and Research 1 1.1 Goals Goals Specific goals: Refamiliarize ourselves with basic statistics ideas: sampling distributions,
More informationRandom Variables. Chapter 2. Random Variables 1
Random Variables Chapter 2 Random Variables 1 Roulette and Random Variables A Roulette wheel has 38 pockets. 18 of them are red and 18 are black; these are numbered from 1 to 36. The two remaining pockets
More informationINTRODUCTION TO PROBABILITY AND STATISTICS
INTRODUCTION TO PROBABILITY AND STATISTICS Conditional probability and independent events.. A fair die is tossed twice. Find the probability of getting a 4, 5, or 6 on the first toss and a,,, or 4 on the
More information3.4. The Binomial Probability Distribution. Copyright Cengage Learning. All rights reserved.
3.4 The Binomial Probability Distribution Copyright Cengage Learning. All rights reserved. The Binomial Probability Distribution There are many experiments that conform either exactly or approximately
More informationChapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS
Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS Part 4: Geometric Distribution Negative Binomial Distribution Hypergeometric Distribution Sections 37, 38 The remaining discrete random
More informationReview. Lecture 3: Probability Distributions. Poisson Distribution. May 8, 2012 GENOME 560, Spring Su In Lee, CSE & GS
Lecture 3: Probability Distributions May 8, 202 GENOME 560, Spring 202 Su In Lee, CSE & GS suinlee@uw.edu Review Random variables Discrete: Probability mass function (pmf) Continuous: Probability density
More informationRandom variables and their properties
Random variables and their properties As we have discussed in class, when observing a random process, we are faced with uncertainty. It is easier to study that uncertainty if we make things numerical.
More informationMath 431 An Introduction to Probability. Final Exam Solutions
Math 43 An Introduction to Probability Final Eam Solutions. A continuous random variable X has cdf a for 0, F () = for 0 <
More informationChapter 11: Two Variable Regression Analysis
Department of Mathematics Izmir University of Economics Week 1415 20142015 In this chapter, we will focus on linear models and extend our analysis to relationships between variables, the definitions
More informationEE 302 Division 1. Homework 5 Solutions.
EE 32 Division. Homework 5 Solutions. Problem. A fair foursided die (with faces labeled,, 2, 3) is thrown once to determine how many times a fair coin is to be flipped: if N is the number that results
More informationThe Scalar Algebra of Means, Covariances, and Correlations
3 The Scalar Algebra of Means, Covariances, and Correlations In this chapter, we review the definitions of some key statistical concepts: means, covariances, and correlations. We show how the means, variances,
More informationStatistical Intervals. Chapter 7 Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage
7 Statistical Intervals Chapter 7 Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage Confidence Intervals The CLT tells us that as the sample size n increases, the sample mean X is close to
More informationMath/Stat 3601: Probability and Statistics, Washington State University
Math/Stat 3601: Probability and Statistics, Washington State University Haijun Li lih@math.wsu.edu Department of Mathematics Washington State University Week 3 Haijun Li Math/Stat 3601: Probability and
More informationContinuous Distributions
MAT 2379 3X (Summer 2012) Continuous Distributions Up to now we have been working with discrete random variables whose R X is finite or countable. However we will have to allow for variables that can take
More informationDiversification and the RiskReturn TradeOff
Diversification and the RiskReturn TradeOff Dr. Patrick Toche References : Zvi Bodie, Alex Kane, Alan Marcus, Investments, McGrawHill/Irwin. The relevant chapters of this textbook will be followed closely
More informationStatistical modeling, Simulation and Time series
Statistical modeling, Simulation and Time series Leif Nilsson Department of Mathematics and Mathematical Statistics Statistical models Simulation (Bootstrap) Time series (Correlated data) 1 Statistical
More informationHomework Zero. Max H. Farrell Chicago Booth BUS41100 Applied Regression Analysis. Complete before the first class, do not turn in
Homework Zero Max H. Farrell Chicago Booth BUS41100 Applied Regression Analysis Complete before the first class, do not turn in This homework is intended as a self test of your knowledge of the basic statistical
More informationMath 141. Lecture 5: Expected Value. Albyn Jones 1. jones/courses/ Library 304. Albyn Jones Math 141
Math 141 Lecture 5: Expected Value Albyn Jones 1 1 Library 304 jones@reed.edu www.people.reed.edu/ jones/courses/141 History The early history of probability theory is intimately related to questions arising
More informationThe Delta Method and Applications
Chapter 5 The Delta Method and Applications 5.1 Linear approximations of functions In the simplest form of the central limit theorem, Theorem 4.18, we consider a sequence X 1, X,... of independent and
More information