Jointly Distributed Random Variables
|
|
- Natalie Eunice Maxwell
- 7 years ago
- Views:
Transcription
1 Jointly Distributed Random Variables COMP 245 STATISTICS Dr N A Heard Contents 1 Jointly Distributed Random Variables Definition Joint cdfs Joint Probability Mass Functions Joint Probability Density Functions Independence and Expectation Independence Expectation Conditional Expectation Covariance and Correlation Examples 6 1 Jointly Distributed Random Variables 1.1 Definition Joint Probability Distribution Suppose we have two random variables X and Y defined on a sample space S with probability measure P(E), E S. Jointly, they form a map (X, Y) : S R 2 where s (X(s), Y(s)). From before, we know to define the marginal probability distributions P X and P Y by, for example, P X (B) = P(X 1 (B)), B R. We now define the joint probability distribution P XY by P XY (B X, B Y ) = P{X 1 (B X ) Y 1 (B Y )}, B X, B Y R. So P XY (B X, B Y ), the probability that X B X and Y B Y, is given by the probability P of the set of all points in the sample space that get mapped both into B X by X and into B Y by Y. More generally, for a single region B XY R 2, find the collection of sample space elements and define S BXY = {s S (X(s), Y(s)) B XY } P XY (B XY ) = P(S BXY ). 1
2 1.2 Joint cdfs Joint Cumulative Distribution Function We thus define the joint cumulative distribution function as F XY (x, y) = P XY ((, x], (, y]), x, y R. It is easy to check that the marginal cdfs for X and Y can be recovered by and that the two definitions will agree. F X (x) = F XY (x, ), x R, F Y (y) = F XY (, y), y R, Properties of a Joint cdf For F XY to be a valid cdf, we need to make sure the following conditions hold F XY (x, y) 1, x, y R; 2. Monotonicity: x 1, x 2, y 1, y 2 R, x 1 < x 2 F XY (x 1, y 1 ) F XY (x 2, y 1 ) and y 1 < y 2 F XY (x 1, y 1 ) F XY (x 1, y 2 ); 3. x, y R, F XY (x, ) = 0, F XY (, y) = 0 and F XY (, ) = 1. Interval Probabilities Suppose we are interested in whether the random variable pair (X, Y) lie in the interval cross product (x 1, x 2 ] (y 1, y 2 ]; that is, if x 1 < X x 2 and y 1 < Y y 2. Y y 2 y 1 x 1 x 2 X First note that P XY ((x 1, x 2 ], (, y]) = F XY (x 2, y) F XY (x 1, y). It is then easy to see that P XY ((x 1, x 2 ], (y 1, y 2 ]) is given by F XY (x 2, y 2 ) F XY (x 1, y 2 ) F XY (x 2, y 1 ) + F XY (x 1, y 1 ). 1.3 Joint Probability Mass Functions If X and Y are both discrete random variables, then we can define the joint probability mass function as p XY (x, y) = P XY ({x}, {y}), x, y R. We can recover the marginal pmfs p X and p Y since, by the law of total probability, x, y R p X (x) = y p XY (x, y), p Y (y) = x p XY (x, y). 2
3 Properties of a Joint pmf For p XY to be a valid pmf, we need to make sure the following conditions hold p XY (x, y) 1, x, y R; 2. p XY (x, y) = 1. y x 1.4 Joint Probability Density Functions On the other hand, if f XY : R R R s.t. P XY (B XY ) = f XY (x, y)dxdy, B XY R R, (x,y) B XY then we say X and Y are jointly continuous and we refer to f XY as the joint probability density function of X and Y. In this case, we have F XY (x, y) = y t= x s= f XY (s, t)dsdt, x, y R, By the Fundamental Theorem of Calculus we can identify the joint pdf as f XY (x, y) = 2 x y F XY(x, y). Furthermore, we can recover the marginal densities f X and f Y : f X (x) = d dx F X(x) = d dx F XY(x, ) = d x f XY (s, y)dsdy. dx s= By the Fundamental Theorem of Calculus, and through a symmetric argument for Y, we thus get f X (x) = f XY (x, y)dy, f Y (y) = x= f XY (x, y)dx. Properties of a Joint pdf For f XY to be a valid pdf, we need to make sure the following conditions hold. 1. f XY (x, y) 0, x, y R; 2. x= f XY (x, y)dxdy = 1. 3
4 2 Independence and Expectation 2.1 Independence Independence of Random Variables Two random variables X and Y are independent iff B X, B Y R., P XY (B X, B Y ) = P X (B X )P Y (B Y ). More specifically, two discrete random variables X and Y are independent iff p XY (x, y) = p X (x)p Y (y), x, y R; and two continuous random variables X and Y are independent iff f XY (x, y) = f X (x) f Y (y), x, y R. Conditional Distributions For two r.v.s X, Y we define the conditional probability distribution P Y X by P Y X (B Y B X ) = P XY(B X, B Y ), B X, B P X (B X ) Y R, P X (B X ) = 0. This is the revised probability of Y falling inside B Y given that we now know X B X. Then we have X and Y are independent P Y X (B Y B X ) = P Y (B Y ), B X, B Y R. For discrete r.v.s X, Y we define the conditional probability mass function p Y X by p Y X (y x) = p XY(x, y), x, y R, p X (x) = 0, p X (x) and for continuous r.v.s X, Y we define the conditional probability density function f Y X by f Y X (y x) = f XY(x, y), x, y R. f X (x) [Justification: P(Y y X [x, x + dx)) = P XY([x, x + dx), (, y]) P X ([x, x + dx)) F(y, x + dx) F(y, x) = F(x + dx) F(x) {F(y, x + dx) F(y, x)}/dx = {F(x + dx) F(x)}/dx d{f(y, x + dx) F(y, x)}/dxdy = f (y X [x, x + dx)) = {F(x + dx) F(x)}/dx f (y, x) = f (y x) = lim f (y X [x, x + dx)) = dx >0 f (x).] R. In either case, X and Y are independent p Y X (y x) = p Y (y) or f Y X (y x) = f Y (y), x, y 4
5 2.2 Expectation E{g(X, Y)} Suppose we have a bivariate function of interest of the random variables X and Y, g : R R R. If X and Y are discrete, we define E{g(X, Y)} by E XY {g(x, Y)} = y g(x, y)p XY (x, y). x If X and Y are jointly continuous, we define E{g(X, Y)} by E XY {g(x, Y)} = x= g(x, y) f XY (x, y)dxdy. Immediately from these definitions we have the following: If g(x, Y) = g 1 (X) + g 2 (Y), E XY {g 1 (X) + g 2 (Y)} = E X {g 1 (X)} + E Y {g 2 (Y)}. If g(x, Y) = g 1 (X)g 2 (Y) and X and Y are independent, E XY {g 1 (X)g 2 (Y)} = E X {g 1 (X)}E Y {g 2 (Y)}. In particular, considering g(x, Y) = XY for independent X, Y we have E XY (XY) = E X (X)E Y (Y). 2.3 Conditional Expectation E Y X (Y X = x) In general E XY (XY) = E X (X)E Y (Y). Suppose X and Y are discrete r.v.s with joint pmf p(x, y). If we are given the value x of the r.v. X, our revised pmf for Y is the conditional pmf p(y x), for y R. The conditional expectation of Y given X = x is therefore Similarly, if X and Y were continuous, E Y X (Y X = x) = y p(y x). y E Y X (Y X = x) = y f (y x)dy. In either case, the conditional expectation is a function of x but not the unknown Y. 5
6 2.4 Covariance and Correlation Covariance For a single variable X we considered the expectation of g(x) = (X µ X )(X µ X ), called the variance and denoted σ 2 X. The bivariate extension of this is the expectation of g(x, Y) = (X µ X )(Y µ Y ). We define the covariance of X and Y by σ XY = Cov(X, Y) = E XY [(X µ X )(Y µ Y )]. Correlation Covariance measures how the random variables move in tandem with one another, and so is closely related to the idea of correlation. We define the correlation of X and Y by ρ XY = Cor(X, Y) = σ XY σ X σ Y. Unlike the covariance, the correlation is invariant to the scale of the r.v.s X and Y. It is easily shown that if X and Y are independent random variables, then σ XY = ρ XY = 0. 3 Examples Example 1 Suppose that the lifetime, X, and brightness, Y of a light bulb are modelled as continuous random variables. Let their joint pdf be given by Are lifetime and brightness independent? The marginal pdf for X is f (x) = f (x, y) = λ 1 λ 2 e λ 1x λ 2 y, x, y > 0. = λ 1 e λ 1x. f (x, y)dy = y=0 λ 1 λ 2 e λ 1x λ 2 y dy Similarly f (y) = λ 2 e λ 2y. Hence f (x, y) = f (x) f (y) and X and Y are independent. Example 2 Suppose continuous r.v.s (X, Y) R 2 have joint pdf { 1 f (x, y) = π, x2 + y 2 1 0, otherwise. Determine the marginal pdfs for X and Y. Well x 2 + y 2 1 y 1 x 2. So 1 x 2 1 f (x) = f (x, y)dy = y= 1 x 2 π dy = 2 1 x π 2. Similarly f (y) = 2 1 y π 2. Hence f (x, y) = f (x) f (y) and X and Y are not independent. 6
Section 6.1 Joint Distribution Functions
Section 6.1 Joint Distribution Functions We often care about more than one random variable at a time. DEFINITION: For any two random variables X and Y the joint cumulative probability distribution function
More informationDefinition: Suppose that two random variables, either continuous or discrete, X and Y have joint density
HW MATH 461/561 Lecture Notes 15 1 Definition: Suppose that two random variables, either continuous or discrete, X and Y have joint density and marginal densities f(x, y), (x, y) Λ X,Y f X (x), x Λ X,
More informationM2S1 Lecture Notes. G. A. Young http://www2.imperial.ac.uk/ ayoung
M2S1 Lecture Notes G. A. Young http://www2.imperial.ac.uk/ ayoung September 2011 ii Contents 1 DEFINITIONS, TERMINOLOGY, NOTATION 1 1.1 EVENTS AND THE SAMPLE SPACE......................... 1 1.1.1 OPERATIONS
More informationLecture Notes 1. Brief Review of Basic Probability
Probability Review Lecture Notes Brief Review of Basic Probability I assume you know basic probability. Chapters -3 are a review. I will assume you have read and understood Chapters -3. Here is a very
More informationSummary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)
Summary of Formulas and Concepts Descriptive Statistics (Ch. 1-4) Definitions Population: The complete set of numerical information on a particular quantity in which an investigator is interested. We assume
More informationLecture 6: Discrete & Continuous Probability and Random Variables
Lecture 6: Discrete & Continuous Probability and Random Variables D. Alex Hughes Math Camp September 17, 2015 D. Alex Hughes (Math Camp) Lecture 6: Discrete & Continuous Probability and Random September
More informationMath 432 HW 2.5 Solutions
Math 432 HW 2.5 Solutions Assigned: 1-10, 12, 13, and 14. Selected for Grading: 1 (for five points), 6 (also for five), 9, 12 Solutions: 1. (2y 3 + 2y 2 ) dx + (3y 2 x + 2xy) dy = 0. M/ y = 6y 2 + 4y N/
More informationSTA 256: Statistics and Probability I
Al Nosedal. University of Toronto. Fall 2014 1 2 3 4 5 My momma always said: Life was like a box of chocolates. You never know what you re gonna get. Forrest Gump. Experiment, outcome, sample space, and
More information5. Continuous Random Variables
5. Continuous Random Variables Continuous random variables can take any value in an interval. They are used to model physical characteristics such as time, length, position, etc. Examples (i) Let X be
More informationChapter 4 - Lecture 1 Probability Density Functions and Cumul. Distribution Functions
Chapter 4 - Lecture 1 Probability Density Functions and Cumulative Distribution Functions October 21st, 2009 Review Probability distribution function Useful results Relationship between the pdf and the
More informationFeb 28 Homework Solutions Math 151, Winter 2012. Chapter 6 Problems (pages 287-291)
Feb 8 Homework Solutions Math 5, Winter Chapter 6 Problems (pages 87-9) Problem 6 bin of 5 transistors is known to contain that are defective. The transistors are to be tested, one at a time, until the
More informationPrinciple of Data Reduction
Chapter 6 Principle of Data Reduction 6.1 Introduction An experimenter uses the information in a sample X 1,..., X n to make inferences about an unknown parameter θ. If the sample size n is large, then
More informationSection 5.1 Continuous Random Variables: Introduction
Section 5. Continuous Random Variables: Introduction Not all random variables are discrete. For example:. Waiting times for anything (train, arrival of customer, production of mrna molecule from gene,
More informationJoint Exam 1/P Sample Exam 1
Joint Exam 1/P Sample Exam 1 Take this practice exam under strict exam conditions: Set a timer for 3 hours; Do not stop the timer for restroom breaks; Do not look at your notes. If you believe a question
More informationMULTIVARIATE PROBABILITY DISTRIBUTIONS
MULTIVARIATE PROBABILITY DISTRIBUTIONS. PRELIMINARIES.. Example. Consider an experiment that consists of tossing a die and a coin at the same time. We can consider a number of random variables defined
More informationOverview of Monte Carlo Simulation, Probability Review and Introduction to Matlab
Monte Carlo Simulation: IEOR E4703 Fall 2004 c 2004 by Martin Haugh Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab 1 Overview of Monte Carlo Simulation 1.1 Why use simulation?
More informationST 371 (IV): Discrete Random Variables
ST 371 (IV): Discrete Random Variables 1 Random Variables A random variable (rv) is a function that is defined on the sample space of the experiment and that assigns a numerical variable to each possible
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES Contents 1. Random variables and measurable functions 2. Cumulative distribution functions 3. Discrete
More informationRandom Variables. Chapter 2. Random Variables 1
Random Variables Chapter 2 Random Variables 1 Roulette and Random Variables A Roulette wheel has 38 pockets. 18 of them are red and 18 are black; these are numbered from 1 to 36. The two remaining pockets
More informationLecture 7: Continuous Random Variables
Lecture 7: Continuous Random Variables 21 September 2005 1 Our First Continuous Random Variable The back of the lecture hall is roughly 10 meters across. Suppose it were exactly 10 meters, and consider
More informationRandom variables P(X = 3) = P(X = 3) = 1 8, P(X = 1) = P(X = 1) = 3 8.
Random variables Remark on Notations 1. When X is a number chosen uniformly from a data set, What I call P(X = k) is called Freq[k, X] in the courseware. 2. When X is a random variable, what I call F ()
More informationMath 370/408, Spring 2008 Prof. A.J. Hildebrand. Actuarial Exam Practice Problem Set 3 Solutions
Math 37/48, Spring 28 Prof. A.J. Hildebrand Actuarial Exam Practice Problem Set 3 Solutions About this problem set: These are problems from Course /P actuarial exams that I have collected over the years,
More informationMath 431 An Introduction to Probability. Final Exam Solutions
Math 43 An Introduction to Probability Final Eam Solutions. A continuous random variable X has cdf a for 0, F () = for 0 <
More informationStatistics 100A Homework 7 Solutions
Chapter 6 Statistics A Homework 7 Solutions Ryan Rosario. A television store owner figures that 45 percent of the customers entering his store will purchase an ordinary television set, 5 percent will purchase
More informationDepartment of Mathematics, Indian Institute of Technology, Kharagpur Assignment 2-3, Probability and Statistics, March 2015. Due:-March 25, 2015.
Department of Mathematics, Indian Institute of Technology, Kharagpur Assignment -3, Probability and Statistics, March 05. Due:-March 5, 05.. Show that the function 0 for x < x+ F (x) = 4 for x < for x
More informationIntroduction to Probability
Introduction to Probability EE 179, Lecture 15, Handout #24 Probability theory gives a mathematical characterization for experiments with random outcomes. coin toss life of lightbulb binary data sequence
More informationECE302 Spring 2006 HW5 Solutions February 21, 2006 1
ECE3 Spring 6 HW5 Solutions February 1, 6 1 Solutions to HW5 Note: Most of these solutions were generated by R. D. Yates and D. J. Goodman, the authors of our textbook. I have added comments in italics
More informationFor a partition B 1,..., B n, where B i B j = for i. A = (A B 1 ) (A B 2 ),..., (A B n ) and thus. P (A) = P (A B i ) = P (A B i )P (B i )
Probability Review 15.075 Cynthia Rudin A probability space, defined by Kolmogorov (1903-1987) consists of: A set of outcomes S, e.g., for the roll of a die, S = {1, 2, 3, 4, 5, 6}, 1 1 2 1 6 for the roll
More informationLecture 8. Confidence intervals and the central limit theorem
Lecture 8. Confidence intervals and the central limit theorem Mathematical Statistics and Discrete Mathematics November 25th, 2015 1 / 15 Central limit theorem Let X 1, X 2,... X n be a random sample of
More informationData Modeling & Analysis Techniques. Probability & Statistics. Manfred Huber 2011 1
Data Modeling & Analysis Techniques Probability & Statistics Manfred Huber 2011 1 Probability and Statistics Probability and statistics are often used interchangeably but are different, related fields
More informationA Tutorial on Probability Theory
Paola Sebastiani Department of Mathematics and Statistics University of Massachusetts at Amherst Corresponding Author: Paola Sebastiani. Department of Mathematics and Statistics, University of Massachusetts,
More informationProbability for Estimation (review)
Probability for Estimation (review) In general, we want to develop an estimator for systems of the form: x = f x, u + η(t); y = h x + ω(t); ggggg y, ffff x We will primarily focus on discrete time linear
More informationUNIT I: RANDOM VARIABLES PART- A -TWO MARKS
UNIT I: RANDOM VARIABLES PART- A -TWO MARKS 1. Given the probability density function of a continuous random variable X as follows f(x) = 6x (1-x) 0
More informationMath 370, Spring 2008 Prof. A.J. Hildebrand. Practice Test 2 Solutions
Math 370, Spring 008 Prof. A.J. Hildebrand Practice Test Solutions About this test. This is a practice test made up of a random collection of 5 problems from past Course /P actuarial exams. Most of the
More informationTransformations and Expectations of random variables
Transformations and Epectations of random variables X F X (): a random variable X distributed with CDF F X. Any function Y = g(x) is also a random variable. If both X, and Y are continuous random variables,
More information5 Double Integrals over Rectangular Regions
Chapter 7 Section 5 Doule Integrals over Rectangular Regions 569 5 Doule Integrals over Rectangular Regions In Prolems 5 through 53, use the method of Lagrange multipliers to find the indicated maximum
More informationProbability density function : An arbitrary continuous random variable X is similarly described by its probability density function f x = f X
Week 6 notes : Continuous random variables and their probability densities WEEK 6 page 1 uniform, normal, gamma, exponential,chi-squared distributions, normal approx'n to the binomial Uniform [,1] random
More informationMath 370, Actuarial Problemsolving Spring 2008 A.J. Hildebrand. Practice Test, 1/28/2008 (with solutions)
Math 370, Actuarial Problemsolving Spring 008 A.J. Hildebrand Practice Test, 1/8/008 (with solutions) About this test. This is a practice test made up of a random collection of 0 problems from past Course
More informationWhat is Statistics? Lecture 1. Introduction and probability review. Idea of parametric inference
0. 1. Introduction and probability review 1.1. What is Statistics? What is Statistics? Lecture 1. Introduction and probability review There are many definitions: I will use A set of principle and procedures
More informationCovariance and Correlation
Covariance and Correlation ( c Robert J. Serfling Not for reproduction or distribution) We have seen how to summarize a data-based relative frequency distribution by measures of location and spread, such
More informationStat 704 Data Analysis I Probability Review
1 / 30 Stat 704 Data Analysis I Probability Review Timothy Hanson Department of Statistics, University of South Carolina Course information 2 / 30 Logistics: Tuesday/Thursday 11:40am to 12:55pm in LeConte
More informationECE302 Spring 2006 HW3 Solutions February 2, 2006 1
ECE302 Spring 2006 HW3 Solutions February 2, 2006 1 Solutions to HW3 Note: Most of these solutions were generated by R. D. Yates and D. J. Goodman, the authors of our textbook. I have added comments in
More informationRANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS
RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS. DISCRETE RANDOM VARIABLES.. Definition of a Discrete Random Variable. A random variable X is said to be discrete if it can assume only a finite or countable
More informationThe Bivariate Normal Distribution
The Bivariate Normal Distribution This is Section 4.7 of the st edition (2002) of the book Introduction to Probability, by D. P. Bertsekas and J. N. Tsitsiklis. The material in this section was not included
More informationMAS108 Probability I
1 QUEEN MARY UNIVERSITY OF LONDON 2:30 pm, Thursday 3 May, 2007 Duration: 2 hours MAS108 Probability I Do not start reading the question paper until you are instructed to by the invigilators. The paper
More informationRandom variables, probability distributions, binomial random variable
Week 4 lecture notes. WEEK 4 page 1 Random variables, probability distributions, binomial random variable Eample 1 : Consider the eperiment of flipping a fair coin three times. The number of tails that
More informationSections 2.11 and 5.8
Sections 211 and 58 Timothy Hanson Department of Statistics, University of South Carolina Stat 704: Data Analysis I 1/25 Gesell data Let X be the age in in months a child speaks his/her first word and
More informationSF2940: Probability theory Lecture 8: Multivariate Normal Distribution
SF2940: Probability theory Lecture 8: Multivariate Normal Distribution Timo Koski 24.09.2015 Timo Koski Matematisk statistik 24.09.2015 1 / 1 Learning outcomes Random vectors, mean vector, covariance matrix,
More informationSOLUTIONS. f x = 6x 2 6xy 24x, f y = 3x 2 6y. To find the critical points, we solve
SOLUTIONS Problem. Find the critical points of the function f(x, y = 2x 3 3x 2 y 2x 2 3y 2 and determine their type i.e. local min/local max/saddle point. Are there any global min/max? Partial derivatives
More informationECG590I Asset Pricing. Lecture 2: Present Value 1
ECG59I Asset Pricing. Lecture 2: Present Value 1 2 Present Value If you have to decide between receiving 1$ now or 1$ one year from now, then you would rather have your money now. If you have to decide
More informationSTAT 3502. x 0 < x < 1
Solution - Assignment # STAT 350 Total mark=100 1. A large industrial firm purchases several new word processors at the end of each year, the exact number depending on the frequency of repairs in the previous
More information. (3.3) n Note that supremum (3.2) must occur at one of the observed values x i or to the left of x i.
Chapter 3 Kolmogorov-Smirnov Tests There are many situations where experimenters need to know what is the distribution of the population of their interest. For example, if they want to use a parametric
More informationNotes on Continuous Random Variables
Notes on Continuous Random Variables Continuous random variables are random quantities that are measured on a continuous scale. They can usually take on any value over some interval, which distinguishes
More informationAn Introduction to Basic Statistics and Probability
An Introduction to Basic Statistics and Probability Shenek Heyward NCSU An Introduction to Basic Statistics and Probability p. 1/4 Outline Basic probability concepts Conditional probability Discrete Random
More informationDiscrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 18. A Brief Introduction to Continuous Probability
CS 7 Discrete Mathematics and Probability Theory Fall 29 Satish Rao, David Tse Note 8 A Brief Introduction to Continuous Probability Up to now we have focused exclusively on discrete probability spaces
More informationWEEK #22: PDFs and CDFs, Measures of Center and Spread
WEEK #22: PDFs and CDFs, Measures of Center and Spread Goals: Explore the effect of independent events in probability calculations. Present a number of ways to represent probability distributions. Textbook
More informationConsumer Theory. The consumer s problem
Consumer Theory The consumer s problem 1 The Marginal Rate of Substitution (MRS) We define the MRS(x,y) as the absolute value of the slope of the line tangent to the indifference curve at point point (x,y).
More informationMath 370, Spring 2008 Prof. A.J. Hildebrand. Practice Test 1 Solutions
Math 70, Spring 008 Prof. A.J. Hildebrand Practice Test Solutions About this test. This is a practice test made up of a random collection of 5 problems from past Course /P actuarial exams. Most of the
More informationProbability and Random Variables. Generation of random variables (r.v.)
Probability and Random Variables Method for generating random variables with a specified probability distribution function. Gaussian And Markov Processes Characterization of Stationary Random Process Linearly
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 14 10/27/2008 MOMENT GENERATING FUNCTIONS
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 14 10/27/2008 MOMENT GENERATING FUNCTIONS Contents 1. Moment generating functions 2. Sum of a ranom number of ranom variables 3. Transforms
More informationE3: PROBABILITY AND STATISTICS lecture notes
E3: PROBABILITY AND STATISTICS lecture notes 2 Contents 1 PROBABILITY THEORY 7 1.1 Experiments and random events............................ 7 1.2 Certain event. Impossible event............................
More information), 35% use extra unleaded gas ( A
. At a certain gas station, 4% of the customers use regular unleaded gas ( A ), % use extra unleaded gas ( A ), and % use premium unleaded gas ( A ). Of those customers using regular gas, onl % fill their
More informationPractice with Proofs
Practice with Proofs October 6, 2014 Recall the following Definition 0.1. A function f is increasing if for every x, y in the domain of f, x < y = f(x) < f(y) 1. Prove that h(x) = x 3 is increasing, using
More information( ) is proportional to ( 10 + x)!2. Calculate the
PRACTICE EXAMINATION NUMBER 6. An insurance company eamines its pool of auto insurance customers and gathers the following information: i) All customers insure at least one car. ii) 64 of the customers
More informatione.g. arrival of a customer to a service station or breakdown of a component in some system.
Poisson process Events occur at random instants of time at an average rate of λ events per second. e.g. arrival of a customer to a service station or breakdown of a component in some system. Let N(t) be
More informationMaximum Likelihood Estimation
Math 541: Statistical Theory II Lecturer: Songfeng Zheng Maximum Likelihood Estimation 1 Maximum Likelihood Estimation Maximum likelihood is a relatively simple method of constructing an estimator for
More informationTwo Fundamental Theorems about the Definite Integral
Two Fundamental Theorems about the Definite Integral These lecture notes develop the theorem Stewart calls The Fundamental Theorem of Calculus in section 5.3. The approach I use is slightly different than
More informationNonparametric adaptive age replacement with a one-cycle criterion
Nonparametric adaptive age replacement with a one-cycle criterion P. Coolen-Schrijner, F.P.A. Coolen Department of Mathematical Sciences University of Durham, Durham, DH1 3LE, UK e-mail: Pauline.Schrijner@durham.ac.uk
More informationExperimental Uncertainty and Probability
02/04/07 PHY310: Statistical Data Analysis 1 PHY310: Lecture 03 Experimental Uncertainty and Probability Road Map The meaning of experimental uncertainty The fundamental concepts of probability 02/04/07
More informationBNG 202 Biomechanics Lab. Descriptive statistics and probability distributions I
BNG 202 Biomechanics Lab Descriptive statistics and probability distributions I Overview The overall goal of this short course in statistics is to provide an introduction to descriptive and inferential
More informationFEGYVERNEKI SÁNDOR, PROBABILITY THEORY AND MATHEmATICAL
FEGYVERNEKI SÁNDOR, PROBABILITY THEORY AND MATHEmATICAL STATIsTICs 4 IV. RANDOm VECTORs 1. JOINTLY DIsTRIBUTED RANDOm VARIABLEs If are two rom variables defined on the same sample space we define the joint
More informationSome probability and statistics
Appendix A Some probability and statistics A Probabilities, random variables and their distribution We summarize a few of the basic concepts of random variables, usually denoted by capital letters, X,Y,
More information1. First-order Ordinary Differential Equations
Advanced Engineering Mathematics 1. First-order ODEs 1 1. First-order Ordinary Differential Equations 1.1 Basic concept and ideas 1.2 Geometrical meaning of direction fields 1.3 Separable differential
More information15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
More informationCalculus 1: Sample Questions, Final Exam, Solutions
Calculus : Sample Questions, Final Exam, Solutions. Short answer. Put your answer in the blank. NO PARTIAL CREDIT! (a) (b) (c) (d) (e) e 3 e Evaluate dx. Your answer should be in the x form of an integer.
More informationGenerating Random Numbers Variance Reduction Quasi-Monte Carlo. Simulation Methods. Leonid Kogan. MIT, Sloan. 15.450, Fall 2010
Simulation Methods Leonid Kogan MIT, Sloan 15.450, Fall 2010 c Leonid Kogan ( MIT, Sloan ) Simulation Methods 15.450, Fall 2010 1 / 35 Outline 1 Generating Random Numbers 2 Variance Reduction 3 Quasi-Monte
More informationMath 370/408, Spring 2008 Prof. A.J. Hildebrand. Actuarial Exam Practice Problem Set 5 Solutions
Math 370/408, Spring 2008 Prof. A.J. Hildebrand Actuarial Exam Practice Problem Set 5 Solutions About this problem set: These are problems from Course 1/P actuarial exams that I have collected over the
More informationThe Normal Distribution. Alan T. Arnholt Department of Mathematical Sciences Appalachian State University
The Normal Distribution Alan T. Arnholt Department of Mathematical Sciences Appalachian State University arnholt@math.appstate.edu Spring 2006 R Notes 1 Copyright c 2006 Alan T. Arnholt 2 Continuous Random
More information3.4. The Binomial Probability Distribution. Copyright Cengage Learning. All rights reserved.
3.4 The Binomial Probability Distribution Copyright Cengage Learning. All rights reserved. The Binomial Probability Distribution There are many experiments that conform either exactly or approximately
More informationThe problems of first TA class of Math144
The problems of first TA class of Math144 T eaching Assistant : Liu Zhi Date : F riday, F eb 2, 27 Q1. According to the journal Chemical Engineering, an important property of a fiber is its water absorbency.
More informationOverview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written
More informationProbability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur
Probability and Statistics Prof. Dr. Somesh Kumar Department of Mathematics Indian Institute of Technology, Kharagpur Module No. #01 Lecture No. #15 Special Distributions-VI Today, I am going to introduce
More informationProbability Generating Functions
page 39 Chapter 3 Probability Generating Functions 3 Preamble: Generating Functions Generating functions are widely used in mathematics, and play an important role in probability theory Consider a sequence
More informationPolynomial Invariants
Polynomial Invariants Dylan Wilson October 9, 2014 (1) Today we will be interested in the following Question 1.1. What are all the possible polynomials in two variables f(x, y) such that f(x, y) = f(y,
More information1 if 1 x 0 1 if 0 x 1
Chapter 3 Continuity In this chapter we begin by defining the fundamental notion of continuity for real valued functions of a single real variable. When trying to decide whether a given function is or
More informationLimits and Continuity
Math 20C Multivariable Calculus Lecture Limits and Continuity Slide Review of Limit. Side limits and squeeze theorem. Continuous functions of 2,3 variables. Review: Limits Slide 2 Definition Given a function
More informationSF2940: Probability theory Lecture 8: Multivariate Normal Distribution
SF2940: Probability theory Lecture 8: Multivariate Normal Distribution Timo Koski 24.09.2014 Timo Koski () Mathematisk statistik 24.09.2014 1 / 75 Learning outcomes Random vectors, mean vector, covariance
More informationTHE BANACH CONTRACTION PRINCIPLE. Contents
THE BANACH CONTRACTION PRINCIPLE ALEX PONIECKI Abstract. This paper will study contractions of metric spaces. To do this, we will mainly use tools from topology. We will give some examples of contractions,
More informationImportant Probability Distributions OPRE 6301
Important Probability Distributions OPRE 6301 Important Distributions... Certain probability distributions occur with such regularity in real-life applications that they have been given their own names.
More informationPSTAT 120B Probability and Statistics
- Week University of California, Santa Barbara April 10, 013 Discussion section for 10B Information about TA: Fang-I CHU Office: South Hall 5431 T Office hour: TBA email: chu@pstat.ucsb.edu Slides will
More informationMeasurements of central tendency express whether the numbers tend to be high or low. The most common of these are:
A PRIMER IN PROBABILITY This handout is intended to refresh you on the elements of probability and statistics that are relevant for econometric analysis. In order to help you prioritize the information
More informationInterpretation of Somers D under four simple models
Interpretation of Somers D under four simple models Roger B. Newson 03 September, 04 Introduction Somers D is an ordinal measure of association introduced by Somers (96)[9]. It can be defined in terms
More information1. Let A, B and C are three events such that P(A) = 0.45, P(B) = 0.30, P(C) = 0.35,
1. Let A, B and C are three events such that PA =.4, PB =.3, PC =.3, P A B =.6, P A C =.6, P B C =., P A B C =.7. a Compute P A B, P A C, P B C. b Compute P A B C. c Compute the probability that exactly
More informationMath 1B, lecture 5: area and volume
Math B, lecture 5: area and volume Nathan Pflueger 6 September 2 Introduction This lecture and the next will be concerned with the computation of areas of regions in the plane, and volumes of regions in
More informationCorrelation in Random Variables
Correlation in Random Variables Lecture 11 Spring 2002 Correlation in Random Variables Suppose that an experiment produces two random variables, X and Y. What can we say about the relationship between
More informationMetric Spaces. Chapter 7. 7.1. Metrics
Chapter 7 Metric Spaces A metric space is a set X that has a notion of the distance d(x, y) between every pair of points x, y X. The purpose of this chapter is to introduce metric spaces and give some
More informationMath 370, Spring 2008 Prof. A.J. Hildebrand. Practice Test 2
Math 370, Spring 2008 Prof. A.J. Hildebrand Practice Test 2 About this test. This is a practice test made up of a random collection of 15 problems from past Course 1/P actuarial exams. Most of the problems
More information(b)using the left hand end points of the subintervals ( lower sums ) we get the aprroximation
(1) Consider the function y = f(x) =e x on the interval [, 1]. (a) Find the area under the graph of this function over this interval using the Fundamental Theorem of Calculus. (b) Subdivide the interval
More informationMATH 425, PRACTICE FINAL EXAM SOLUTIONS.
MATH 45, PRACTICE FINAL EXAM SOLUTIONS. Exercise. a Is the operator L defined on smooth functions of x, y by L u := u xx + cosu linear? b Does the answer change if we replace the operator L by the operator
More informationSo, using the new notation, P X,Y (0,1) =.08 This is the value which the joint probability function for X and Y takes when X=0 and Y=1.
Joint probabilit is the probabilit that the RVs & Y take values &. like the PDF of the two events, and. We will denote a joint probabilit function as P,Y (,) = P(= Y=) Marginal probabilit of is the probabilit
More information