Expected Value. Let X be a discrete random variable which takes values in S X = {x 1, x 2,..., x n }


 Abraham Wiggins
 2 years ago
 Views:
Transcription
1 Expected Value Let X be a discrete random variable which takes values in S X = {x 1, x 2,..., x n } Expected Value or Mean of X: E(X) = n x i p(x i ) i=1 Example: Roll one die Let X be outcome of rolling one die. The frequency function is p(x) = 1, x = 1,..., 6, 6 and hence E(X) = 6 x=1 x 6 = 7 2 = 3.5 Example: Bernoulli random variable Let X Bin(1, θ). p(x) = θ x (1 θ) 1 x Thus the mean of X is E(X) = 0 (1 θ) + 1 θ = θ. Expected Value and Variance, Feb 2,
2 Expected Value Linearity of the expected value Let X and Y be two discrete random variables. Then E(a X + b Y ) = ae(x) + be(y ) for any constants a, b R Note: No independence is required. Proof: E(a X + b Y ) = x,y(a x + b y)p(x, y) p(x, y) = p(y) x = a x,y = a x x p(x, y) + b y p(x, y) x,y x p(x) + b y y p(y) = ae(x) + be(y ) Example: Binomial distribution Let X Bin(n, θ). Then X = X X n with X i Bin(1, θ): E(X) = n E(X i ) = n θ = nθ i=1 i=1 Expected Value and Variance, Feb 2,
3 Example: Poisson distribution Expected Value Let X be a Poisson random variable with parameter λ. E(X) = Remarks: x=0 x λx x! e λ = λ e λ = λ e λ e λ = λ x=0 λ x 1 (x 1)! For most distributions some advanced knowledge of calculus is required to find the mean. Use tables for means of commonly used distribution. Expected Value and Variance, Feb 2,
4 Expected Value Example: European Call Options Agreement that gives an investor the right (but not the obligation) to buy a stock, bond, commodity, or other instruments at a specific time at a specific price. What is a fair price P for European call options? If S T is the price of the stock at time T, the profit will be Profit = (S T K) + P. Profit is a random variable Fair price P for this option is expected value P = E(S T K) +. Expected Value and Variance, Feb 2,
5 Expected Value Example: European Call Options (contd) Consider the following simple model: S t = S t 1 + ε t, t = 1,..., T P(ε t = 1) = p and P(ε t = 1) = 1 p. S t is also called a random walk. The distribution of S T is given by (s 0 known at time 0) S T = s Y T, with Y Bin(T, p) Therefore the price P is (assuming s 0 = 0 without loss of generality) P = E(S T K) + = T (2 y T K) p θ (y) 1 {y>(k+t )/2} y=1 Let n = 20, K = 10, θ = 0.6 P = p(x) Profit Frequency function of profit Expected Value and Variance, Feb 2,
6 Example: Group testing Expected Value Suppose that a large number of blood samples are to be screened for a rare disease with prevalence 1 p. If each sample is assayed individually, n tests will be required. Alternative scheme: n samples, m groups with k samples Split each sample in half and pool all samples in one group Test pooled sample for each group If test positive test all samples in group separately What is the expected number of tests under this alternative scheme? Let X i be the number of tests in group i. The frequency function of X i is { p k if x = 1 p(x) = 1 p k if x = k + 1 The expected number of tests in each group is E(X i ) = p k + (k + 1)(1 p k ) = k + 1 kp k Hence E(N) = m Plot of E(N): i=1 E(X i ) = n (1 + 1 ) k pk The mean is minimized for groups of size 11. Proportion k Expected Value and Variance, Feb 2,
7 Let X be a random variable. Variance of X: var(x) = E ( X E(X) ) 2. Variance The variance of X is the expected squared distance of X from its mean. Suppose X is discrete random variable with S X = {x 1,..., x n }. Then the variance of X can be written as var(x) = n (x i n 2 x j p(x j )) p(xi ) i=1 j=1 Example: Roll one die X takes values in {1, 2, 3, 4, 5, 6} with frequency function p(x) = 1 6. E(X) = 6 x 1 x=1 6 = 7 2 var(x) = 6 ( 7 x 2 x=1 ) = 1 6 ( ) = We often denote the variance of a random variable X by σ 2 X, σ 2 X = var(x) and its standard deviation by σ X. Expected Value and Variance, Feb 2,
8 Properties of the Variance The variance can also be written as var(x) = E(X 2 ) ( E(X) ) 2 To see this (using linearity of the mean): var(x) = E(X E(X)) 2 = E [ X 2 2XE(X) + ( E(X) ) 2] = E ( X 2) 2E(X)E(X) + ( E(X) ) 2 = E(X 2 ) ( E(X) ) 2 Example: Let X Bin(1, θ). Then var(x) = E(X 2 ) ( E(X) ) 2 = E(X) ( E(X) ) 2 = θ θ 2 = θ (1 θ) Rules for the variance: For constants a and b var(ax + b) = a 2 var(x). For independent random variables X and Y var(x + Y ) = var(x) + var(y ). Example: Let X Bin(n, θ). Then var(x) = n θ (1 θ) Expected Value and Variance, Feb 2,
9 Covariance For independent random variables X and Y we have var(x + Y ) = var(x) + var(y ). Question: What about dependent random variables? It can be shown that var(x + Y ) = var(x) + var(y ) + 2 cov(x, Y ) where cov(x, Y ) = E [ (X E(X))(Y E(Y ) ] is the covariance of X and Y. Properties of the covariance cov(x, Y ) = E(XY ) E(X) E(Y ) cov(x, X) = var(x) cov(x, 1) = 0 cov(x, Y ) = cov(y, X) cov(a X 1 + b X 2, Y ) = a cov(x 1, Y ) + b cov(x 2, Y ) Expected Value and Variance, Feb 2,
10 Important: Covariance cov(x, Y ) = 0 does NOT imply that X and Y are independent. Example: Suppose X { 1, 0, 1} with probabilities P(X = x) = 1 3 for x = 1, 0, 1. Then E(X) = 0 and cov(x, X 2 ) = E(X 3 ) = E(X) = 0 On the other hand P(X = 1, X 2 = 0) = = P(X = 1)P(X2 = 0), that is, X and Y are not independent! Note: The covariance of X and Y measures only linear dependence. Expected Value and Variance, Feb 2,
11 Correlation The correlation coefficient ρ is defined as ρ XY = corr(x, Y ) = cov(x, Y ) var(x)var(y ). Properties: dimensionless quantity not affected by linear transformations, i.e. corr(a X + b, c Y + d) = corr(x, Y ) 1 ρ XY 1 ρ XY = 1 if and only if P(Y = a + b X) = 1 for some a and b measures linear association between X and Y Example: Three boxes: pp, pd, and dd (Ex 3.6) Let X i = 1 {penny on ith draw}. Then X i Bin(1, p) with p = 1 2 and joint frequency function p(x 1, x 2 ): x 1 \x Thus: cov(x 1, X 2 ) = E[(X 1 p)(x 2 p)] = = 1 12 corr(x 1, X 2 ) = = 1 3 Expected Value and Variance, Feb 2,
12 Prediction An instructor standardizes his midterm and final so the class average is µ = 75 and the SD is σ = 10 on both tests. The correlation between the tests is always around ρ = X  score of student on the first examination Y  score of student on the second examination Since X and Y are dependent we should be able to predict the score in the final from the midterm score. Approach: Predict Y from linear function a + b X Minimize mean squared error Solution: MSE = E ( Y a b X ) 2 = var(y b X) + [ E(Y a b X) ] 2 a = µ b µ and b = σ XY σ 2 X = ρ Thus the best linear predictor is Ŷ = µ + ρ (X µ) Note: We expect the student s score on the final to differ from the mean only by half the difference observed in the midterm (regression to the mean). Expected Value and Variance, Feb 2,
13 Summary Bernoulli distribution  Bin(1, θ) p(x) = θ x (1 θ) 1 x Binomial distribution  Bin(n, θ) ( ) n p(x) = θ x (1 θ) n x x E(X) = θ var(x) = θ(1 θ) E(X) = nθ var(x) = nθ(1 θ) Poisson distribution  Poiss(λ) p(x) = λx x! e λ E(X) = λ Geometric distribution p(x) = θ(1 θ) x 1 Hypergeometric distribution  H(N, M, n) p(x) = var(x) = λ E(X) = 1 θ var(x) = 1 θ θ 2 ( M )( N M ) x n x ) E(X) = n M N ( N n Expected Value and Variance, Feb 2,
Bivariate Distributions
Chapter 4 Bivariate Distributions 4.1 Distributions of Two Random Variables In many practical cases it is desirable to take more than one measurement of a random observation: (brief examples) 1. What is
More informationJoint Probability Distributions and Random Samples. Week 5, 2011 Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage
5 Joint Probability Distributions and Random Samples Week 5, 2011 Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage Two Discrete Random Variables The probability mass function (pmf) of a single
More informationLesson 5 Chapter 4: Jointly Distributed Random Variables
Lesson 5 Chapter 4: Jointly Distributed Random Variables Department of Statistics The Pennsylvania State University 1 Marginal and Conditional Probability Mass Functions The Regression Function Independence
More informationNotes 11 Autumn 2005
MAS 08 Probabilit I Notes Autumn 005 Two discrete random variables If X and Y are discrete random variables defined on the same sample space, then events such as X = and Y = are well defined. The joint
More informationChapters 5. Multivariate Probability Distributions
Chapters 5. Multivariate Probability Distributions Random vectors are collection of random variables defined on the same sample space. Whenever a collection of random variables are mentioned, they are
More informationExamination 110 Probability and Statistics Examination
Examination 0 Probability and Statistics Examination Sample Examination Questions The Probability and Statistics Examination consists of 5 multiplechoice test questions. The test is a threehour examination
More informationJoint Distribution and Correlation
Joint Distribution and Correlation Michael Ash Lecture 3 Reminder: Start working on the Problem Set Mean and Variance of Linear Functions of an R.V. Linear Function of an R.V. Y = a + bx What are the properties
More informationDefinition: Suppose that two random variables, either continuous or discrete, X and Y have joint density
HW MATH 461/561 Lecture Notes 15 1 Definition: Suppose that two random variables, either continuous or discrete, X and Y have joint density and marginal densities f(x, y), (x, y) Λ X,Y f X (x), x Λ X,
More informationChapter 4 Lecture Notes
Chapter 4 Lecture Notes Random Variables October 27, 2015 1 Section 4.1 Random Variables A random variable is typically a realvalued function defined on the sample space of some experiment. For instance,
More informationJoint Probability Distributions and Random Samples (Devore Chapter Five)
Joint Probability Distributions and Random Samples (Devore Chapter Five) 101634501 Probability and Statistics for Engineers Winter 20102011 Contents 1 Joint Probability Distributions 1 1.1 Two Discrete
More informationSummary of Formulas and Concepts. Descriptive Statistics (Ch. 14)
Summary of Formulas and Concepts Descriptive Statistics (Ch. 14) Definitions Population: The complete set of numerical information on a particular quantity in which an investigator is interested. We assume
More informationLecture Notes 1. Brief Review of Basic Probability
Probability Review Lecture Notes Brief Review of Basic Probability I assume you know basic probability. Chapters 3 are a review. I will assume you have read and understood Chapters 3. Here is a very
More informationCovariance and Correlation
Covariance and Correlation ( c Robert J. Serfling Not for reproduction or distribution) We have seen how to summarize a databased relative frequency distribution by measures of location and spread, such
More informationStatistics  Written Examination MEC Students  BOVISA
Statistics  Written Examination MEC Students  BOVISA Prof.ssa A. Guglielmi 26.0.2 All rights reserved. Legal action will be taken against infringement. Reproduction is prohibited without prior consent.
More informationST 371 (IV): Discrete Random Variables
ST 371 (IV): Discrete Random Variables 1 Random Variables A random variable (rv) is a function that is defined on the sample space of the experiment and that assigns a numerical variable to each possible
More information4. Joint Distributions of Two Random Variables
4. Joint Distributions of Two Random Variables 4.1 Joint Distributions of Two Discrete Random Variables Suppose the discrete random variables X and Y have supports S X and S Y, respectively. The joint
More informationFor a partition B 1,..., B n, where B i B j = for i. A = (A B 1 ) (A B 2 ),..., (A B n ) and thus. P (A) = P (A B i ) = P (A B i )P (B i )
Probability Review 15.075 Cynthia Rudin A probability space, defined by Kolmogorov (19031987) consists of: A set of outcomes S, e.g., for the roll of a die, S = {1, 2, 3, 4, 5, 6}, 1 1 2 1 6 for the roll
More informationChapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS
Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS Part 4: Geometric Distribution Negative Binomial Distribution Hypergeometric Distribution Sections 37, 38 The remaining discrete random
More informationCalculate the holding period return for this investment. It is approximately
1. An investor purchases 100 shares of XYZ at the beginning of the year for $35. The stock pays a cash dividend of $3 per share. The price of the stock at the time of the dividend is $30. The dividend
More informationJointly Distributed Random Variables
Jointly Distributed Random Variables COMP 245 STATISTICS Dr N A Heard Contents 1 Jointly Distributed Random Variables 1 1.1 Definition......................................... 1 1.2 Joint cdfs..........................................
More informationMath 141. Lecture 7: Variance, Covariance, and Sums. Albyn Jones 1. 1 Library 304. jones/courses/141
Math 141 Lecture 7: Variance, Covariance, and Sums Albyn Jones 1 1 Library 304 jones@reed.edu www.people.reed.edu/ jones/courses/141 Last Time Variance: expected squared deviation from the mean: Standard
More informationJoint Exam 1/P Sample Exam 1
Joint Exam 1/P Sample Exam 1 Take this practice exam under strict exam conditions: Set a timer for 3 hours; Do not stop the timer for restroom breaks; Do not look at your notes. If you believe a question
More informationST 371 (VIII): Theory of Joint Distributions
ST 371 (VIII): Theory of Joint Distributions So far we have focused on probability distributions for single random variables. However, we are often interested in probability statements concerning two or
More informationSolution to HW  1. Problem 1. [Points = 3] In September, Chapel Hill s daily high temperature has a mean
Problem 1. [Points = 3] In September, Chapel Hill s daily high temperature has a mean of 81 degree F and a standard deviation of 10 degree F. What is the mean, standard deviation and variance in terms
More information3 Multiple Discrete Random Variables
3 Multiple Discrete Random Variables 3.1 Joint densities Suppose we have a probability space (Ω, F,P) and now we have two discrete random variables X and Y on it. They have probability mass functions f
More informationCovariance and Correlation. Consider the joint probability distribution f XY (x, y).
Chapter 5: JOINT PROBABILITY DISTRIBUTIONS Part 2: Section 52 Covariance and Correlation Consider the joint probability distribution f XY (x, y). Is there a relationship between X and Y? If so, what kind?
More informationP (x) 0. Discrete random variables Expected value. The expected value, mean or average of a random variable x is: xp (x) = v i P (v i )
Discrete random variables Probability mass function Given a discrete random variable X taking values in X = {v 1,..., v m }, its probability mass function P : X [0, 1] is defined as: P (v i ) = Pr[X =
More informationNormal distribution Approximating binomial distribution by normal 2.10 Central Limit Theorem
1.1.2 Normal distribution 1.1.3 Approimating binomial distribution by normal 2.1 Central Limit Theorem Prof. Tesler Math 283 October 22, 214 Prof. Tesler 1.1.23, 2.1 Normal distribution Math 283 / October
More informationTopic 8 The Expected Value
Topic 8 The Expected Value Functions of Random Variables 1 / 12 Outline Names for Eg(X ) Variance and Standard Deviation Independence Covariance and Correlation 2 / 12 Names for Eg(X ) If g(x) = x, then
More informationVariances and covariances
Chapter 4 Variances and covariances 4.1 Overview The expected value of a random variable gives a crude measure for the center of location of the distribution of that random variable. For instance, if the
More informationPooling and Metaanalysis. Tony O Hagan
Pooling and Metaanalysis Tony O Hagan Pooling Synthesising prior information from several experts 2 Multiple experts The case of multiple experts is important When elicitation is used to provide expert
More informationSections 2.11 and 5.8
Sections 211 and 58 Timothy Hanson Department of Statistics, University of South Carolina Stat 704: Data Analysis I 1/25 Gesell data Let X be the age in in months a child speaks his/her first word and
More informationMathematical Expectation
Mathematical Expectation Properties of Mathematical Expectation I The concept of mathematical expectation arose in connection with games of chance. In its simplest form, mathematical expectation is the
More informationUniversity of California, Los Angeles Department of Statistics. Random variables
University of California, Los Angeles Department of Statistics Statistics Instructor: Nicolas Christou Random variables Discrete random variables. Continuous random variables. Discrete random variables.
More informationP(X = x k ) = 1 = k=1
74 CHAPTER 6. IMPORTANT DISTRIBUTIONS AND DENSITIES 6.2 Problems 5.1.1 Which are modeled with a unifm distribution? (a Yes, P(X k 1/6 f k 1,...,6. (b No, this has a binomial distribution. (c Yes, P(X k
More informationUniversity of California, Berkeley, Statistics 134: Concepts of Probability
University of California, Berkeley, Statistics 134: Concepts of Probability Michael Lugo, Spring 211 Exam 2 solutions 1. A fair twentysided die has its faces labeled 1, 2, 3,..., 2. The die is rolled
More informationExclusive OR (XOR) and hardware random number generators
Exclusive OR (XOR) and hardware random number generators Robert B Davies February 28, 2002 1 Introduction The exclusive or (XOR) operation is commonly used to reduce the bias from the bits generated by
More informationWe have discussed the notion of probabilistic dependence above and indicated that dependence is
1 CHAPTER 7 Online Supplement Covariance and Correlation for Measuring Dependence We have discussed the notion of probabilistic dependence above and indicated that dependence is defined in terms of conditional
More informationRandom variables P(X = 3) = P(X = 3) = 1 8, P(X = 1) = P(X = 1) = 3 8.
Random variables Remark on Notations 1. When X is a number chosen uniformly from a data set, What I call P(X = k) is called Freq[k, X] in the courseware. 2. When X is a random variable, what I call F ()
More informationSolutions for the exam for Matematisk statistik och diskret matematik (MVE050/MSG810). Statistik för fysiker (MSG820). December 15, 2012.
Solutions for the exam for Matematisk statistik och diskret matematik (MVE050/MSG810). Statistik för fysiker (MSG8). December 15, 12. 1. (3p) The joint distribution of the discrete random variables X and
More informationAnNajah National University Faculty of Engineering Industrial Engineering Department. Course : Quantitative Methods (65211)
AnNajah National University Faculty of Engineering Industrial Engineering Department Course : Quantitative Methods (65211) Instructor: Eng. Tamer Haddad 2 nd Semester 2009/2010 Chapter 5 Example: Joint
More informationJoint Distributions. Tieming Ji. Fall 2012
Joint Distributions Tieming Ji Fall 2012 1 / 33 X : univariate random variable. (X, Y ): bivariate random variable. In this chapter, we are going to study the distributions of bivariate random variables
More informationChapter 3 Descriptive Statistics: Numerical Measures. Learning objectives
Chapter 3 Descriptive Statistics: Numerical Measures Slide 1 Learning objectives 1. Single variable Part I (Basic) 1.1. How to calculate and use the measures of location 1.. How to calculate and use the
More informationDISCRETE RANDOM VARIABLES
DISCRETE RANDOM VARIABLES DISCRETE RANDOM VARIABLES Documents prepared for use in course B01.1305, New York University, Stern School of Business Definitions page 3 Discrete random variables are introduced
More informationCorrelation in Random Variables
Correlation in Random Variables Lecture 11 Spring 2002 Correlation in Random Variables Suppose that an experiment produces two random variables, X and Y. What can we say about the relationship between
More informationMULTIVARIATE PROBABILITY DISTRIBUTIONS
MULTIVARIATE PROBABILITY DISTRIBUTIONS. PRELIMINARIES.. Example. Consider an experiment that consists of tossing a die and a coin at the same time. We can consider a number of random variables defined
More informationChapter 5. Random variables
Random variables random variable numerical variable whose value is the outcome of some probabilistic experiment; we use uppercase letters, like X, to denote such a variable and lowercase letters, like
More informationTopic 4: Multivariate random variables. Multiple random variables
Topic 4: Multivariate random variables Joint, marginal, and conditional pmf Joint, marginal, and conditional pdf and cdf Independence Expectation, covariance, correlation Conditional expectation Two jointly
More informationElementary Statistics. Scatter Plot, Regression Line, Linear Correlation Coefficient, and Coefficient of Determination
Scatter Plot, Regression Line, Linear Correlation Coefficient, and Coefficient of Determination What is a Scatter Plot? A Scatter Plot is a plot of ordered pairs (x, y) where the horizontal axis is used
More informationECE302 Spring 2006 HW7 Solutions March 11, 2006 1
ECE32 Spring 26 HW7 Solutions March, 26 Solutions to HW7 Note: Most of these solutions were generated by R. D. Yates and D. J. Goodman, the authors of our textbook. I have added comments in italics where
More information4. Joint Distributions
Virtual Laboratories > 2. Distributions > 1 2 3 4 5 6 7 8 4. Joint Distributions Basic Theory As usual, we start with a random experiment with probability measure P on an underlying sample space. Suppose
More informationSufficient Statistics and Exponential Family. 1 Statistics and Sufficient Statistics. Math 541: Statistical Theory II. Lecturer: Songfeng Zheng
Math 541: Statistical Theory II Lecturer: Songfeng Zheng Sufficient Statistics and Exponential Family 1 Statistics and Sufficient Statistics Suppose we have a random sample X 1,, X n taken from a distribution
More informationStatistics 100 Binomial and Normal Random Variables
Statistics 100 Binomial and Normal Random Variables Three different random variables with common characteristics: 1. Flip a fair coin 10 times. Let X = number of heads out of 10 flips. 2. Poll a random
More informationBasic Statistics for SGPE Students Part I: Descriptive Statistics
Basic Statistics for SGPE Students Part I: Descriptive Statistics Achim Ahrens Anna Babloyan ahrensachim@gmail.com Erkal Ersoy annababloyan@gmail.com erkalersoy@gmail.com HeriotWatt University, Edinburgh
More informationRandom variables, probability distributions, binomial random variable
Week 4 lecture notes. WEEK 4 page 1 Random variables, probability distributions, binomial random variable Eample 1 : Consider the eperiment of flipping a fair coin three times. The number of tails that
More informationRandom Vectors and the Variance Covariance Matrix
Random Vectors and the Variance Covariance Matrix Definition 1. A random vector X is a vector (X 1, X 2,..., X p ) of jointly distributed random variables. As is customary in linear algebra, we will write
More informationWHERE DOES THE 10% CONDITION COME FROM?
1 WHERE DOES THE 10% CONDITION COME FROM? The text has mentioned The 10% Condition (at least) twice so far: p. 407 Bernoulli trials must be independent. If that assumption is violated, it is still okay
More informationLecture 6: Discrete & Continuous Probability and Random Variables
Lecture 6: Discrete & Continuous Probability and Random Variables D. Alex Hughes Math Camp September 17, 2015 D. Alex Hughes (Math Camp) Lecture 6: Discrete & Continuous Probability and Random September
More informationStatistics 100A Homework 4 Solutions
Problem 1 For a discrete random variable X, Statistics 100A Homework 4 Solutions Ryan Rosario Note that all of the problems below as you to prove the statement. We are proving the properties of epectation
More informationDepartment of Mathematics, Indian Institute of Technology, Kharagpur Assignment 23, Probability and Statistics, March 2015. Due:March 25, 2015.
Department of Mathematics, Indian Institute of Technology, Kharagpur Assignment 3, Probability and Statistics, March 05. Due:March 5, 05.. Show that the function 0 for x < x+ F (x) = 4 for x < for x
More informationHomework 4  KEY. Jeff Brenion. June 16, 2004. Note: Many problems can be solved in more than one way; we present only a single solution here.
Homework 4  KEY Jeff Brenion June 16, 2004 Note: Many problems can be solved in more than one way; we present only a single solution here. 1 Problem 21 Since there can be anywhere from 0 to 4 aces, the
More informationMAS108 Probability I
1 QUEEN MARY UNIVERSITY OF LONDON 2:30 pm, Thursday 3 May, 2007 Duration: 2 hours MAS108 Probability I Do not start reading the question paper until you are instructed to by the invigilators. The paper
More informationWhat is Statistics? Lecture 1. Introduction and probability review. Idea of parametric inference
0. 1. Introduction and probability review 1.1. What is Statistics? What is Statistics? Lecture 1. Introduction and probability review There are many definitions: I will use A set of principle and procedures
More informationEstimation with Minimum Mean Square Error
C H A P T E R 8 Estimation with Minimum Mean Square Error INTRODUCTION A recurring theme in this text and in much of communication, control and signal processing is that of making systematic estimates,
More informationMath 431 An Introduction to Probability. Final Exam Solutions
Math 43 An Introduction to Probability Final Eam Solutions. A continuous random variable X has cdf a for 0, F () = for 0 <
More informationECE302 Spring 2006 HW3 Solutions February 2, 2006 1
ECE302 Spring 2006 HW3 Solutions February 2, 2006 1 Solutions to HW3 Note: Most of these solutions were generated by R. D. Yates and D. J. Goodman, the authors of our textbook. I have added comments in
More informationChicago Booth BUSINESS STATISTICS 41000 Final Exam Fall 2011
Chicago Booth BUSINESS STATISTICS 41000 Final Exam Fall 2011 Name: Section: I pledge my honor that I have not violated the Honor Code Signature: This exam has 34 pages. You have 3 hours to complete this
More informationThe Method of Lagrange Multipliers
The Method of Lagrange Multipliers S. Sawyer October 25, 2002 1. Lagrange s Theorem. Suppose that we want to maximize (or imize a function of n variables f(x = f(x 1, x 2,..., x n for x = (x 1, x 2,...,
More informationNumerical Summarization of Data OPRE 6301
Numerical Summarization of Data OPRE 6301 Motivation... In the previous session, we used graphical techniques to describe data. For example: While this histogram provides useful insight, other interesting
More informationLecture 2: Simple Linear Regression
DMBA: Statistics Lecture 2: Simple Linear Regression Least Squares, SLR properties, Inference, and Forecasting Carlos Carvalho The University of Texas McCombs School of Business mccombs.utexas.edu/faculty/carlos.carvalho/teaching
More informationMath 461 Fall 2006 Test 2 Solutions
Math 461 Fall 2006 Test 2 Solutions Total points: 100. Do all questions. Explain all answers. No notes, books, or electronic devices. 1. [105+5 points] Assume X Exponential(λ). Justify the following two
More informationECON1003: Analysis of Economic Data Fall 2003 Answers to Quiz #2 11:40a.m. 12:25p.m. (45 minutes) Tuesday, October 28, 2003
ECON1003: Analysis of Economic Data Fall 2003 Answers to Quiz #2 11:40a.m. 12:25p.m. (45 minutes) Tuesday, October 28, 2003 1. (4 points) The number of claims for missing baggage for a wellknown airline
More informationSTAT 360 Probability and Statistics. Fall 2012
STAT 360 Probability and Statistics Fall 2012 1) General information: Crosslisted course offered as STAT 360, MATH 360 Semester: Fall 2012, Aug 20Dec 07 Course name: Probability and Statistics Number
More informationTechnology StepbyStep Using StatCrunch
Technology StepbyStep Using StatCrunch Section 1.3 Simple Random Sampling 1. Select Data, highlight Simulate Data, then highlight Discrete Uniform. 2. Fill in the following window with the appropriate
More informationRANDOM VARIABLES MATH CIRCLE (ADVANCED) 3/3/2013. 3 k) ( 52 3 )
RANDOM VARIABLES MATH CIRCLE (ADVANCED) //0 0) a) Suppose you flip a fair coin times. i) What is the probability you get 0 heads? ii) head? iii) heads? iv) heads? For = 0,,,, P ( Heads) = ( ) b) Suppose
More informationStat 704 Data Analysis I Probability Review
1 / 30 Stat 704 Data Analysis I Probability Review Timothy Hanson Department of Statistics, University of South Carolina Course information 2 / 30 Logistics: Tuesday/Thursday 11:40am to 12:55pm in LeConte
More information0 x = 0.30 x = 1.10 x = 3.05 x = 4.15 x = 6 0.4 x = 12. f(x) =
. A mailorder computer business has si telephone lines. Let X denote the number of lines in use at a specified time. Suppose the pmf of X is as given in the accompanying table. 0 2 3 4 5 6 p(.0.5.20.25.20.06.04
More information5. Continuous Random Variables
5. Continuous Random Variables Continuous random variables can take any value in an interval. They are used to model physical characteristics such as time, length, position, etc. Examples (i) Let X be
More informationContinuous Random Variables. and Probability Distributions. Continuous Random Variables and Probability Distributions ( ) ( ) Chapter 4 4.
UCLA STAT 11 A Applied Probability & Statistics for Engineers Instructor: Ivo Dinov, Asst. Prof. In Statistics and Neurology Teaching Assistant: Neda Farzinnia, UCLA Statistics University of California,
More informationEconometrics Simple Linear Regression
Econometrics Simple Linear Regression Burcu Eke UC3M Linear equations with one variable Recall what a linear equation is: y = b 0 + b 1 x is a linear equation with one variable, or equivalently, a straight
More informationRandom Variables. Chapter 2. Random Variables 1
Random Variables Chapter 2 Random Variables 1 Roulette and Random Variables A Roulette wheel has 38 pockets. 18 of them are red and 18 are black; these are numbered from 1 to 36. The two remaining pockets
More informationOverview of Monte Carlo Simulation, Probability Review and Introduction to Matlab
Monte Carlo Simulation: IEOR E4703 Fall 2004 c 2004 by Martin Haugh Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab 1 Overview of Monte Carlo Simulation 1.1 Why use simulation?
More informationChapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS. Part 3: Discrete Uniform Distribution Binomial Distribution
Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS Part 3: Discrete Uniform Distribution Binomial Distribution Sections 35, 36 Special discrete random variable distributions we will cover
More informationChapter 4. Multivariate Distributions
1 Chapter 4. Multivariate Distributions Joint p.m.f. (p.d.f.) Independent Random Variables Covariance and Correlation Coefficient Expectation and Covariance Matrix Multivariate (Normal) Distributions Matlab
More informationProbability Generating Functions
page 39 Chapter 3 Probability Generating Functions 3 Preamble: Generating Functions Generating functions are widely used in mathematics, and play an important role in probability theory Consider a sequence
More informationRandom Variable: A function that assigns numerical values to all the outcomes in the sample space.
STAT 509 Section 3.2: Discrete Random Variables Random Variable: A function that assigns numerical values to all the outcomes in the sample space. Notation: Capital letters (like Y) denote a random variable.
More informationWorked examples Multiple Random Variables
Worked eamples Multiple Random Variables Eample Let X and Y be random variables that take on values from the set,, } (a) Find a joint probability mass assignment for which X and Y are independent, and
More informationValuing Options / Volatility
Chapter 5 Valuing Options / Volatility Measures Now that the foundation regarding the basics of futures and options contracts has been set, we now move to discuss the role of volatility in futures and
More informationThe Scalar Algebra of Means, Covariances, and Correlations
3 The Scalar Algebra of Means, Covariances, and Correlations In this chapter, we review the definitions of some key statistical concepts: means, covariances, and correlations. We show how the means, variances,
More informationMälardalen University
Mälardalen University http://www.mdh.se 1/38 Value at Risk and its estimation Anatoliy A. Malyarenko Department of Mathematics & Physics Mälardalen University SE72 123 Västerås,Sweden email: amo@mdh.se
More informationFeb 7 Homework Solutions Math 151, Winter 2012. Chapter 4 Problems (pages 172179)
Feb 7 Homework Solutions Math 151, Winter 2012 Chapter Problems (pages 172179) Problem 3 Three dice are rolled. By assuming that each of the 6 3 216 possible outcomes is equally likely, find the probabilities
More informationBinomial distribution From Wikipedia, the free encyclopedia See also: Negative binomial distribution
Binomial distribution From Wikipedia, the free encyclopedia See also: Negative binomial distribution In probability theory and statistics, the binomial distribution is the discrete probability distribution
More informationNotes on Continuous Random Variables
Notes on Continuous Random Variables Continuous random variables are random quantities that are measured on a continuous scale. They can usually take on any value over some interval, which distinguishes
More information15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
More informationL10: Probability, statistics, and estimation theory
L10: Probability, statistics, and estimation theory Review of probability theory Bayes theorem Statistics and the Normal distribution Least Squares Error estimation Maximum Likelihood estimation Bayesian
More informationThe Delta Method and Applications
Chapter 5 The Delta Method and Applications 5.1 Linear approximations of functions In the simplest form of the central limit theorem, Theorem 4.18, we consider a sequence X 1, X,... of independent and
More informationEcon 132 C. Health Insurance: U.S., Risk Pooling, Risk Aversion, Moral Hazard, Rand Study 7
Econ 132 C. Health Insurance: U.S., Risk Pooling, Risk Aversion, Moral Hazard, Rand Study 7 C2. Health Insurance: Risk Pooling Health insurance works by pooling individuals together to reduce the variability
More informationFEGYVERNEKI SÁNDOR, PROBABILITY THEORY AND MATHEmATICAL
FEGYVERNEKI SÁNDOR, PROBABILITY THEORY AND MATHEmATICAL STATIsTICs 4 IV. RANDOm VECTORs 1. JOINTLY DIsTRIBUTED RANDOm VARIABLEs If are two rom variables defined on the same sample space we define the joint
More informationThe Big 50 Revision Guidelines for S1
The Big 50 Revision Guidelines for S1 If you can understand all of these you ll do very well 1. Know what is meant by a statistical model and the Modelling cycle of continuous refinement 2. Understand
More informationMT426 Notebook 3 Fall 2012 prepared by Professor Jenny Baglivo. 3 MT426 Notebook 3 3. 3.1 Definitions... 3. 3.2 Joint Discrete Distributions...
MT426 Notebook 3 Fall 2012 prepared by Professor Jenny Baglivo c Copyright 20042012 by Jenny A. Baglivo. All Rights Reserved. Contents 3 MT426 Notebook 3 3 3.1 Definitions............................................
More information