Tenth Problem Assignment


 Howard Alexander
 2 years ago
 Views:
Transcription
1 EECS 40 Due on April 6, 007 PROBLEM (8 points) Dave is taking a multiplechoice exam. You may assume that the number of questions is infinite. Simultaneously, but independently, his conscious and subconscious facilities are generating answers for him, each in a Poisson manner. (His conscious and subconscious are always working on different questions.) Average rate at which conscious responses are generated = λ c responses/min Average rate at which subconscious responses are generated = λ s responses/min Each conscious response is an independent Bernoulli trial with probability p c of being correct. Similarly, each subconscious response is an independent Bernoulli trial with probability p s of being correct. Dave responds only once to each question, and you can assume that his time for recording these conscious and subconscious responses is negligible. (a) Determine p K (k), the probability mass function for the number of conscious responses Dave makes in an interval of T minutes. The number of conscious responses is a Poisson random variable with parameter λ = λ c T. Thus, p k (k) = (λ ct) k e λ ct k! (b) If we pick any question to which Dave has responded, what is the probability that his answer to that question: (a) Represents a conscious response The probability of a conscious response is λ c λ c + λ s (b) Represents a subconscious response The probability of a subconscious response is λ s λ c + λ s Due on April 6, 007
2 (c) If we pick an interval of T minutes, what is the probability that in that interval Dave will make exactly r conscious responses and exactly s subconscious responses. minutes is The probability of making r conscious and s subconscious responsed in T (λ c T) r e λ ct r! (λ st) s e λ st s! (d) Determine the moment generating function for the probability density function for random variable X, where X is the time from the start of the exam until Dave makes his first conscious response which is preceded by at least one subconscious response. Let Y s denote the random variable when the first subconscious response is generated and Y c denote the random time starting for the first subconscious response when the first conscious response is generated. Then Thus, X = Y s + Y c M X (s) = M Ys (s) M Yc (s) = λ s s λ s λ c s λ c (e) Determine the probability mass function for the total number of responses up to and including his third conscious response. Consider the arrivals of the merged process. Each arrival belongs to the conscious process with probability λ c /(λ c + λ s ). Thus, if we only count the arrivals, then the arrivals from the conscious process form a Bernoulli process with parameter p = λ c /(λ c + λ s ). Then the number of responses ( trials ) up to and including his third conscious response ( success ) has Pascal distribution with n =, that is ( ) ( ) k ( ) k λc λs p K (k) = λ c + λ s λ c + λ s (f) The papers are to be collected as soon as Dave has completed exactly N responses. Determine: The number of responses are generated by a Poisson process with rate λ c + λ s. The correct responses are generated by a Poisson process with rate p c λ c + p s λ s. Thus each response is correct with probability (p c λ c + p s λ s )/(λ c + λ s ). (i) The expected number of questions he will answer correctly Due on April 6, 007
3 The expected number of questions answered correctly is the mean of a binomial random variable and equal to N p cλ c + p s λ s λ c + λ s (ii) The probability mass function for L, the number of questions he answers correctly This probability distribution of L, the number of questions answered correctly is Binomial ( N, (p c λ c + p s λ s )/(λ c + λ s ) ), that is ( ) ( ) N pc λ c + p s λ l ( s p L (l) = p ) cλ c + p s λ N l s l λ c + λ s λ c + λ s (g) Repeat part (f) for the case in which the exam papers are to be collected at the end of a fixed interval T minutes. The number of correct responses in a fixed interval T is a Poisson process with parameter (p c λ c + p s λ s )T. So the PMF of the number of correct responses is p L (l) = with mean (p c λ c + p s λ s )T. ( (pc λ c + p s λ s )T ) l e (p c λ c +p s λ s )T l! PROBLEM (6 points) All ships travel at the same speed through a wide canal. Eastbound ship arrivals at the canal are a Poisson process with an average arrival rate λ E ships per day. Westbound ships arrive as an independent Poisson process with average arrival rate λ W per day. An indicator at a point in the canal is always pointing in the direction of travel of the most recent ship to pass it. Each ship takes t days to traverse the length of the canal. (a) Given that the pointer is pointing west: (i) What is the probability that the next ship to pass it will be westbound? The direction of the next ship does not depend on the previous ships. Therefore, this is just the probability λ W /(λ E + λ W ) that the next ship is westbound. (ii) What is the PDF for the remaining time until the pointer changes direction? The pointer will change directions on the next arrival of an eastbound ship. The time until an eastbound ship arrives is an exponential random variable with parameter λ E, and its PDF is λ E e λ Et, t 0 Due on April 6, 007
4 (b) What is the probability that an eastbound ship will see no westbound ships during its eastward journey through the canal? Suppose that an eastbound ship enters the canal at time t 0. This ship will meet any westbound ship that entered the canal between times t 0 t and t 0 + t. Thus, the desired probability is the probability that there are no westbound ship arrivals during an interval of length t, and using the Poisson PMF, it is equal to e λ Wt (c) We begin observing at an arbitrary time. Let V be the time we have to continue observing until we see the seventh eastbound ship. Determine the PDF for V. The time until we see the seventh eastbound ship is an Erlang random variable of order 7, with parameter λ E, of the form λ 7 E t6 e λ Et PROBLEM (9 points) The number of customers N who shop at a supermarket in a day is Poisson with parameter λ. The number of items purchased by any customer is Poisson with parameter µ, and the number of items purchased by different customers are independent of each other. (a) Assume that the number of items purchased by each customer is independent of N. Find E[S] and Var(S), where S is the total number of items sold. 6! Let X i be the number of items bought by the ith customer. Then which is the random sum of a random number of random variables. Thus, and S = N i=0 X i E[S] = E[N]E[X] = λµ Var(S) = E[N] Var(X) + ( E[X] ) Var(N) = λµ + λµ (b) The supermarket has two advertising strategies, one can increase λ by 0% and the other increases µ by 0%. What are the effects of these two strategies on the mean and variance of S? Which is the better strategy? 4 Due on April 6, 007
5 The better strategy will be one which keeps the variance low. (i) Increase µ by 0%. The variance becomes.λµ +.λµ. (ii) Increase λ by 0%. The variance becomes.λµ +.λµ. Thus we have to compare.λµ +.λµ.λµ +.λµ The R.H.S. is smaller so we choose option (b), i.e., increase λ by 0%. (c) Because of congestion, when there are more customers around, the amount of time each customer spends in a store tends to be shorter and hence they will more likely buy fewer items. To incorporate that, we can revise the above model so that where there are n customers, µ = c/n, where c is some constant. (i) Is the number of items bought by a customer independent of N? No. The rate at which each customer buys items depends on the number of customers in the store. So, the number of items bought is dependent on N. (ii) Find E[S] and Var(S) in this new model. E[S] = E [ E[S N] ] = E [ E[ n i=0 X i N] ] [ = E N c ] = c. N Further, as conditioned on N, X i s are independent, we have Thus, Var(S N = n) = Var ( n X i N = n ) = i=0 n i=0 Var(X i N = n) = n c n = c Var(S) = E [ Var(S N) ] + Var ( E[S N] ) = E[c] + Var(c) = c Notice that the mean and variance are same. One would suspect that it likely that S is a Poisson random variable. In fat it is easy to check this fact by evaluating the transform of S. PROBLEM 4 (8 points) The Markov chain with transition probabilities listed below is in state immediately before the first trial. p, = p, = 0.4, p, = p, = 0.6, p, = 0., p,4 = 0., p 4, = p,6 = p 6,4 =.0 Due on April 6, 007
6 (a) Draw the statetransition diagram for this Markov Chain. Indicate which states, if any, are recurrent, transient, and periodic. Recurrent States:,, 4,, 6 Transient States: Periodic States: 4,, (b) Find the probability that the process is in state after n trials. If the process leaves state, it can never return back to it. Thus the probability that the process is in state is the same as the probability that the process remains in state for all times until n. That is, the probability is 0. n. (c) Find the expected number of trials up to and including the trial on which the process leaves state. Let N be the trial on which the process leaves state. From the previous part, we know that N is a geometric random variable with success rate p = 0.8 (because given that we are in state, we will leave with probability 0.8). Thus, we have for the expected value of N: E [N] = n0. n 0.8 = 4 n= (d) Find the probability that the process never enters state. The process cannot stay in state forever. At some finite time, it will make a transition to either state or 4. If the process jumps to state, it cannot stay in state forever and at some finite time it will make a transition to state. However, if the process makes a transition from state to 4, it can never return to state. Thus the probability of never entering state is the same as the probability of jumping from state to 4 (rather than state ). That is, we have: = 8 (e) Find the probability that the process is in state 4 after 0 trials. 6 Due on April 6, 007
7 The process will be in state 4 after 0 trials if and only if makes a jump from state to 4 in trials, 4, 7 or 0. The probability of this happening is: 0. + (0.) + (0.) 6 + (0.) 9 (0.) = 0. (0.) (0.) = 0.04 (f) Given that the process is in state 4 after 0 trials, find the probability that the process was in state 4 after the first trial. Let A be the event that the process is in state 4 after 0 trials and B be the event that the process was in state 4 after the first trial. Observe that B A. Thus, P(B A) = P(A B P(A) = P(B) P(A) = = 0.99 PROBLEM (0 points) (a) Buses depart from Ann Arbor to Detroit every hour on the hour. Passengers arrive according to a Poisson process of rate λ per hour. Find the expected number of passengers on a bus. (Ignore issues of limited seating.) The expected number of passengers on a bus is the expected number of arrivals of a Poisson process of rate λ per hour in an hour, hence equal to λ. (b) Now suppose that the buses are no longer operating on a deterministic schedule, but rather their interdeparture times are independent and exponentially distributed with rate µ per hour. Find the PMF for the number of buses arriving in one hour. The interdeparture time between the buses is exponential process with rate µ, hence the departure process of the buses is a Poisson process with rate µ. Thus the PMF for the number of buses arriving in one hour is p K (k) = µk e µ (c) Let us define an event at the bus stop to be either the arrival of a passenger, or the departure of a bus. With the same assumptions as in part above, find the expected number of events that occur in one hour. The event process is the merged process of two Poisson processes, hence a Poisson process with rate λ + µ per hour. Thus the expected number of events in an hour is λ + µ. k! 7 Due on April 6, 007
8 (d) If a passenger arrives at the bus stop and sees λ people waiting, find his/her expected time to wait until the next bus. 8 Due on April 6, 007
9 The interarrival time between the buses is an exponential process, and hence memoryless. The fact that there are λ people waiting gives some information about the past of the process. But as the arrival process is memoryless, this does not convey any information about the future of the process and hence the waiting time is also exponential with rate µ. Thus, the expected waiting time is /µ. (e) Find the PMF for the number of people on a bus. We are interested only in the number of passengers who arrive between the arrivals of buses. Suppose we concentrate only on arrivals and consider the arrival of buses as sucesses and the arrival of a passenger as faliure. Thus, we are interested in the number of failures between two successes of a Bernoulli process. This has a shifted geometric distribution given by ( ) λ n ( ) µ p N (n) = λ + µ λ + µ PROBLEM 6 (9 points) For a series of dependent trials, the probability of success on any given trial is given by (k + )/(k + ), where k is the number of successes in the previous three trials. Define a state description and a set of transition probabilities which allow this process to be described as a Markov chain. Draw the state transition diagram. Try to use the smallest possible number of states. Let the outcome of the previous three trials be the state. Then Due on April 6, 007
by Dimitri P. Bertsekas and John N. Tsitsiklis Last updated: October 8, 2002
INTRODUCTION TO PROBABILITY by Dimitri P. Bertsekas and John N. Tsitsiklis CHAPTER 5: ADDITIONAL PROBLEMS Last updated: October 8, 2002 Problems marked with [D] are from Fundamentals of Applied Probability,
More informationLECTURE 16. Readings: Section 5.1. Lecture outline. Random processes Definition of the Bernoulli process Basic properties of the Bernoulli process
LECTURE 16 Readings: Section 5.1 Lecture outline Random processes Definition of the Bernoulli process Basic properties of the Bernoulli process Number of successes Distribution of interarrival times The
More informationContents. TTM4155: Teletraffic Theory (Teletrafikkteori) Probability Theory Basics. Yuming Jiang. Basic Concepts Random Variables
TTM4155: Teletraffic Theory (Teletrafikkteori) Probability Theory Basics Yuming Jiang 1 Some figures taken from the web. Contents Basic Concepts Random Variables Discrete Random Variables Continuous Random
More informatione.g. arrival of a customer to a service station or breakdown of a component in some system.
Poisson process Events occur at random instants of time at an average rate of λ events per second. e.g. arrival of a customer to a service station or breakdown of a component in some system. Let N(t) be
More informationECE302 Spring 2006 HW4 Solutions February 6, 2006 1
ECE302 Spring 2006 HW4 Solutions February 6, 2006 1 Solutions to HW4 Note: Most of these solutions were generated by R. D. Yates and D. J. Goodman, the authors of our textbook. I have added comments in
More informationMaster s Theory Exam Spring 2006
Spring 2006 This exam contains 7 questions. You should attempt them all. Each question is divided into parts to help lead you through the material. You should attempt to complete as much of each problem
More informationChapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS
Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS Part 4: Geometric Distribution Negative Binomial Distribution Hypergeometric Distribution Sections 37, 38 The remaining discrete random
More informationProbability Generating Functions
page 39 Chapter 3 Probability Generating Functions 3 Preamble: Generating Functions Generating functions are widely used in mathematics, and play an important role in probability theory Consider a sequence
More information2WB05 Simulation Lecture 8: Generating random variables
2WB05 Simulation Lecture 8: Generating random variables Marko Boon http://www.win.tue.nl/courses/2wb05 January 7, 2013 Outline 2/36 1. How do we generate random variables? 2. Fitting distributions Generating
More informationChapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS. Part 3: Discrete Uniform Distribution Binomial Distribution
Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS Part 3: Discrete Uniform Distribution Binomial Distribution Sections 35, 36 Special discrete random variable distributions we will cover
More informationPerformance Analysis of Computer Systems
Performance Analysis of Computer Systems Introduction to Queuing Theory Holger Brunst (holger.brunst@tudresden.de) Matthias S. Mueller (matthias.mueller@tudresden.de) Summary of Previous Lecture Simulation
More informationMath 461 Fall 2006 Test 2 Solutions
Math 461 Fall 2006 Test 2 Solutions Total points: 100. Do all questions. Explain all answers. No notes, books, or electronic devices. 1. [105+5 points] Assume X Exponential(λ). Justify the following two
More informationFifth Problem Assignment
EECS 40 PROBLEM (24 points) Discrete random variable X is described by the PMF { K x p X (x) = 2, if x = 0,, 2 0, for all other values of x Due on Feb 9, 2007 Let D, D 2,..., D N represent N successive
More informationMAS108 Probability I
1 QUEEN MARY UNIVERSITY OF LONDON 2:30 pm, Thursday 3 May, 2007 Duration: 2 hours MAS108 Probability I Do not start reading the question paper until you are instructed to by the invigilators. The paper
More informationECE302 Spring 2006 HW3 Solutions February 2, 2006 1
ECE302 Spring 2006 HW3 Solutions February 2, 2006 1 Solutions to HW3 Note: Most of these solutions were generated by R. D. Yates and D. J. Goodman, the authors of our textbook. I have added comments in
More informationMath 370/408, Spring 2008 Prof. A.J. Hildebrand. Actuarial Exam Practice Problem Set 2 Solutions
Math 70/408, Spring 2008 Prof. A.J. Hildebrand Actuarial Exam Practice Problem Set 2 Solutions About this problem set: These are problems from Course /P actuarial exams that I have collected over the years,
More information4 The M/M/1 queue. 4.1 Timedependent behaviour
4 The M/M/1 queue In this chapter we will analyze the model with exponential interarrival times with mean 1/λ, exponential service times with mean 1/µ and a single server. Customers are served in order
More informationThe Exponential Distribution
21 The Exponential Distribution From DiscreteTime to ContinuousTime: In Chapter 6 of the text we will be considering Markov processes in continuous time. In a sense, we already have a very good understanding
More informationNotes on Continuous Random Variables
Notes on Continuous Random Variables Continuous random variables are random quantities that are measured on a continuous scale. They can usually take on any value over some interval, which distinguishes
More informationAggregate Loss Models
Aggregate Loss Models Chapter 9 Stat 477  Loss Models Chapter 9 (Stat 477) Aggregate Loss Models Brian Hartman  BYU 1 / 22 Objectives Objectives Individual risk model Collective risk model Computing
More informationSolutions for the exam for Matematisk statistik och diskret matematik (MVE050/MSG810). Statistik för fysiker (MSG820). December 15, 2012.
Solutions for the exam for Matematisk statistik och diskret matematik (MVE050/MSG810). Statistik för fysiker (MSG8). December 15, 12. 1. (3p) The joint distribution of the discrete random variables X and
More informationHomework 4  KEY. Jeff Brenion. June 16, 2004. Note: Many problems can be solved in more than one way; we present only a single solution here.
Homework 4  KEY Jeff Brenion June 16, 2004 Note: Many problems can be solved in more than one way; we present only a single solution here. 1 Problem 21 Since there can be anywhere from 0 to 4 aces, the
More information), 35% use extra unleaded gas ( A
. At a certain gas station, 4% of the customers use regular unleaded gas ( A ), % use extra unleaded gas ( A ), and % use premium unleaded gas ( A ). Of those customers using regular gas, onl % fill their
More informationsheng@mail.ncyu.edu.tw 1 Content Introduction Expectation and variance of continuous random variables Normal random variables Exponential random variables Other continuous distributions The distribution
More information5. Continuous Random Variables
5. Continuous Random Variables Continuous random variables can take any value in an interval. They are used to model physical characteristics such as time, length, position, etc. Examples (i) Let X be
More informationWorked examples Random Processes
Worked examples Random Processes Example 1 Consider patients coming to a doctor s office at random points in time. Let X n denote the time (in hrs) that the n th patient has to wait before being admitted
More informationRandom Variable: A function that assigns numerical values to all the outcomes in the sample space.
STAT 509 Section 3.2: Discrete Random Variables Random Variable: A function that assigns numerical values to all the outcomes in the sample space. Notation: Capital letters (like Y) denote a random variable.
More informationMath 431 An Introduction to Probability. Final Exam Solutions
Math 43 An Introduction to Probability Final Eam Solutions. A continuous random variable X has cdf a for 0, F () = for 0 <
More informationExponential Distribution
Exponential Distribution Definition: Exponential distribution with parameter λ: { λe λx x 0 f(x) = 0 x < 0 The cdf: F(x) = x Mean E(X) = 1/λ. f(x)dx = Moment generating function: φ(t) = E[e tx ] = { 1
More informationChapter 2: Data quantifiers: sample mean, sample variance, sample standard deviation Quartiles, percentiles, median, interquartile range Dot diagrams
Review for Final Chapter 2: Data quantifiers: sample mean, sample variance, sample standard deviation Quartiles, percentiles, median, interquartile range Dot diagrams Histogram Boxplots Chapter 3: Set
More informationChapter 5. Random variables
Random variables random variable numerical variable whose value is the outcome of some probabilistic experiment; we use uppercase letters, like X, to denote such a variable and lowercase letters, like
More informationRenewal Theory. (iv) For s < t, N(t) N(s) equals the number of events in (s, t].
Renewal Theory Def. A stochastic process {N(t), t 0} is said to be a counting process if N(t) represents the total number of events that have occurred up to time t. X 1, X 2,... times between the events
More informationPractice Problems #4
Practice Problems #4 PRACTICE PROBLEMS FOR HOMEWORK 4 (1) Read section 2.5 of the text. (2) Solve the practice problems below. (3) Open Homework Assignment #4, solve the problems, and submit multiplechoice
More informationSome special discrete probability distributions
University of California, Los Angeles Department of Statistics Statistics 100A Instructor: Nicolas Christou Some special discrete probability distributions Bernoulli random variable: It is a variable that
More informationPractice problems for Homework 11  Point Estimation
Practice problems for Homework 11  Point Estimation 1. (10 marks) Suppose we want to select a random sample of size 5 from the current CS 3341 students. Which of the following strategies is the best:
More informationECE302 Spring 2006 HW7 Solutions March 11, 2006 1
ECE32 Spring 26 HW7 Solutions March, 26 Solutions to HW7 Note: Most of these solutions were generated by R. D. Yates and D. J. Goodman, the authors of our textbook. I have added comments in italics where
More information6.041/6.431 Spring 2008 Quiz 2 Wednesday, April 16, 7:309:30 PM. SOLUTIONS
6.4/6.43 Spring 28 Quiz 2 Wednesday, April 6, 7:39:3 PM. SOLUTIONS Name: Recitation Instructor: TA: 6.4/6.43: Question Part Score Out of 3 all 36 2 a 4 b 5 c 5 d 8 e 5 f 6 3 a 4 b 6 c 6 d 6 e 6 Total
More informationChapter 4 Expected Values
Chapter 4 Expected Values 4. The Expected Value of a Random Variables Definition. Let X be a random variable having a pdf f(x). Also, suppose the the following conditions are satisfied: x f(x) converges
More information2. Discrete random variables
2. Discrete random variables Statistics and probability: 21 If the chance outcome of the experiment is a number, it is called a random variable. Discrete random variable: the possible outcomes can be
More informationPractice Problems for Homework #6. Normal distribution and Central Limit Theorem.
Practice Problems for Homework #6. Normal distribution and Central Limit Theorem. 1. Read Section 3.4.6 about the Normal distribution and Section 4.7 about the Central Limit Theorem. 2. Solve the practice
More informationChapter 4 Lecture Notes
Chapter 4 Lecture Notes Random Variables October 27, 2015 1 Section 4.1 Random Variables A random variable is typically a realvalued function defined on the sample space of some experiment. For instance,
More informationM/M/1 and M/M/m Queueing Systems
M/M/ and M/M/m Queueing Systems M. Veeraraghavan; March 20, 2004. Preliminaries. Kendall s notation: G/G/n/k queue G: General  can be any distribution. First letter: Arrival process; M: memoryless  exponential
More informationLecture 7: Continuous Random Variables
Lecture 7: Continuous Random Variables 21 September 2005 1 Our First Continuous Random Variable The back of the lecture hall is roughly 10 meters across. Suppose it were exactly 10 meters, and consider
More informationNetwork Design Performance Evaluation, and Simulation #6
Network Design Performance Evaluation, and Simulation #6 1 Network Design Problem Goal Given QoS metric, e.g., Average delay Loss probability Characterization of the traffic, e.g., Average interarrival
More information4. Joint Distributions
Virtual Laboratories > 2. Distributions > 1 2 3 4 5 6 7 8 4. Joint Distributions Basic Theory As usual, we start with a random experiment with probability measure P on an underlying sample space. Suppose
More information0 x = 0.30 x = 1.10 x = 3.05 x = 4.15 x = 6 0.4 x = 12. f(x) =
. A mailorder computer business has si telephone lines. Let X denote the number of lines in use at a specified time. Suppose the pmf of X is as given in the accompanying table. 0 2 3 4 5 6 p(.0.5.20.25.20.06.04
More informationTopic 8 The Expected Value
Topic 8 The Expected Value Functions of Random Variables 1 / 12 Outline Names for Eg(X ) Variance and Standard Deviation Independence Covariance and Correlation 2 / 12 Names for Eg(X ) If g(x) = x, then
More informationImportant Probability Distributions OPRE 6301
Important Probability Distributions OPRE 6301 Important Distributions... Certain probability distributions occur with such regularity in reallife applications that they have been given their own names.
More information3. Continuous Random Variables
3. Continuous Random Variables A continuous random variable is one which can take any value in an interval (or union of intervals) The values that can be taken by such a variable cannot be listed. Such
More informationUNIT I: RANDOM VARIABLES PART A TWO MARKS
UNIT I: RANDOM VARIABLES PART A TWO MARKS 1. Given the probability density function of a continuous random variable X as follows f(x) = 6x (1x) 0
More informationFEGYVERNEKI SÁNDOR, PROBABILITY THEORY AND MATHEmATICAL
FEGYVERNEKI SÁNDOR, PROBABILITY THEORY AND MATHEmATICAL STATIsTICs 4 IV. RANDOm VECTORs 1. JOINTLY DIsTRIBUTED RANDOm VARIABLEs If are two rom variables defined on the same sample space we define the joint
More informationCSE 312, 2011 Winter, W.L.Ruzzo. 7. continuous random variables
CSE 312, 2011 Winter, W.L.Ruzzo 7. continuous random variables continuous random variables Discrete random variable: takes values in a finite or countable set, e.g. X {1,2,..., 6} with equal probability
More informationHypothesis Testing COMP 245 STATISTICS. Dr N A Heard. 1 Hypothesis Testing 2 1.1 Introduction... 2 1.2 Error Rates and Power of a Test...
Hypothesis Testing COMP 45 STATISTICS Dr N A Heard Contents 1 Hypothesis Testing 1.1 Introduction........................................ 1. Error Rates and Power of a Test.............................
More informationChapter 5 Discrete Probability Distribution. Learning objectives
Chapter 5 Discrete Probability Distribution Slide 1 Learning objectives 1. Understand random variables and probability distributions. 1.1. Distinguish discrete and continuous random variables. 2. Able
More informationIEOR 6711: Stochastic Models, I Fall 2012, Professor Whitt, Final Exam SOLUTIONS
IEOR 6711: Stochastic Models, I Fall 2012, Professor Whitt, Final Exam SOLUTIONS There are four questions, each with several parts. 1. Customers Coming to an Automatic Teller Machine (ATM) (30 points)
More informationJoint Exam 1/P Sample Exam 1
Joint Exam 1/P Sample Exam 1 Take this practice exam under strict exam conditions: Set a timer for 3 hours; Do not stop the timer for restroom breaks; Do not look at your notes. If you believe a question
More informationSimple Markovian Queueing Systems
Chapter 4 Simple Markovian Queueing Systems Poisson arrivals and exponential service make queueing models Markovian that are easy to analyze and get usable results. Historically, these are also the models
More informationBinomial lattice model for stock prices
Copyright c 2007 by Karl Sigman Binomial lattice model for stock prices Here we model the price of a stock in discrete time by a Markov chain of the recursive form S n+ S n Y n+, n 0, where the {Y i }
More informationSolution Using the geometric series a/(1 r) = x=1. x=1. Problem For each of the following distributions, compute
Math 472 Homework Assignment 1 Problem 1.9.2. Let p(x) 1/2 x, x 1, 2, 3,..., zero elsewhere, be the pmf of the random variable X. Find the mgf, the mean, and the variance of X. Solution 1.9.2. Using the
More informationQuestion: What is the probability that a fivecard poker hand contains a flush, that is, five cards of the same suit?
ECS20 Discrete Mathematics Quarter: Spring 2007 Instructor: John Steinberger Assistant: Sophie Engle (prepared by Sophie Engle) Homework 8 Hints Due Wednesday June 6 th 2007 Section 6.1 #16 What is the
More informationCHAPTER 6: Continuous Uniform Distribution: 6.1. Definition: The density function of the continuous random variable X on the interval [A, B] is.
Some Continuous Probability Distributions CHAPTER 6: Continuous Uniform Distribution: 6. Definition: The density function of the continuous random variable X on the interval [A, B] is B A A x B f(x; A,
More information3 Multiple Discrete Random Variables
3 Multiple Discrete Random Variables 3.1 Joint densities Suppose we have a probability space (Ω, F,P) and now we have two discrete random variables X and Y on it. They have probability mass functions f
More informationStatistics 100A Homework 8 Solutions
Part : Chapter 7 Statistics A Homework 8 Solutions Ryan Rosario. A player throws a fair die and simultaneously flips a fair coin. If the coin lands heads, then she wins twice, and if tails, the onehalf
More informationStats on the TI 83 and TI 84 Calculator
Stats on the TI 83 and TI 84 Calculator Entering the sample values STAT button Left bracket { Right bracket } Store (STO) List L1 Comma Enter Example: Sample data are {5, 10, 15, 20} 1. Press 2 ND and
More informationIntroduction to Probability
Introduction to Probability EE 179, Lecture 15, Handout #24 Probability theory gives a mathematical characterization for experiments with random outcomes. coin toss life of lightbulb binary data sequence
More informationFeb 7 Homework Solutions Math 151, Winter 2012. Chapter 4 Problems (pages 172179)
Feb 7 Homework Solutions Math 151, Winter 2012 Chapter Problems (pages 172179) Problem 3 Three dice are rolled. By assuming that each of the 6 3 216 possible outcomes is equally likely, find the probabilities
More informationLecture Notes 1. Brief Review of Basic Probability
Probability Review Lecture Notes Brief Review of Basic Probability I assume you know basic probability. Chapters 3 are a review. I will assume you have read and understood Chapters 3. Here is a very
More informationOverview of Monte Carlo Simulation, Probability Review and Introduction to Matlab
Monte Carlo Simulation: IEOR E4703 Fall 2004 c 2004 by Martin Haugh Overview of Monte Carlo Simulation, Probability Review and Introduction to Matlab 1 Overview of Monte Carlo Simulation 1.1 Why use simulation?
More informationAn Introduction to Queueing Theory
An Introduction to Queueing Theory Rein Nobel Department of Econometrics, Vrije Universiteit, Amsterdam Open Middag november 20 Overview. Basic results for queueing models in continuous time: (a) delay
More informationST 371 (IV): Discrete Random Variables
ST 371 (IV): Discrete Random Variables 1 Random Variables A random variable (rv) is a function that is defined on the sample space of the experiment and that assigns a numerical variable to each possible
More information7 Hypothesis testing  one sample tests
7 Hypothesis testing  one sample tests 7.1 Introduction Definition 7.1 A hypothesis is a statement about a population parameter. Example A hypothesis might be that the mean age of students taking MAS113X
More informationExploratory Data Analysis
Exploratory Data Analysis Johannes Schauer johannes.schauer@tugraz.at Institute of Statistics Graz University of Technology Steyrergasse 17/IV, 8010 Graz www.statistics.tugraz.at February 12, 2008 Introduction
More informationQueueing Systems. Ivo Adan and Jacques Resing
Queueing Systems Ivo Adan and Jacques Resing Department of Mathematics and Computing Science Eindhoven University of Technology P.O. Box 513, 5600 MB Eindhoven, The Netherlands March 26, 2015 Contents
More informationStat 515 Midterm Examination II April 6, 2010 (9:30 a.m.  10:45 a.m.)
Name: Stat 515 Midterm Examination II April 6, 2010 (9:30 a.m.  10:45 a.m.) The total score is 100 points. Instructions: There are six questions. Each one is worth 20 points. TA will grade the best five
More informationMonte Carlo Simulations and MATLAB
Definitions: Monte Carlo Simulations and MATLAB Monte Carlo Simulation: A simulation of a random process using a large number of computer generated samples. MATLAB: Similar to SciLab Mean: The center of
More informationST 371 (VIII): Theory of Joint Distributions
ST 371 (VIII): Theory of Joint Distributions So far we have focused on probability distributions for single random variables. However, we are often interested in probability statements concerning two or
More informationMath 425 (Fall 08) Solutions Midterm 2 November 6, 2008
Math 425 (Fall 8) Solutions Midterm 2 November 6, 28 (5 pts) Compute E[X] and Var[X] for i) X a random variable that takes the values, 2, 3 with probabilities.2,.5,.3; ii) X a random variable with the
More informationLECTURE  1 INTRODUCTION TO QUEUING SYSTEM
LECTURE  1 INTRODUCTION TO QUEUING SYSTEM Learning objective To introduce features of queuing system 9.1 Queue or Waiting lines Customers waiting to get service from server are represented by queue and
More informationPerformance Analysis, Autumn 2010
Performance Analysis, Autumn 2010 Bengt Jonsson November 16, 2010 Kendall Notation Queueing process described by A/B/X /Y /Z, where Example A is the arrival distribution B is the service pattern X the
More informationWorked examples Multiple Random Variables
Worked eamples Multiple Random Variables Eample Let X and Y be random variables that take on values from the set,, } (a) Find a joint probability mass assignment for which X and Y are independent, and
More informationExamination 110 Probability and Statistics Examination
Examination 0 Probability and Statistics Examination Sample Examination Questions The Probability and Statistics Examination consists of 5 multiplechoice test questions. The test is a threehour examination
More informationMassachusetts Institute of Technology
n (i) m m (ii) n m ( (iii) n n n n (iv) m m Massachusetts Institute of Technology 6.0/6.: Probabilistic Systems Analysis (Quiz Solutions Spring 009) Question Multiple Choice Questions: CLEARLY circle the
More informationA Tunnel Freezing Process with Adaptive Control A Stochastic Modeling Approach
A Tunnel Freezing Process with Adaptive Control A Stochastic Modeling Approach S. Ramakrishnan N. Gautam R. A. Wysk Department of Industrial and Manufacturing Engineering Pennsylvania State University
More informationISyE 6761 Fall 2012 Homework #2 Solutions
1 1. The joint p.m.f. of X and Y is (a) Find E[X Y ] for 1, 2, 3. (b) Find E[E[X Y ]]. (c) Are X and Y independent? ISE 6761 Fall 212 Homework #2 Solutions f(x, ) x 1 x 2 x 3 1 1/9 1/3 1/9 2 1/9 1/18 3
More informationSection 5.1 Continuous Random Variables: Introduction
Section 5. Continuous Random Variables: Introduction Not all random variables are discrete. For example:. Waiting times for anything (train, arrival of customer, production of mrna molecule from gene,
More informationService Management Capacity Planning and Queuing Models
Service Management Capacity Planning and Queuing Models Univ.Prof. Dr.Ing. Wolfgang Maass Chair in Economics Information and Service Systems (ISS) Saarland University, Saarbrücken, Germany WS 2011/2012
More information3(vi) B. Answer: False. 3(vii) B. Answer: True
Mathematics 0N1 Solutions 1 1. Write the following sets in list form. 1(i) The set of letters in the word banana. {a, b, n}. 1(ii) {x : x 2 + 3x 10 = 0}. 3(iv) C A. True 3(v) B = {e, e, f, c}. True 3(vi)
More informationChapters 5. Multivariate Probability Distributions
Chapters 5. Multivariate Probability Distributions Random vectors are collection of random variables defined on the same sample space. Whenever a collection of random variables are mentioned, they are
More informationWhat is Statistics? Lecture 1. Introduction and probability review. Idea of parametric inference
0. 1. Introduction and probability review 1.1. What is Statistics? What is Statistics? Lecture 1. Introduction and probability review There are many definitions: I will use A set of principle and procedures
More informationSingle Workstation Factory Models
Chapter 3 Single Workstation Factory Models Throughout the analyses given in this textbook, emphasis is on the development of steadystate system measures such as the expected number of jobs in the system
More informationLecture 8. Confidence intervals and the central limit theorem
Lecture 8. Confidence intervals and the central limit theorem Mathematical Statistics and Discrete Mathematics November 25th, 2015 1 / 15 Central limit theorem Let X 1, X 2,... X n be a random sample of
More informationStatistics 100A Homework 7 Solutions
Chapter 6 Statistics A Homework 7 Solutions Ryan Rosario. A television store owner figures that 45 percent of the customers entering his store will purchase an ordinary television set, 5 percent will purchase
More informationStochastic Performance Modelling
Stochastic Performance Modelling O. J. Boxma Department of Mathematics and Computer Science Eindhoven University of Technology P.O. Box 513, 5600 MB Eindhoven, The Netherlands 1 Preface This course presents
More informationRandom variables, probability distributions, binomial random variable
Week 4 lecture notes. WEEK 4 page 1 Random variables, probability distributions, binomial random variable Eample 1 : Consider the eperiment of flipping a fair coin three times. The number of tails that
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 14 10/27/2008 MOMENT GENERATING FUNCTIONS
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 14 10/27/2008 MOMENT GENERATING FUNCTIONS Contents 1. Moment generating functions 2. Sum of a ranom number of ranom variables 3. Transforms
More informationContinuous random variables
Continuous random variables So far we have been concentrating on discrete random variables, whose distributions are not continuous. Now we deal with the socalled continuous random variables. A random
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES Contents 1. Random variables and measurable functions 2. Cumulative distribution functions 3. Discrete
More informationIEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem
IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem Time on my hands: Coin tosses. Problem Formulation: Suppose that I have
More informationMath 370/408, Spring 2008 Prof. A.J. Hildebrand. Actuarial Exam Practice Problem Set 5 Solutions
Math 370/408, Spring 2008 Prof. A.J. Hildebrand Actuarial Exam Practice Problem Set 5 Solutions About this problem set: These are problems from Course 1/P actuarial exams that I have collected over the
More informationDISCRETE RANDOM VARIABLES
DISCRETE RANDOM VARIABLES DISCRETE RANDOM VARIABLES Documents prepared for use in course B01.1305, New York University, Stern School of Business Definitions page 3 Discrete random variables are introduced
More information