# CHAPTER 2 CONDITIONAL PROBABILITY AND INDEPENDENCE INTRODUCTION 2-1 CONDITIONAL PROBABILITY

Save this PDF as:

Size: px
Start display at page:

## Transcription

1 CHAPTER 2 CONDITIONAL PROBABILITY AND INDEPENDENCE INTRODUCTION This chapter introduces the important concepts of conditional probability and statistical independence. Conditional probabilities arise when it is known that a certain event has occurred. This knowledge changes the probabilities of events within the sample space of the experiment. Conditioning on an event occurs frequently, and understanding how to work with conditional probabilities and apply them to a particular problem or application is extremely important. In some cases, knowing that a particular event has occurred will not effect the probability of another event, and this leads to the concept of the statistical independence of events that will be developed along with the concept of conditional independence. 2-1 CONDITIONAL PROBABILITY The three probability axioms introduced in the previous chapter provide the foundation upon which to develop a theory of probability. The next step is to understand how the knowledge that a particular event has occurred will change the probabilities that are assigned to the outcomes of an experiment. The concept of a conditional probability is one of the most important concepts in probability and, although a very simple concept, conditional probability is often confusing to students. In situations where a conditional probability is to be used, it is often overlooked or incorrectly applied, thereby leading to an incorrect answer or conclusion. Perhaps the best starting point for the development of conditional probabilities is a simple example that illustrates the context in which they arise. Suppose that we have an electronic device that has a probability p n of still working after n months of continuous operation, and suppose that the probability is equal to 0.5 that the device will still be working after one year (n = 12). The device is then 45

2 46 CHAPTER 2 CONDITIONAL PROBABILITY AND INDEPENDENCE Ω B A C A B A B Conditioning event Figure 2-1: Illustration of conditioning by an event A. Any outcome not in A and any event C that is mutually exclusive of A becomes an impossible event. put into operation and after one year it is still working. The question then is What is the probability that the device will continue working for another n months. These are known as conditional probabilities because they are probabilities that are conditioned on the event that n 12. With this specific example in mind, let us now look at conditional probability in a more general context. Suppose that we have an experiment with a sample space Ω with probabilities defined on the events in Ω. If it is given that event A has occurred, then the only outcomes that are possible are those that are in A, and any outcomes that are not in A will have a probability of zero. Therefore, it is necessary to adjust or scale the probability of each elementary event within A so that the probability of event A is equal to one. A picture illustrating the effect of conditioning is given in Figure 2-1. There are three observations worth noting at this point: 1. If the probability of an event A is P{A}, and if it is given that A has occurred, then the probability of A becomes equal to one (Axiom 2). In other words, since the only outcomes that are possible are those that are in A, then A has effectively become the new sample space or the new certain event. 2. Conditioning by A will not change the relative probabilities between the experimental outcomes in A. For example, if the probability of the elementary event ω i A is equal to the probability of the elementary event

3 2-1 CONDITIONAL PROBABILITY 47 ω j A, then conditioning by A will not change this relationship. In other words, ω i and ω j will still be equally likely outcomes, P{ω i } = P{ω j }. 3. For any event C that is mutually exclusive of the conditioning event, A C =, the conditional probability of C will be equal to zero. In other words, given that A has occurred, if there are no outcomes in C that are also in A, then P{C} = 0. Important Concept Conditioning an experiment on an event A effectively changes the sample space from Ω to the conditioning event, A since any outcomes not in A will have a probability of zero. To make this a bit more concrete, consider the experiment of rolling a fair die. With a sample space consisting of six equally likely events with a probability of 1/6 for each possible outcome, suppose that the experiment is performed and we are told that the roll of the die is even (we know nothing else about the outcome, only that it is even). How does this information (conditioning) change the probabilities of the remaining events in the sample space? It should be clear that the new information (that the outcome of the roll of the die is even) should not change the relative probabilities of the remaining events, so the remaining outcomes should still be equally likely. Since only three possible outcomes remain, then their conditional probabilities should be equal to one third. Note that this also makes the probability of the conditioning event (the new sample space) equal to one. Thus, the probability that a two is rolled given that the roll resulted in an even number is equal to 1/3, P{roll a two, given that the roll is even} = 1/3 If we define the event and the event A = {Roll is even} B = {Roll a two} then this conditional probability of B given A is written as follows P{B A} = 1/3

5 2-1 CONDITIONAL PROBABILITY 49 Conditional Probability Let A be any event with nonzero probability, P{A} > 0. For any event B, the conditional probability of B given A, denoted by P{B A}, is P{B A} = P{B A} P{A} (2.1) Although it will not be done here, Eq. (2.1) may be derived as a logical consequence of the axioms of probability (see [2], p. 78). Conditional probabilities are valid probabilities in the sense that they satisfy the three probability axioms given in Sect For example, it is clear that Axiom 1 is satisfied, P{B A} 0 since both P{A B} and P{A} are non-negative. It is also clear that P{Ω A} = 1 since P{Ω A} P{Ω A} = = P{A} P{A} P{A} = 1 Finally, it is easily verified that for two mutually exclusive events B 1 and B 2, Specifically, note that P{B 1 B 2 A} = P{B 1 A} + P{B 2 A} P{B 1 B 2 A} = P{(B 1 B 2 ) A} P{A} = P{(B 1 A) (B 2 A)} P{A} Since B 1 and B 2 are mutually exclusive, then so are the events B 1 A and B 2 A. Therefore, P{(A B 1 ) (A B 2 )} = P{A B 1 } + P{A B 2 } and the result follows. A special case of conditioning occurs when A and B are mutually exclusive as illustrated in Fig. 2-2(a). Intuitively, since there are no outcomes in B that are also

6 50 CHAPTER 2 CONDITIONAL PROBABILITY AND INDEPENDENCE Ω A B B A (a) (b) Figure 2-2: Special cases of conditioning on an event A. (a) The sets A and B are mutually exclusive, (b) One set, B, is a subset of another set, A in A, if it is given that A has occurred then the probability of event B should be zero. To show this more formally, note that P{B A} = P{B A} P{A} = P{ } P{A} = 0 where the last equality follows from the fact that P{ } = 1 P{Ω} = 0. It similarly follows that P{A B} = 0 when A and B are mutually exclusive. As a specific example, consider the experiment of rolling a die, and let A and B be the following two events: A = {Roll a one} ; B = {Roll an even number} These two events are clearly disjoint, and it is clear that the probability of A given B is zero as is the probability of B given A. Another special case occurs when A as a subset of B as illustrated in Fig. 2-2(b). In this case, since B A = A, then P{B A} = P{B A} P{B} = P{A} P{A} = 1 This, of course, is an intuitive result since, if it is given that A has occurred, then any outcome in A will necessarily be an outcome in B and, therefore, event B also must have occurred. For example, when rolling a die, if A = {Roll a one} ; B = {Roll an odd number} then event A is a subset of event B, and the probability that an odd number is rolled (event B) is equal to one if it is given that a one was rolled (event A). If, on the

7 2-1 CONDITIONAL PROBABILITY 51 other hand, the conditioning event is B, then the probability of A given B is P{A B} = P{A B} P{B} = P{A} P{B} so the probability of event A is scaled by the probability of event B. Example 2-1: GEOMETRIC PROBABILITY LAW Consider an experiment that has a sample space consisting of the set of all positive integers, Ω = {1, 2, 3,...} and let N denote the outcome of an experiment defined on Ω. Suppose that the following probabilities are assigned to N, P{N = k} = ( 1 2 )k ; k = 1, 2, 3,... (2.2) This probability assignment is called a geometric probability law and is one that arises in arrival time problems as will be seen later.it is easy to show that this is a valid probability assignment since P{N = k} 0 for all k, and 2 P{Ω} = P{N 1} = ( 1 2 )k = 1 2 k=1 ( 1 2 )k = = 1 (2.3) 2 The third axiom is satisfied automatically since probabilities are assigned individually to each elementary outcome in Ω. Now let s find the probability that N > N 1 given that N > N 0 assuming that N 1 is greater than N 0 and both are positive integers. Using the definition of conditional probability, we have k=0 P{N > N 1 N > N 0 } = P{(N > N 1) (N > N 0 )} P{N > N 0 } The probability in the numerator is P{N > N 1 } = k=n 1 +1 ( 1 2 )k = ( 1 2 )N 1 = P{N > N 1} P{N > N 0 } (2.4) ( 1 2 )k = ( 1 2 )N 1 (2.5) where the last equality followed by using Eq. (2.3). Similarly, it follows that the probability in the denominator is P{N > N 0 } = ( 1 2 )N 0. Therefore, the conditional probability that N is greater than N 1 given that N is greater than N 0 is k=1 P{N > N 1 N > N 0 } = ( 1 2 )N 1 ( 1 2 )N 0 = ( 1 2 )N 1 N 0 2 In the evaluation of this probability, the geometric series is used (See Appendix 1).

8 52 CHAPTER 2 CONDITIONAL PROBABILITY AND INDEPENDENCE Figure 2-3: The memoryless property. The probability that N > N 1 given that N > N 0 is the same as the probability that N > N 1 + L given that N > N 0 + L. What is interesting is that this conditional probability depends only on the difference between N 1 and N 0. In other words, P{N > N 1 N > N 0 } = P{N > N 1 + L N > N 0 + L} for any L 0 as illustrated graphically in Fig memoryless property. This is known as the There will be instances in which it will be necessary to work with probabilities that are conditioned on two events, P{A B C}, and express this a form similar to Eq. (2.1) that maintains the conditioning on C. To see how this is done, recall that P{A D} = P{A D} P{D} (2.6) Now suppose that D is the intersection of two events, B and C, D = B C It then follows that However, we know that P{A B C} = P{A B C} P{B C} P{A B C} = P{A B C}P{C} and P{B C} = P{B C}P{C}

9 2-2 INDEPENDENCE 53 Therefore, P{A B C} = P{A B C} P{B C} The interpretation is that we first define a new sample space, C, which is the conditioning event, and then we have standard conditional probability given in Eq. (2.6) that is defined on this new space. Conditioning on Two Events P{A B C} = P{A B C} P{B C} (2.7) 2-2 INDEPENDENCE In Chapter 1, the terms independent experiments and independent outcomes were used without bothering to define what was meant by independence. With an understanding of conditional probability, it is now possible to define and gain an appreciation for what it means for one or more events to be independent, and what is meant by conditional independence INDEPENDENCE OF A PAIR OF EVENTS When it is said that events A and B are independent, our intuition suggests that this means that the outcome of one event should not have any effect or influence on the outcome of the other. It might also suggest that if it is known that one event has occurred, then this should not effect or change the probability that the other event will occur. Consider, for example, the experiment of rolling two fair dice. It is generally assumed (unless one is superstitious) that after the two dice are rolled, knowing what number appears on one of the dice will not help in knowing what number appears on the other. To make this a little more precise, suppose that one of the dice is red and the other is white, and let A be the event that a one occurs on the red die and B the event that a one occurs on the white die. Independence of these two events is taken to mean that knowing that event A occurred should not change the probability that event B occurs, and vice versa. Stated in terms of conditional probabilities, this may be written as follows, P{B A} = P{B} (2.8a)

10 54 CHAPTER 2 CONDITIONAL PROBABILITY AND INDEPENDENCE P{A B} = P{A} From the definition of conditional probability, it follows from Eq. (2.8a) that and, therefore, that P{B A} = P{B A} P{A} (2.8b) = P{B} (2.9) P{A B} = P{A}P{B} (2.10) Eq. (2.10) also implies Eq. (2.10). This leads to the following definition for the statistical independence of a pair of events, A and B: Independent Events Two events A and B are said to be statistically independent (or simply independent) when P{A B} = P{A}P{B} (2.11) Two events that are not independent are said to be dependent. Independence is a reflexive property in the sense that if A is independent of B, then B is independent of A. In other words, if the probability of event B does not change when it is given that event A occurs, then the probability of A will not change if it is given that event B occurs. The concept of independence plays a central role in probability theory and arises frequently in problems and applications. Testing for independence may not always be easy, and sometimes it is necessary to assume that certain events are independent when it is believed that such an assumption is justified. Example 2-2: INDEPENDENCE Suppose that two switches are arranged in parallel as illustrated in Fig. 2-4(a). Let A 1 be the event that switch 1 is closed and let A 2 be the event that switch 2 is closed. Assume that these events are independent, and that P{A 1 } = p 1 ; P{A 2 } = p 2 A connection exists from point X to point Y if either of the two switches are closed. Therefore, the probability that there is a connection is P{Connection} = P{A 1 A 2 } = P{A 1 }+P{A 2 } P{A 1 A 2 } = p 1 +p 2 p 1 p 2

11 2-2 INDEPENDENCE 55 1 X Y X 1 2 Y 2 (a) (b) Figure 2-4: Two switches connected in (a) parallel and (b) series. If the two switches are in series as illustrated in Fig. 2-4(b), then there will be a connection between X and Y only when both switches are closed. Therefore, for the series case, P{Connection} = P{A 1 A 2 } = p 1 p 2 There are a few properties related to the independence of events that are useful to develop since they will give more insight into what the independence of two events means. The first property is that the sample space Ω is independent of any event B Ω. 3 This follows from P{B Ω} = P{B Ω} P{Ω} = P{B} The second property is that if A and B are mutually exclusive events, A B =, with P{A} 0 and P{B} 0, then A and B will be dependent events. To understand this intuitively, note that when A and B are mutually exclusive, if event B occurs then event A cannot occur, and vice versa. Therefore, if it is known that one of these events occurs, this it is known that the other one cannot occur, thereby establishing the dependence between the two events. To show formally, note that if A and B are disjoint events, then P{A B} = P{ } = 0 3 The exclusion of B Ω is necessary because any set B is always dependent upon itself. More specifically, since P{B B} = 1 then this will not be the same as P{B}, which is required for independence, unless B = Ω.

12 56 CHAPTER 2 CONDITIONAL PROBABILITY AND INDEPENDENCE However, in order for A and B to be independent, it is necessary that P{A B} = P{A}P{B} With the assumption that both A and B have non-zero probabilities, it then follows that A and B must be dependent. The next property is that if B is a subset of A, then A and B will be dependent events unless P{A} = 1. The fundamental idea here is that if B is a subset of A, then if it is given that event B has occurred, then it is known that event A also must have occurred because any outcome in B is also an outcome in A. To demonstrate this dependence formally, note that if B A, then B A = B and P{B A} = P{B A} P{A} = P{B} P{A} P{B} unless P{A} = 1, i.e., A is the certain event. The last property is that if A and B are independent, then A and B c are also independent. To show this, note that A = (A B) (A B c ) Since B and B c are mutually exclusive events, then A B and A B c are also mutually exclusive and P{A} = P{A B} + P{A B c } Therefore, P{A B c } = P{A Bc } P{B c } = P{A} P{A B} 1 P{B} Since A and B are independent, P{A B} = P{A}P{B}, and we have P{A B c } = P{A} P{A}P{B} 1 P{B} = P{A} which establishes the independence of A and B c.

13 2-2 INDEPENDENCE 57 Properties of Independent Events 1. The events Ω and are independent of any event A unless P{A} = 1 or P{A} = If A B =, with P{A} 0 and P{B} 0, then A and B are dependent events. 3. If B A then A and B will be dependent unless P{A} = If A and B are independent, then A and B c are independent, Example 2-3: ARRIVAL OF TWO TRAINS 4 Trains X and Y arrive at a station at random times between 8:00 A.M. and 8:20 A.M. Train X stops at the station for three minutes and Train Y stops for five minutes. Assuming that the trains arrive at times that are independent of each other, we will find the probabilities of several events that are defined in terms of the train arrival times. First, however, it is necessary that we specify the underlying experiment, draw a picture of the sample space, and make probability assignments on the events that are defined within this sample space. To begin, let x be the arrival time of train X, and y the arrival time of train Y, with x and y being equal to the amount of time past 8:00 A.M. that the train arrives. It should then be clear that the outcomes of this experiment are all pairs of numbers (x, y) with 0 x 20 and 0 y 20. In other words, the sample space Ω consists of all points within the square shown in Fig. 2-5(a). The next step is to assign probabilities to events within the sample space. It is assumed that the trains arrive at random times between 8:00 A.M. and 8:20 A.M., and that the trains arrive at times that are independent of each other. What it means for a train to arrive at a random time between 8:00 A.M. and 8:20 A.M. is that a train arrival at any time within this interval is equally likely (equally probable). For example, the probability of a train arriving between 8:00 A.M. and 8:01 A.M. will be the same as the probability that it arrives between 8:10 A.M. and 8:11 A.M. (equal-length time intervals). Since the probability that the train arrives between 8:00 A.M. and 8:20 A.M. is equal to one, this suggests the following probability 4 From [3], p. 33

14 58 CHAPTER 2 CONDITIONAL PROBABILITY AND INDEPENDENCE 20 y A Ω y Ω t 4 A B B t 3 t 1 t 2 20 x x (a) (b) Figure 2-5: The experiment of two train arrivals over a twenty minute time interval. (a) The sample space, Ω, and the events A = {t 1 x t 2 }, B = {t 3 y t 4 }, and A B. (b) The event A = {y x} and the event B that the trains meet at the station, which is defined by B = { 3 < x y < 5}. measure for the event A = {t 1 x t 2 } P{A} = t 2 t 1 20 ; 0 t 1 t 2 20 Note that the event A corresponds to those outcomes that lie in the vertical strip shown in Fig. 2-5(a), and the probability of event A is equal to the width of the strip divided by 20. Furthermore, the probability of a train arriving over any collection of time intervals will be equal to the total duration of these time intervals divided by 20. For example, P{(0 x 5) (12 x 15)} = 8 20 = 0.4 A similar measure is defined for y, with Note that the event P{t 3 y t 4 } = t 4 t 3 20 B = {t 3 x t 4 } ; 0 t 3 t 4 20 is represented by the horizontal strip of outcomes in Ω shown in Fig. 2-5(a).

15 2-2 INDEPENDENCE 59 To complete the probability specification, it is necessary to determine the probability of the intersection of events A and B. Since it is assumed that the arrival times of the two trains are independent events, then A and B are independent and P{A B} = P{A}P{B} = (t 2 t 1 )(t 4 t 3 ) The event A B is the rectangular event shown in Fig. 2-5(a), and we conclude that the probability of any rectangle within Ω is equal to the area of the rectangle divided by 400. More generally, the probability of any general region within the sample space will be equal to the area of the region divided by 400. Having specified the probabilities on events in Ω, let s find the probability that train X arrives before train Y. This is the event A = {x y} which corresponds to those outcomes that are in the triangular region above the line x = y in Fig. 2-5(b). Since the area of this triangle is equal to 200, then P{A} = = 0.5 This result makes sense intuitively since each train arrives at a random time and each arrives independently of the other. Therefore, there is nothing that would make one train more likely than the other to arrive first at the station. Now let s find the probability that the trains meet at the station, i.e., the second train arrives at the station before the first one departs. Since train X is at the station for three minutes, if train X is the first to arrive, then train Y must arrive within three minutes after the arrival of train X, i.e., x y x + 3, or 0 y x 3 Similarly, if train Y is the first to arrive, since train Y remains at the station for five minutes, then train X must arrive within five minutes after the arrival of train Y, i.e., y x y + 5, or 0 x y 5 Therefore, the event that the trains meet at the station is C = { 3 x y 5} which corresponds to the shaded region consisting of two trapezoids shown in Fig. 2-5(b). Since the area of these trapezoids is 143, then P{C} =

16 60 CHAPTER 2 CONDITIONAL PROBABILITY AND INDEPENDENCE INDEPENDENCE OF MORE THAN TWO EVENTS The definition given in Eq. (2.11) is concerned with the independence of a pair of events, A and B. If there are three events, A, B, and C, then it would be tempting to say that A, B, and C are independent if the following three conditions hold: P{A B} = P{A}P{B} P{B C} = P{B}P{C} P{C A} = P{C}P{A} (2.12) However, when Eq. (2.12) is satisfied, then A, B, and C are said to be independent in pairs, which means that the occurrence of any one of the three events will not have any effect on the probability that either one of the other events will occur. However, Eq. (2.12) does not necessarily imply that the probability of one of the events will not change if it is given that the other two events have occurred. In other words, it may not necessarily follow from Eq. (2.12) that P{A B C} = P{A} nor is it necessarily true that independence in pairs imples that P{A B C} = P{A}P{B}P{C} The following example illustrates this point and shows that some care is needed in dealing with independence of three or more events, and that sometimes our intuition may fail us. Example 2-4: INDEPENDENCE IN PAIRS Consider a digital transmitter that sends two binary digits, b 1 and b 2, with each bit being equally likely to be a zero or a one, P{b i = 0} = P{b i = 1} = 1 2 ; i = 1, 2 In addition, suppose that the events {b 1 = i} is independent of the event {b 2 = j} for i, j = 1, 2, P{(b 1 = i) (b 2 = j)} = P{b 1 = i}p{b 2 = j} = 1 4 ; i, j = 0, 1 The sample space for this experiment consists of four possible outcomes, each corresponding to one of the four possible pairs of binary digits as illustrated in Fig. 2-6(a). Now let A be the event that the first bit is zero,

17 2-2 INDEPENDENCE 61 (a) (b) Figure 2-6: Independence in pairs. A = {b 1 = 0} = {00} {01} and B the event that the second bit is zero, B = {b 2 = 0} = {00} {10} and C the event that the two bits are the same, C = {b 1 = b 2 } = {00} {11} These events are illustrated in Fig. 2-6(b). Since the probability of each elementary event is equal to 1/4, and since each of the events A, B, and C contain exactly two elementary events, then P{A} = P{B} = P{C} = 1/2 It is easy to show that these three events are independent in pairs. For example, since P{A B} = P{00} = 1 4 = P{A}P{B} then A and B are independent. It may similarly be shown that A and C are independent and that B and C are independent. However, consider what happens when one of the events is conditioned on the other two, such as P{A B C}. In this case, P{A B C} = P{A B C} P{B C}

18 62 CHAPTER 2 CONDITIONAL PROBABILITY AND INDEPENDENCE and since A B C = {00} and B C = {00} are the same event, then Therefore, P{A B C} = 1 P{A B C} P{A} and it follows that A is not independent of the event B C. In addition, note that since P{A B C} = 1 4 and then P{A}P{B}P{C} = 1 16 P{A B C} P{A}P{B}P{C} which would be the obvious generalization of the definition given in Eq. (2.11) for three events. The previous example shows that generalizing the definition for the independence of two events to the independence of three events requires more than pair-wise independence. Therefore, what is required for three events to be said to be statistically independent is given in the following definition: Independence of Three Events Three events A, B, and C are said to be statistically independent if they are independent in pairs, and P{A B C} = P{A}P{B}P{C} (2.13) The extension to more than three events follows by induction. For example, four events A, B, C, and D are independent if they are independent in groups of three, and P{A B C D} = P{A}P{B}P{C}P{D} Continuing, events A i for i = 1,..., n are said to be independent if they are independent in groups of n 1 and { n } n P A i = P{A i } i=1 i=1

19 2-2 INDEPENDENCE CONDITIONAL INDEPENDENCE Recall that if A and B are independent events, then event B does not have any influence on event A, and the occurrence of B will not change the probability of event A. Since independence is reflexive, then the converse is also true. Frequently, however, there will be cases in which two events are independent, but this independence will depend (explicitly or implicitly) on some other condition or event. To understand how such a situation might arise, consider the following example. Example 2-5: ELECTRICAL COMPONENTS 5 Suppose that an electronic system has two components that operate independently of each other in the sense that the failure of one component is not affected by and does not have any effect on the failure of the other. In addition, let A and B be the following events: A = {Component 1 operates without failure for one year} B = {Component 2 operates without failure for one year} It would seem natural to assume that events A and B are statistically independent given the assumption of operational independence. However, this may not be the case since, in some situations, there may be other random factors or events that affect each component in different ways. In these cases, statistical independence will be conditioned (depend upon) these other factors or events. For example, suppose that the operating temperature of the system affects the likelihood of a failure of each component, and it does so in different ways. More specifically, let C be the event that the system is operated within what is considered to be the normal temperature range for at least 90% of the time, and suppose that C = {Normal Temperature Range 90% of the time} P{A C} = 0.9 ; P{B C} = 0.8 and P{A C c } = 0.8 ; P{B C c } = 0.7 In addition, let us assume that P{C} = 0.9. Since the components operate independently under any given temperature, then it is reasonable to assume that 5 From Pfeiffer 4. P{A B C} = P{A C} (2.14) P{A B C c } = P{A C c } (2.15)

20 64 CHAPTER 2 CONDITIONAL PROBABILITY AND INDEPENDENCE In other words, given that the temperature is within the normal range, then failure of one component is not affected by the failure of the other, and the same is true if the temperature is not within the normal range. From Eq. (2.7) we know that the conditional probability P{A B C} is equal to P{A B C} = P{A B C} P{B C} = P{A C} Therefore, it follows from Eq. (2.14) that P{A B C} = P{A C}P{B C} (2.16) which says that A and B are independent when conditioned on event C. Similarly, it follows from Eq. (2.7) and Eq. (2.15) that P{A B C c } = P{A C c }P{B C c } (2.17) However, neither Eq. (2.16) nor Eq. (2.17) necessarily imply that A and B are (unconditionally) independent, since this requires that P{A B} = P{A}P{B} To determine whether or not A and B are independent, we may use the special case of the total probability theorem given in Eq. (3.2) to find the probability of event A, as well as the probability of event B, P{A} = P{A C}P{C} + P{A C c }P{C c } = (0.9)(0.9) + (0.8)(0.1) = 0.89 P{B} = P{B C}P{C} + P{B C c }P{C c } = (0.8)(0.9) + (0.7)(0.1) = 0.79 Finally, again using Eq. (3.2) along with Eq. (2.16) and Eq. (2.17) we have P{A B} = P{A B C}P{C} + P{A B C c }P{C c } = P{A C}P{B C}P{C} + P{A C c }P{B C c }P{C c } = (0.9)(0.8)(0.9) + (0.8)(0.7)(0.1) = Since P{A B} = P{A}P{B} = then A and B are not independent.

21 2-2 INDEPENDENCE 65 In order to more clearly understand where the dependency is coming in, note that if Component 1 fails, then it is more likely that the operating temperature is outside the normal range, which increases the probability that the second component will fail. If Component 1 does not fail, then this makes it more likely that the operating temperature is within the normal range and, therefore, it is less likely that Component 2 will fail. As illustrated in the previous example, two events A and B that are not statistically independent may become independent when conditioned on another event C. This leads to the concept of conditional independence, which is defined as follows: Conditional Independence Two events A and B are said to be conditionally independent given an event C if P{A B C} = P{A C}P{B C} (2.18) A convenient way to interpret Eq. (2.18) and to view the concept of conditional independence is as follows. When it is given that event C occurs, then C becomes the new sample space, and it is in this new sample space that event B becomes independent of A. Thus, the conditioning event removes the dependencies that exist between A and B. As is the case for independence, conditional independence is reflexive in the sense that if A is conditionally independent of B given C then B is conditionally independent of A given C. One might be tempted to conclude that conditional independence is a weaker form of independence in the sense that if A and B are independent, then they will be independent for any conditioning event C. This, however, is not the case as illustrated abstractly in Fig. 2-7, which shows two events A and B with A B not empty and a conditioning set C that includes elements from both A and B. If P{A} = P{B} = 1 2 and P{A B} = 1 4, then A and B are independent events. However, note that P{A B C} = 0 while both P{A C} and P{B C} are non-zero. Therefore, A and B are not conditionally independent when the conditioning event is C. A more concrete example is given below. Example 2-6: INDEPENDENT BUT NOT CONDITIONALLY INDEPENDENT

22 66 CHAPTER 2 CONDITIONAL PROBABILITY AND INDEPENDENCE Figure 2-7: Something. Let Ω = {1, 2, 3, 4} be a set of five equally like outcomes, and let A = {1, 2} ; B = {2, 3} Clearly, P{A} = 1/2 and P{B} = 1/2 and P{A B} = 1/4. Therefore, A and B are independent. However, if C = {1, 4}, then P{A C} = 1/2 and P{B C} = 1/2 while P{A B C} = 0. Therefore, although A and B are independent, they are not conditionally independent given the event C. References 1. Alvin W.Drake, Fundamentals of Applied Probability Theory, McGraw-Hill, New York, Harold J. Larson and Bruno O. Schubert, Random Variables and Stochastic Processes, Volume 1, John Wiley & Sons, A. Papoulis, Probability, Random Variables, and Stochastic Processes, McGraw-Hill, Second Edition, P. Pfeiffer, Probability for Applications, Springer, 1989.

### Discrete Mathematics and Probability Theory Fall 2009 Satish Rao,David Tse Note 11

CS 70 Discrete Mathematics and Probability Theory Fall 2009 Satish Rao,David Tse Note Conditional Probability A pharmaceutical company is marketing a new test for a certain medical condition. According

### Definition and Calculus of Probability

In experiments with multivariate outcome variable, knowledge of the value of one variable may help predict another. For now, the word prediction will mean update the probabilities of events regarding the

### 8. You have three six-sided dice (one red, one green, and one blue.) The numbers of the dice are

Probability Warm up problem. Alice and Bob each pick, at random, a real number between 0 and 10. Call Alice s number A and Bob s number B. The is the probability that A B? What is the probability that

### Lecture 1 Introduction Properties of Probability Methods of Enumeration Asrat Temesgen Stockholm University

Lecture 1 Introduction Properties of Probability Methods of Enumeration Asrat Temesgen Stockholm University 1 Chapter 1 Probability 1.1 Basic Concepts In the study of statistics, we consider experiments

### A Few Basics of Probability

A Few Basics of Probability Philosophy 57 Spring, 2004 1 Introduction This handout distinguishes between inductive and deductive logic, and then introduces probability, a concept essential to the study

### Elementary Number Theory We begin with a bit of elementary number theory, which is concerned

CONSTRUCTION OF THE FINITE FIELDS Z p S. R. DOTY Elementary Number Theory We begin with a bit of elementary number theory, which is concerned solely with questions about the set of integers Z = {0, ±1,

### 6.3 Conditional Probability and Independence

222 CHAPTER 6. PROBABILITY 6.3 Conditional Probability and Independence Conditional Probability Two cubical dice each have a triangle painted on one side, a circle painted on two sides and a square painted

### Probability Models.S1 Introduction to Probability

Probability Models.S1 Introduction to Probability Operations Research Models and Methods Paul A. Jensen and Jonathan F. Bard The stochastic chapters of this book involve random variability. Decisions are

### Combinatorics: The Fine Art of Counting

Combinatorics: The Fine Art of Counting Week 7 Lecture Notes Discrete Probability Continued Note Binomial coefficients are written horizontally. The symbol ~ is used to mean approximately equal. The Bernoulli

### Probability. A random sample is selected in such a way that every different sample of size n has an equal chance of selection.

1 3.1 Sample Spaces and Tree Diagrams Probability This section introduces terminology and some techniques which will eventually lead us to the basic concept of the probability of an event. The Rare Event

### Discrete Mathematics for CS Fall 2006 Papadimitriou & Vazirani Lecture 22

CS 70 Discrete Mathematics for CS Fall 2006 Papadimitriou & Vazirani Lecture 22 Introduction to Discrete Probability Probability theory has its origins in gambling analyzing card games, dice, roulette

### Chapter 3: The basic concepts of probability

Chapter 3: The basic concepts of probability Experiment: a measurement process that produces quantifiable results (e.g. throwing two dice, dealing cards, at poker, measuring heights of people, recording

### We now explore a third method of proof: proof by contradiction.

CHAPTER 6 Proof by Contradiction We now explore a third method of proof: proof by contradiction. This method is not limited to proving just conditional statements it can be used to prove any kind of statement

### Worked examples Basic Concepts of Probability Theory

Worked examples Basic Concepts of Probability Theory Example 1 A regular tetrahedron is a body that has four faces and, if is tossed, the probability that it lands on any face is 1/4. Suppose that one

### Applications of Methods of Proof

CHAPTER 4 Applications of Methods of Proof 1. Set Operations 1.1. Set Operations. The set-theoretic operations, intersection, union, and complementation, defined in Chapter 1.1 Introduction to Sets are

### P (A) = lim P (A) = N(A)/N,

1.1 Probability, Relative Frequency and Classical Definition. Probability is the study of random or non-deterministic experiments. Suppose an experiment can be repeated any number of times, so that we

### Ch. 13.2: Mathematical Expectation

Ch. 13.2: Mathematical Expectation Random Variables Very often, we are interested in sample spaces in which the outcomes are distinct real numbers. For example, in the experiment of rolling two dice, we

### E3: PROBABILITY AND STATISTICS lecture notes

E3: PROBABILITY AND STATISTICS lecture notes 2 Contents 1 PROBABILITY THEORY 7 1.1 Experiments and random events............................ 7 1.2 Certain event. Impossible event............................

### Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 10

CS 70 Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 10 Introduction to Discrete Probability Probability theory has its origins in gambling analyzing card games, dice,

### Finite and discrete probability distributions

8 Finite and discrete probability distributions To understand the algorithmic aspects of number theory and algebra, and applications such as cryptography, a firm grasp of the basics of probability theory

### Probability OPRE 6301

Probability OPRE 6301 Random Experiment... Recall that our eventual goal in this course is to go from the random sample to the population. The theory that allows for this transition is the theory of probability.

### INCIDENCE-BETWEENNESS GEOMETRY

INCIDENCE-BETWEENNESS GEOMETRY MATH 410, CSUSM. SPRING 2008. PROFESSOR AITKEN This document covers the geometry that can be developed with just the axioms related to incidence and betweenness. The full

### Math/Stats 425 Introduction to Probability. 1. Uncertainty and the axioms of probability

Math/Stats 425 Introduction to Probability 1. Uncertainty and the axioms of probability Processes in the real world are random if outcomes cannot be predicted with certainty. Example: coin tossing, stock

### Probability. a number between 0 and 1 that indicates how likely it is that a specific event or set of events will occur.

Probability Probability Simple experiment Sample space Sample point, or elementary event Event, or event class Mutually exclusive outcomes Independent events a number between 0 and 1 that indicates how

### Basic Probability Concepts

page 1 Chapter 1 Basic Probability Concepts 1.1 Sample and Event Spaces 1.1.1 Sample Space A probabilistic (or statistical) experiment has the following characteristics: (a) the set of all possible outcomes

### Elements of probability theory

2 Elements of probability theory Probability theory provides mathematical models for random phenomena, that is, phenomena which under repeated observations yield di erent outcomes that cannot be predicted

### 5.1 Radical Notation and Rational Exponents

Section 5.1 Radical Notation and Rational Exponents 1 5.1 Radical Notation and Rational Exponents We now review how exponents can be used to describe not only powers (such as 5 2 and 2 3 ), but also roots

### The Casino Lab STATION 1: CRAPS

The Casino Lab Casinos rely on the laws of probability and expected values of random variables to guarantee them profits on a daily basis. Some individuals will walk away very wealthy, while others will

### Hypothesis Testing for Beginners

Hypothesis Testing for Beginners Michele Piffer LSE August, 2011 Michele Piffer (LSE) Hypothesis Testing for Beginners August, 2011 1 / 53 One year ago a friend asked me to put down some easy-to-read notes

### Math212a1010 Lebesgue measure.

Math212a1010 Lebesgue measure. October 19, 2010 Today s lecture will be devoted to Lebesgue measure, a creation of Henri Lebesgue, in his thesis, one of the most famous theses in the history of mathematics.

### Chapter 3. Cartesian Products and Relations. 3.1 Cartesian Products

Chapter 3 Cartesian Products and Relations The material in this chapter is the first real encounter with abstraction. Relations are very general thing they are a special type of subset. After introducing

### LEARNING OBJECTIVES FOR THIS CHAPTER

CHAPTER 2 American mathematician Paul Halmos (1916 2006), who in 1942 published the first modern linear algebra book. The title of Halmos s book was the same as the title of this chapter. Finite-Dimensional

### Pigeonhole Principle Solutions

Pigeonhole Principle Solutions 1. Show that if we take n + 1 numbers from the set {1, 2,..., 2n}, then some pair of numbers will have no factors in common. Solution: Note that consecutive numbers (such

### Basic Probability. Probability: The part of Mathematics devoted to quantify uncertainty

AMS 5 PROBABILITY Basic Probability Probability: The part of Mathematics devoted to quantify uncertainty Frequency Theory Bayesian Theory Game: Playing Backgammon. The chance of getting (6,6) is 1/36.

### Arkansas Tech University MATH 4033: Elementary Modern Algebra Dr. Marcel B. Finan

Arkansas Tech University MATH 4033: Elementary Modern Algebra Dr. Marcel B. Finan 3 Binary Operations We are used to addition and multiplication of real numbers. These operations combine two real numbers

### 2.1 The Algebra of Sets

Chapter 2 Abstract Algebra 83 part of abstract algebra, sets are fundamental to all areas of mathematics and we need to establish a precise language for sets. We also explore operations on sets and relations

### Pythagorean Triples. Chapter 2. a 2 + b 2 = c 2

Chapter Pythagorean Triples The Pythagorean Theorem, that beloved formula of all high school geometry students, says that the sum of the squares of the sides of a right triangle equals the square of the

### Introduction to probability theory in the Discrete Mathematics course

Introduction to probability theory in the Discrete Mathematics course Jiří Matoušek (KAM MFF UK) Version: Oct/18/2013 Introduction This detailed syllabus contains definitions, statements of the main results

### Probability definitions

Probability definitions 1. Probability of an event = chance that the event will occur. 2. Experiment = any action or process that generates observations. In some contexts, we speak of a data-generating

### Chapter 3. Distribution Problems. 3.1 The idea of a distribution. 3.1.1 The twenty-fold way

Chapter 3 Distribution Problems 3.1 The idea of a distribution Many of the problems we solved in Chapter 1 may be thought of as problems of distributing objects (such as pieces of fruit or ping-pong balls)

### Math 319 Problem Set #3 Solution 21 February 2002

Math 319 Problem Set #3 Solution 21 February 2002 1. ( 2.1, problem 15) Find integers a 1, a 2, a 3, a 4, a 5 such that every integer x satisfies at least one of the congruences x a 1 (mod 2), x a 2 (mod

### CHAPTER 3 Numbers and Numeral Systems

CHAPTER 3 Numbers and Numeral Systems Numbers play an important role in almost all areas of mathematics, not least in calculus. Virtually all calculus books contain a thorough description of the natural,

### ALGEBRA. sequence, term, nth term, consecutive, rule, relationship, generate, predict, continue increase, decrease finite, infinite

ALGEBRA Pupils should be taught to: Generate and describe sequences As outcomes, Year 7 pupils should, for example: Use, read and write, spelling correctly: sequence, term, nth term, consecutive, rule,

### Basic concepts in probability. Sue Gordon

Mathematics Learning Centre Basic concepts in probability Sue Gordon c 2005 University of Sydney Mathematics Learning Centre, University of Sydney 1 1 Set Notation You may omit this section if you are

### Mathematical Induction

Mathematical Induction (Handout March 8, 01) The Principle of Mathematical Induction provides a means to prove infinitely many statements all at once The principle is logical rather than strictly mathematical,

### Lecture Note 1 Set and Probability Theory. MIT 14.30 Spring 2006 Herman Bennett

Lecture Note 1 Set and Probability Theory MIT 14.30 Spring 2006 Herman Bennett 1 Set Theory 1.1 Definitions and Theorems 1. Experiment: any action or process whose outcome is subject to uncertainty. 2.

### MATH10040 Chapter 2: Prime and relatively prime numbers

MATH10040 Chapter 2: Prime and relatively prime numbers Recall the basic definition: 1. Prime numbers Definition 1.1. Recall that a positive integer is said to be prime if it has precisely two positive

### In the situations that we will encounter, we may generally calculate the probability of an event

What does it mean for something to be random? An event is called random if the process which produces the outcome is sufficiently complicated that we are unable to predict the precise result and are instead

### The Language of Mathematics

CHPTER 2 The Language of Mathematics 2.1. Set Theory 2.1.1. Sets. set is a collection of objects, called elements of the set. set can be represented by listing its elements between braces: = {1, 2, 3,

### Induction. Margaret M. Fleck. 10 October These notes cover mathematical induction and recursive definition

Induction Margaret M. Fleck 10 October 011 These notes cover mathematical induction and recursive definition 1 Introduction to induction At the start of the term, we saw the following formula for computing

### Sections 2.1, 2.2 and 2.4

SETS Sections 2.1, 2.2 and 2.4 Chapter Summary Sets The Language of Sets Set Operations Set Identities Introduction Sets are one of the basic building blocks for the types of objects considered in discrete

### 8 Divisibility and prime numbers

8 Divisibility and prime numbers 8.1 Divisibility In this short section we extend the concept of a multiple from the natural numbers to the integers. We also summarize several other terms that express

### Probabilities. Probability of a event. From Random Variables to Events. From Random Variables to Events. Probability Theory I

Victor Adamchi Danny Sleator Great Theoretical Ideas In Computer Science Probability Theory I CS 5-25 Spring 200 Lecture Feb. 6, 200 Carnegie Mellon University We will consider chance experiments with

### Probability & Probability Distributions

Probability & Probability Distributions Carolyn J. Anderson EdPsych 580 Fall 2005 Probability & Probability Distributions p. 1/61 Probability & Probability Distributions Elementary Probability Theory Definitions

### WRITING PROOFS. Christopher Heil Georgia Institute of Technology

WRITING PROOFS Christopher Heil Georgia Institute of Technology A theorem is just a statement of fact A proof of the theorem is a logical explanation of why the theorem is true Many theorems have this

### Worksheet for Teaching Module Probability (Lesson 1)

Worksheet for Teaching Module Probability (Lesson 1) Topic: Basic Concepts and Definitions Equipment needed for each student 1 computer with internet connection Introduction In the regular lectures in

### MAT 1000. Mathematics in Today's World

MAT 1000 Mathematics in Today's World We talked about Cryptography Last Time We will talk about probability. Today There are four rules that govern probabilities. One good way to analyze simple probabilities

### Problem Set I: Preferences, W.A.R.P., consumer choice

Problem Set I: Preferences, W.A.R.P., consumer choice Paolo Crosetto paolo.crosetto@unimi.it Exercises solved in class on 18th January 2009 Recap:,, Definition 1. The strict preference relation is x y

### MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES Contents 1. Random variables and measurable functions 2. Cumulative distribution functions 3. Discrete

### COMBINATORIAL PROPERTIES OF THE HIGMAN-SIMS GRAPH. 1. Introduction

COMBINATORIAL PROPERTIES OF THE HIGMAN-SIMS GRAPH ZACHARY ABEL 1. Introduction In this survey we discuss properties of the Higman-Sims graph, which has 100 vertices, 1100 edges, and is 22 regular. In fact

### Factoring Patterns in the Gaussian Plane

Factoring Patterns in the Gaussian Plane Steve Phelps Introduction This paper describes discoveries made at the Park City Mathematics Institute, 00, as well as some proofs. Before the summer I understood

### POWER SETS AND RELATIONS

POWER SETS AND RELATIONS L. MARIZZA A. BAILEY 1. The Power Set Now that we have defined sets as best we can, we can consider a sets of sets. If we were to assume nothing, except the existence of the empty

### Basic Probability Theory I

A Probability puzzler!! Basic Probability Theory I Dr. Tom Ilvento FREC 408 Our Strategy with Probability Generally, we want to get to an inference from a sample to a population. In this case the population

### The Calculus of Probability

The Calculus of Probability Let A and B be events in a sample space S. Partition rule: P(A) = P(A B) + P(A B ) Example: Roll a pair of fair dice P(Total of 10) = P(Total of 10 and double) + P(Total of

### Unraveling versus Unraveling: A Memo on Competitive Equilibriums and Trade in Insurance Markets

Unraveling versus Unraveling: A Memo on Competitive Equilibriums and Trade in Insurance Markets Nathaniel Hendren January, 2014 Abstract Both Akerlof (1970) and Rothschild and Stiglitz (1976) show that

### So let us begin our quest to find the holy grail of real analysis.

1 Section 5.2 The Complete Ordered Field: Purpose of Section We present an axiomatic description of the real numbers as a complete ordered field. The axioms which describe the arithmetic of the real numbers

### Determinants, Areas and Volumes

Determinants, Areas and Volumes Theodore Voronov Part 2 Areas and Volumes The area of a two-dimensional object such as a region of the plane and the volume of a three-dimensional object such as a solid

### I. WHAT IS PROBABILITY?

C HAPTER 3 PROBABILITY Random Experiments I. WHAT IS PROBABILITY? The weatherman on 0 o clock news program states that there is a 20% chance that it will snow tomorrow, a 65% chance that it will rain and

### Week 2: Conditional Probability and Bayes formula

Week 2: Conditional Probability and Bayes formula We ask the following question: suppose we know that a certain event B has occurred. How does this impact the probability of some other A. This question

### Basic Proof Techniques

Basic Proof Techniques David Ferry dsf43@truman.edu September 13, 010 1 Four Fundamental Proof Techniques When one wishes to prove the statement P Q there are four fundamental approaches. This document

### not to be republishe NCERT SETS Chapter Introduction 1.2 Sets and their Representations

SETS Chapter 1 In these days of conflict between ancient and modern studies; there must surely be something to be said for a study which did not begin with Pythagoras and will not end with Einstein; but

### FEGYVERNEKI SÁNDOR, PROBABILITY THEORY AND MATHEmATICAL

FEGYVERNEKI SÁNDOR, PROBABILITY THEORY AND MATHEmATICAL STATIsTICs 4 IV. RANDOm VECTORs 1. JOINTLY DIsTRIBUTED RANDOm VARIABLEs If are two rom variables defined on the same sample space we define the joint

### Math 4310 Handout - Quotient Vector Spaces

Math 4310 Handout - Quotient Vector Spaces Dan Collins The textbook defines a subspace of a vector space in Chapter 4, but it avoids ever discussing the notion of a quotient space. This is understandable

### Discrete Math in Computer Science Homework 7 Solutions (Max Points: 80)

Discrete Math in Computer Science Homework 7 Solutions (Max Points: 80) CS 30, Winter 2016 by Prasad Jayanti 1. (10 points) Here is the famous Monty Hall Puzzle. Suppose you are on a game show, and you

### How to Conduct a Hypothesis Test

How to Conduct a Hypothesis Test The idea of hypothesis testing is relatively straightforward. In various studies we observe certain events. We must ask, is the event due to chance alone, or is there some

### Graph Theory Problems and Solutions

raph Theory Problems and Solutions Tom Davis tomrdavis@earthlink.net http://www.geometer.org/mathcircles November, 005 Problems. Prove that the sum of the degrees of the vertices of any finite graph is

### Colored Hats and Logic Puzzles

Colored Hats and Logic Puzzles Alex Zorn January 21, 2013 1 Introduction In this talk we ll discuss a collection of logic puzzles/games in which a number of people are given colored hats, and they try

### ELEMENTARY PROBABILITY

ELEMENTARY PROBABILITY Events and event sets. Consider tossing a die. There are six possible outcomes, which we shall denote by elements of the set {A i ; i =1, 2,...,6}. A numerical value is assigned

### Joint distributions Math 217 Probability and Statistics Prof. D. Joyce, Fall 2014

Joint distributions Math 17 Probability and Statistics Prof. D. Joyce, Fall 14 Today we ll look at joint random variables and joint distributions in detail. Product distributions. If Ω 1 and Ω are sample

### The Story. Probability - I. Plan. Example. A Probability Tree. Draw a probability tree.

Great Theoretical Ideas In Computer Science Victor Adamchi CS 5-5 Carnegie Mellon University Probability - I The Story The theory of probability was originally developed by a French mathematician Blaise

### MINITAB ASSISTANT WHITE PAPER

MINITAB ASSISTANT WHITE PAPER This paper explains the research conducted by Minitab statisticians to develop the methods and data checks used in the Assistant in Minitab 17 Statistical Software. One-Way

### + Section 6.2 and 6.3

Section 6.2 and 6.3 Learning Objectives After this section, you should be able to DEFINE and APPLY basic rules of probability CONSTRUCT Venn diagrams and DETERMINE probabilities DETERMINE probabilities

### Testing LTL Formula Translation into Büchi Automata

Testing LTL Formula Translation into Büchi Automata Heikki Tauriainen and Keijo Heljanko Helsinki University of Technology, Laboratory for Theoretical Computer Science, P. O. Box 5400, FIN-02015 HUT, Finland

### Page 331, 38.4 Suppose a is a positive integer and p is a prime. Prove that p a if and only if the prime factorization of a contains p.

Page 331, 38.2 Assignment #11 Solutions Factor the following positive integers into primes. a. 25 = 5 2. b. 4200 = 2 3 3 5 2 7. c. 10 10 = 2 10 5 10. d. 19 = 19. e. 1 = 1. Page 331, 38.4 Suppose a is a

### Bayesian Tutorial (Sheet Updated 20 March)

Bayesian Tutorial (Sheet Updated 20 March) Practice Questions (for discussing in Class) Week starting 21 March 2016 1. What is the probability that the total of two dice will be greater than 8, given that

### Unit 19: Probability Models

Unit 19: Probability Models Summary of Video Probability is the language of uncertainty. Using statistics, we can better predict the outcomes of random phenomena over the long term from the very complex,

### 3. Mathematical Induction

3. MATHEMATICAL INDUCTION 83 3. Mathematical Induction 3.1. First Principle of Mathematical Induction. Let P (n) be a predicate with domain of discourse (over) the natural numbers N = {0, 1,,...}. If (1)

### TYPES OF NUMBERS. Example 2. Example 1. Problems. Answers

TYPES OF NUMBERS When two or more integers are multiplied together, each number is a factor of the product. Nonnegative integers that have exactly two factors, namely, one and itself, are called prime

### Mathematical induction. Niloufar Shafiei

Mathematical induction Niloufar Shafiei Mathematical induction Mathematical induction is an extremely important proof technique. Mathematical induction can be used to prove results about complexity of

### 1. Prove that the empty set is a subset of every set.

1. Prove that the empty set is a subset of every set. Basic Topology Written by Men-Gen Tsai email: b89902089@ntu.edu.tw Proof: For any element x of the empty set, x is also an element of every set since

### Avoiding the Consolation Prize: The Mathematics of Game Shows

Avoiding the Consolation Prize: The Mathematics of Game Shows STUART GLUCK, Ph.D. CENTER FOR TALENTED YOUTH JOHNS HOPKINS UNIVERSITY STU@JHU.EDU CARLOS RODRIGUEZ CENTER FOR TALENTED YOUTH JOHNS HOPKINS

### Conditional Probability, Hypothesis Testing, and the Monty Hall Problem

Conditional Probability, Hypothesis Testing, and the Monty Hall Problem Ernie Croot September 17, 2008 On more than one occasion I have heard the comment Probability does not exist in the real world, and

### Statistical Inference. Prof. Kate Calder. If the coin is fair (chance of heads = chance of tails) then

Probability Statistical Inference Question: How often would this method give the correct answer if I used it many times? Answer: Use laws of probability. 1 Example: Tossing a coin If the coin is fair (chance

### MATH 289 PROBLEM SET 4: NUMBER THEORY

MATH 289 PROBLEM SET 4: NUMBER THEORY 1. The greatest common divisor If d and n are integers, then we say that d divides n if and only if there exists an integer q such that n = qd. Notice that if d divides

### CHAPTER 3. Methods of Proofs. 1. Logical Arguments and Formal Proofs

CHAPTER 3 Methods of Proofs 1. Logical Arguments and Formal Proofs 1.1. Basic Terminology. An axiom is a statement that is given to be true. A rule of inference is a logical rule that is used to deduce

### Basic Probability Theory (I)

Basic Probability Theory (I) Intro to Bayesian Data Analysis & Cognitive Modeling Adrian Brasoveanu [partly based on slides by Sharon Goldwater & Frank Keller and John K. Kruschke] Fall 2012 UCSC Linguistics

### 2. Methods of Proof Types of Proofs. Suppose we wish to prove an implication p q. Here are some strategies we have available to try.

2. METHODS OF PROOF 69 2. Methods of Proof 2.1. Types of Proofs. Suppose we wish to prove an implication p q. Here are some strategies we have available to try. Trivial Proof: If we know q is true then

### Copyright 2011 Casa Software Ltd. www.casaxps.com. Centre of Mass

Centre of Mass A central theme in mathematical modelling is that of reducing complex problems to simpler, and hopefully, equivalent problems for which mathematical analysis is possible. The concept of