Some preliminary Q&A

Size: px
Start display at page:

Download "Some preliminary Q&A"

Transcription

1 Some preliminary Q&A What is the philosophical difference between classical ( frequentist ) and Bayesian statistics? Chapter 1: Approaches for Statistical Inference p. 1/19

2 Some preliminary Q&A What is the philosophical difference between classical ( frequentist ) and Bayesian statistics? To a frequentist, unknown model parameters are fixed and unknown, and only estimable by replications of data from some experiment. Chapter 1: Approaches for Statistical Inference p. 1/19

3 Some preliminary Q&A What is the philosophical difference between classical ( frequentist ) and Bayesian statistics? To a frequentist, unknown model parameters are fixed and unknown, and only estimable by replications of data from some experiment. A Bayesian thinks of parameters as random, and thus having distributions (just like the data). We can thus think about unknowns for which no reliable frequentist experiment exists, e.g. θ = proportion of US men with untreated atrial fibrillation Chapter 1: Approaches for Statistical Inference p. 1/19

4 Some preliminary Q&A How does it work? Chapter 1: Approaches for Statistical Inference p. 2/19

5 Some preliminary Q&A How does it work? A Bayesian writes down a prior guess for θ, p(θ), then combines this with the information that the data X provide to obtain the posterior distribution of θ, p(θ X). All statistical inferences (point and interval estimates, hypothesis tests) then follow as appropriate summaries of the posterior. Chapter 1: Approaches for Statistical Inference p. 2/19

6 Some preliminary Q&A How does it work? A Bayesian writes down a prior guess for θ, p(θ), then combines this with the information that the data X provide to obtain the posterior distribution of θ, p(θ X). All statistical inferences (point and interval estimates, hypothesis tests) then follow as appropriate summaries of the posterior. Note that posterior information prior information 0, with the second replaced by = only if the prior is noninformative (which is often uniform, or flat ). Chapter 1: Approaches for Statistical Inference p. 2/19

7 Some preliminary Q&A Is the classical approach wrong? Chapter 1: Approaches for Statistical Inference p. 3/19

8 Some preliminary Q&A Is the classical approach wrong? While a hardcore Bayesian might say so, it is probably more accurate to think of classical methods as merely limited in scope! Chapter 1: Approaches for Statistical Inference p. 3/19

9 Some preliminary Q&A Is the classical approach wrong? While a hardcore Bayesian might say so, it is probably more accurate to think of classical methods as merely limited in scope! The Bayesian approach expands the class of models we can fit to our data, enabling us to handle repeated measures unbalanced or missing data nonhomogenous variances multivariate data and many other settings that are awkward or infeasible from a classical point of view. Chapter 1: Approaches for Statistical Inference p. 3/19

10 Some preliminary Q&A Is the classical approach wrong? While a hardcore Bayesian might say so, it is probably more accurate to think of classical methods as merely limited in scope! The Bayesian approach expands the class of models we can fit to our data, enabling us to handle repeated measures unbalanced or missing data nonhomogenous variances multivariate data and many other settings that are awkward or infeasible from a classical point of view. The approach also eases the interpretation of and learning from those models once fit. Chapter 1: Approaches for Statistical Inference p. 3/19

11 Simple example of Bayesian thinking From Business Week, online edition, July 31, 2001: Chapter 1: Approaches for Statistical Inference p. 4/19

12 Simple example of Bayesian thinking From Business Week, online edition, July 31, 2001: Economists might note, to take a simple example, that American turkey consumption tends to increase in November. A Bayesian would clarify this by observing that Thanksgiving occurs in this month. Chapter 1: Approaches for Statistical Inference p. 4/19

13 Simple example of Bayesian thinking From Business Week, online edition, July 31, 2001: Economists might note, to take a simple example, that American turkey consumption tends to increase in November. A Bayesian would clarify this by observing that Thanksgiving occurs in this month. Data: plot of turkey consumption by month Chapter 1: Approaches for Statistical Inference p. 4/19

14 Simple example of Bayesian thinking From Business Week, online edition, July 31, 2001: Economists might note, to take a simple example, that American turkey consumption tends to increase in November. A Bayesian would clarify this by observing that Thanksgiving occurs in this month. Data: plot of turkey consumption by month Prior: location of Thanksgiving in the calendar knowledge of Americans Thanksgiving eating habits Chapter 1: Approaches for Statistical Inference p. 4/19

15 Simple example of Bayesian thinking From Business Week, online edition, July 31, 2001: Economists might note, to take a simple example, that American turkey consumption tends to increase in November. A Bayesian would clarify this by observing that Thanksgiving occurs in this month. Data: plot of turkey consumption by month Prior: location of Thanksgiving in the calendar knowledge of Americans Thanksgiving eating habits Posterior: Understanding of the pattern in the data! Chapter 1: Approaches for Statistical Inference p. 4/19

16 Bayes means revision of estimates Humans tend to be Bayesian in the sense that most revise their opinions about uncertain quantities as more data accumulate. For example: Suppose you are about to make your first submission to a particular academic journal Chapter 1: Approaches for Statistical Inference p. 5/19

17 Bayes means revision of estimates Humans tend to be Bayesian in the sense that most revise their opinions about uncertain quantities as more data accumulate. For example: Suppose you are about to make your first submission to a particular academic journal You assess your chances of your paper being accepted (you have an opinion, but the true probability is unknown) Chapter 1: Approaches for Statistical Inference p. 5/19

18 Bayes means revision of estimates Humans tend to be Bayesian in the sense that most revise their opinions about uncertain quantities as more data accumulate. For example: Suppose you are about to make your first submission to a particular academic journal You assess your chances of your paper being accepted (you have an opinion, but the true probability is unknown) You submit your article and it is accepted! Chapter 1: Approaches for Statistical Inference p. 5/19

19 Bayes means revision of estimates Humans tend to be Bayesian in the sense that most revise their opinions about uncertain quantities as more data accumulate. For example: Suppose you are about to make your first submission to a particular academic journal You assess your chances of your paper being accepted (you have an opinion, but the true probability is unknown) You submit your article and it is accepted! Question: What is your revised opinion regarding the acceptance probability for papers like yours? Chapter 1: Approaches for Statistical Inference p. 5/19

20 Bayes means revision of estimates Humans tend to be Bayesian in the sense that most revise their opinions about uncertain quantities as more data accumulate. For example: Suppose you are about to make your first submission to a particular academic journal You assess your chances of your paper being accepted (you have an opinion, but the true probability is unknown) You submit your article and it is accepted! Question: What is your revised opinion regarding the acceptance probability for papers like yours? If you said anything other than 1, you are a Bayesian! Chapter 1: Approaches for Statistical Inference p. 5/19

21 Bayes can account for structure County-level breast cancer rates per 10,000 women: With no direct data for, what estimate would you use? Chapter 1: Approaches for Statistical Inference p. 6/19

22 Bayes can account for structure County-level breast cancer rates per 10,000 women: With no direct data for, what estimate would you use? Is 200 reasonable? Chapter 1: Approaches for Statistical Inference p. 6/19

23 Bayes can account for structure County-level breast cancer rates per 10,000 women: With no direct data for, what estimate would you use? Is 200 reasonable? Probably not: all the other rates are around 100 Chapter 1: Approaches for Statistical Inference p. 6/19

24 Bayes can account for structure County-level breast cancer rates per 10,000 women: With no direct data for, what estimate would you use? Is 200 reasonable? Probably not: all the other rates are around 100 Perhaps use the average of the neighboring values (again, near 100) Chapter 1: Approaches for Statistical Inference p. 6/19

25 Accounting for structure (cont d) Now assume that data become available for county : 100 women at risk, 2 cancer cases. Thus rate = , 000 = 200 Would you use this value as the estimate? Chapter 1: Approaches for Statistical Inference p. 7/19

26 Accounting for structure (cont d) Now assume that data become available for county : 100 women at risk, 2 cancer cases. Thus rate = , 000 = 200 Would you use this value as the estimate? Probably not: The sample size is very small, so this estimate will be unreliable. How about a compromise between 200 and the rates in the neighboring counties? Chapter 1: Approaches for Statistical Inference p. 7/19

27 Accounting for structure (cont d) Now assume that data become available for county : 100 women at risk, 2 cancer cases. Thus rate = , 000 = 200 Would you use this value as the estimate? Probably not: The sample size is very small, so this estimate will be unreliable. How about a compromise between 200 and the rates in the neighboring counties? Now repeat this thought experiment if the county data were 20/1000, 200/10000,... Chapter 1: Approaches for Statistical Inference p. 7/19

28 Accounting for structure (cont d) Now assume that data become available for county : 100 women at risk, 2 cancer cases. Thus rate = , 000 = 200 Would you use this value as the estimate? Probably not: The sample size is very small, so this estimate will be unreliable. How about a compromise between 200 and the rates in the neighboring counties? Now repeat this thought experiment if the county data were 20/1000, 200/10000,... Bayes and empirical Bayes methods can incorporate the structure in the data, weight the data and prior information appropriately, and allow the data to dominate as the sample size becomes large. Chapter 1: Approaches for Statistical Inference p. 7/19

29 Motivating Example From Berger and Berry (1988, Amer. Scientist): Consider a clinical trial to study the effectiveness of Vitamin C in treating the common cold. Chapter 1: Approaches for Statistical Inference p. 8/19

30 Motivating Example From Berger and Berry (1988, Amer. Scientist): Consider a clinical trial to study the effectiveness of Vitamin C in treating the common cold. Observations are matched pairs of subjects (twins?), half randomized (in double blind fashion) to vitamin C, half to placebo. We count how many pairs had C giving superior relief after 48 hours. Chapter 1: Approaches for Statistical Inference p. 8/19

31 Two Designs Design #1: Sample n = 17 pairs, and test H 0 : P(C better) = 1 2 vs. H A : P(C better) 1 2 Suppose we observe x = 13 preferences for C. Then p-value = P(X 13 or X 4) =.049 So if α =.05, stop and reject H 0. Chapter 1: Approaches for Statistical Inference p. 9/19

32 Two Designs Design #1: Sample n = 17 pairs, and test H 0 : P(C better) = 1 2 vs. H A : P(C better) 1 2 Suppose we observe x = 13 preferences for C. Then p-value = P(X 13 or X 4) =.049 So if α =.05, stop and reject H 0. Design #2: Sample n 1 = 17 pairs. Then: if x 1 13 or x 1 4, stop. otherwise, sample an additional n 2 = 27 pairs. Reject H 0 if X 1 + X 2 29 or X 1 + X Chapter 1: Approaches for Statistical Inference p. 9/19

33 Two Designs (cont d) We choose this second stage since under H 0, P(X 1 + X 2 29 or X 1 + X 2 15) =.049 the same as Stage 1! Chapter 1: Approaches for Statistical Inference p. 10/19

34 Two Designs (cont d) We choose this second stage since under H 0, P(X 1 + X 2 29 or X 1 + X 2 15) =.049 the same as Stage 1! Suppose we again observe X 1 = 13. Now: p-value = P(X 1 13 or X 1 4) +P(X 1 + X 2 29 and 4 < X 1 < 13) +P(X 1 + X 2 15 and 4 < X 1 < 13) =.085 no longer significant at α =.05! Chapter 1: Approaches for Statistical Inference p. 10/19

35 Two Designs (cont d) We choose this second stage since under H 0, P(X 1 + X 2 29 or X 1 + X 2 15) =.049 the same as Stage 1! Suppose we again observe X 1 = 13. Now: p-value = P(X 1 13 or X 1 4) +P(X 1 + X 2 29 and 4 < X 1 < 13) +P(X 1 + X 2 15 and 4 < X 1 < 13) =.085 no longer significant at α =.05! Yet the observed data was exactly the same; all we did was contemplate a second stage (no effect on data), and it changed our answer! Chapter 1: Approaches for Statistical Inference p. 10/19

36 Additional Q & A Q: What if we kept adding stages? Chapter 1: Approaches for Statistical Inference p. 11/19

37 Additional Q & A Q: What if we kept adding stages? A: p-value 1, even though x 1 still 13! Chapter 1: Approaches for Statistical Inference p. 11/19

38 Additional Q & A Q: What if we kept adding stages? A: p-value 1, even though x 1 still 13! Q: So are p-values really objective evidence? Chapter 1: Approaches for Statistical Inference p. 11/19

39 Additional Q & A Q: What if we kept adding stages? A: p-value 1, even though x 1 still 13! Q: So are p-values really objective evidence? A: No, since extra info (like design) critical! Chapter 1: Approaches for Statistical Inference p. 11/19

40 Additional Q & A Q: What if we kept adding stages? A: p-value 1, even though x 1 still 13! Q: So are p-values really objective evidence? A: No, since extra info (like design) critical! Q: What about unforeseen events? Example: First 5 patients develop an allergic reaction to the treatment trial is stopped by clinicians. Can a frequentist analyze these data? Chapter 1: Approaches for Statistical Inference p. 11/19

41 Additional Q & A Q: What if we kept adding stages? A: p-value 1, even though x 1 still 13! Q: So are p-values really objective evidence? A: No, since extra info (like design) critical! Q: What about unforeseen events? Example: First 5 patients develop an allergic reaction to the treatment trial is stopped by clinicians. Can a frequentist analyze these data? A: No: This aspect of design wasn t anticipated, so p-values not computable! Chapter 1: Approaches for Statistical Inference p. 11/19

42 Additional Q & A Q: What if we kept adding stages? A: p-value 1, even though x 1 still 13! Q: So are p-values really objective evidence? A: No, since extra info (like design) critical! Q: What about unforeseen events? Example: First 5 patients develop an allergic reaction to the treatment trial is stopped by clinicians. Can a frequentist analyze these data? A: No: This aspect of design wasn t anticipated, so p-values not computable! Q: Can a Bayesian? Chapter 1: Approaches for Statistical Inference p. 11/19

43 Additional Q & A Q: What if we kept adding stages? A: p-value 1, even though x 1 still 13! Q: So are p-values really objective evidence? A: No, since extra info (like design) critical! Q: What about unforeseen events? Example: First 5 patients develop an allergic reaction to the treatment trial is stopped by clinicians. Can a frequentist analyze these data? A: No: This aspect of design wasn t anticipated, so p-values not computable! Q: Can a Bayesian? A: (obviously) Yes as we shall see... Chapter 1: Approaches for Statistical Inference p. 11/19

44 Bayes/frequentist controversy Frequentist outlook: Suppose we have k unknown quantities θ = (θ 1,θ 2,...,θ k ), and data X = (X 1,...,X n ) with distribution which depends on θ, say p(x θ). Chapter 1: Approaches for Statistical Inference p. 12/19

45 Bayes/frequentist controversy Frequentist outlook: Suppose we have k unknown quantities θ = (θ 1,θ 2,...,θ k ), and data X = (X 1,...,X n ) with distribution which depends on θ, say p(x θ). The frequentist selects a loss function, L(θ, a), and develops procedures which perform well with respect to frequentist risk, R(θ,δ) = E X θ [L(θ,a)] Chapter 1: Approaches for Statistical Inference p. 12/19

46 Bayes/frequentist controversy Frequentist outlook: Suppose we have k unknown quantities θ = (θ 1,θ 2,...,θ k ), and data X = (X 1,...,X n ) with distribution which depends on θ, say p(x θ). The frequentist selects a loss function, L(θ, a), and develops procedures which perform well with respect to frequentist risk, R(θ,δ) = E X θ [L(θ,a)] A good frequentist procedure is one that enjoys low loss over repeated data sampling regardless of the true value of θ. Chapter 1: Approaches for Statistical Inference p. 12/19

47 Bayes/frequentist controversy So are there any such procedures? Sure: Example: Suppose X i θ iid N(θ,σ 2 ), i = 1,...,n. Standard 95% frequentist confidence interval (CI) is δ(x) = ( x ± 1.96 s/ n). Chapter 1: Approaches for Statistical Inference p. 13/19

48 Bayes/frequentist controversy So are there any such procedures? Sure: Example: Suppose X i θ iid N(θ,σ 2 ), i = 1,...,n. Standard 95% frequentist confidence interval (CI) is δ(x) = ( x ± 1.96 s/ n). If we choose the loss function { 0 if θ δ(x) L(θ,δ) = 1 if θ δ(x), then R[(θ,σ),δ] = E x θ,σ [L(θ,δ(x))] = P θ,σ (θ δ(x)) =.05 Chapter 1: Approaches for Statistical Inference p. 13/19

49 Bayes/frequentist controversy So are there any such procedures? Sure: Example: Suppose X i θ iid N(θ,σ 2 ), i = 1,...,n. Standard 95% frequentist confidence interval (CI) is δ(x) = ( x ± 1.96 s/ n). If we choose the loss function { 0 if θ δ(x) L(θ,δ) = 1 if θ δ(x), then R[(θ,σ),δ] = E x θ,σ [L(θ,δ(x))] = P θ,σ (θ δ(x)) =.05 On average in repeated use, δ fails only 5% of the time. Chapter 1: Approaches for Statistical Inference p. 13/19

50 Bayes/frequentist controversy So are there any such procedures? Sure: Example: Suppose X i θ iid N(θ,σ 2 ), i = 1,...,n. Standard 95% frequentist confidence interval (CI) is δ(x) = ( x ± 1.96 s/ n). If we choose the loss function { 0 if θ δ(x) L(θ,δ) = 1 if θ δ(x), then R[(θ,σ),δ] = E x θ,σ [L(θ,δ(x))] = P θ,σ (θ δ(x)) =.05 On average in repeated use, δ fails only 5% of the time. Has some appeal, since it works for all θ and σ! Chapter 1: Approaches for Statistical Inference p. 13/19

51 Bayes/frequentist controversy Many frequentist analyses based on the likelihood function, which is just p(x θ) viewed as a function of θ: L(θ;x) p(x θ) Larger p(x θ) for values of θ which are more likely. Leads naturally to... Chapter 1: Approaches for Statistical Inference p. 14/19

52 Bayes/frequentist controversy Many frequentist analyses based on the likelihood function, which is just p(x θ) viewed as a function of θ: L(θ;x) p(x θ) Larger p(x θ) for values of θ which are more likely. Leads naturally to... The Likelihood Principle: In making inferences or decisions about θ after x is observed, all relevant experimental information is contained in the likelihood function for the observed x. Furthermore, two likelihood functions contain the same information about θ if they are proportional to each other as functions of θ. Chapter 1: Approaches for Statistical Inference p. 14/19

53 Bayes/frequentist controversy Many frequentist analyses based on the likelihood function, which is just p(x θ) viewed as a function of θ: L(θ;x) p(x θ) Larger p(x θ) for values of θ which are more likely. Leads naturally to... The Likelihood Principle: In making inferences or decisions about θ after x is observed, all relevant experimental information is contained in the likelihood function for the observed x. Furthermore, two likelihood functions contain the same information about θ if they are proportional to each other as functions of θ. Seems completely reasonable (and theoretically justifiable) yet frequentist analyses may violate!... Chapter 1: Approaches for Statistical Inference p. 14/19

54 Binomial vs. Negative Binomial Example due to Pratt (comment on Birnbaum, 1962 JASA): Suppose 12 independent coin tosses: 9H, 3T. Test: H 0 : θ = 1 2 vs. H A : θ > 1 2 Two possibilities for f(x θ): Chapter 1: Approaches for Statistical Inference p. 15/19

55 Binomial vs. Negative Binomial Example due to Pratt (comment on Birnbaum, 1962 JASA): Suppose 12 independent coin tosses: 9H, 3T. Test: H 0 : θ = 1 2 vs. H A : θ > 1 2 Two possibilities for f(x θ): Binomial: n = 12 tosses (fixed beforehand) X = #H Bin(12,θ) L 1 (θ) = p 1 (x θ) = ( n x = ( 12 9 ) θ 9 (1 θ) 3. ) θ x (1 θ) n x Chapter 1: Approaches for Statistical Inference p. 15/19

56 Binomial vs. Negative Binomial Example due to Pratt (comment on Birnbaum, 1962 JASA): Suppose 12 independent coin tosses: 9H, 3T. Test: H 0 : θ = 1 2 vs. H A : θ > 1 2 Two possibilities for f(x θ): Binomial: n = 12 tosses (fixed beforehand) X = #H Bin(12,θ) L 1 (θ) = p 1 (x θ) = ( n x ) θ x (1 θ) n x = ( 12) 9 θ 9 (1 θ) 3. Negative Binomial: Flip until we get r = 3 tails X NB(3,θ) L 2 (θ) = p 2 (x θ) = ( r+x 1 x = ( 11 9 ) θ 9 (1 θ) 3. ) θ x (1 θ) r Chapter 1: Approaches for Statistical Inference p. 15/19

57 Binomial vs. Negative Binomial Example due to Pratt (comment on Birnbaum, 1962 JASA): Suppose 12 independent coin tosses: 9H, 3T. Test: H 0 : θ = 1 2 vs. H A : θ > 1 2 Two possibilities for f(x θ): Binomial: n = 12 tosses (fixed beforehand) X = #H Bin(12,θ) L 1 (θ) = p 1 (x θ) = ( n x ) θ x (1 θ) n x = ( 12) 9 θ 9 (1 θ) 3. Negative Binomial: Flip until we get r = 3 tails X NB(3,θ) L 2 (θ) = p 2 (x θ) = ( r+x 1 x ) θ x (1 θ) r = ( 11 9 ) θ 9 (1 θ) 3. Adopt the rejection region, Reject H 0 if X c. Chapter 1: Approaches for Statistical Inference p. 15/19

58 Binomial vs. Negative Binomial p-values: Chapter 1: Approaches for Statistical Inference p. 16/19

59 Binomial vs. Negative Binomial p-values: α 1 = P θ= 1(X 9) = Σ 12 ( 12 ) 2 j=9 j θ j (1 θ) 12 j =.075 Chapter 1: Approaches for Statistical Inference p. 16/19

60 Binomial vs. Negative Binomial p-values: α 1 = P θ= 1(X 9) = Σ 12 2 ( 12 j=9 j α 2 = P θ= 1 9) = Σ 2(X j=9 ( 2+j j ) θ j (1 θ) 12 j =.075 ) θ j (1 θ) 3 =.0325 Chapter 1: Approaches for Statistical Inference p. 16/19

61 Binomial vs. Negative Binomial p-values: α 1 = P θ= 1(X 9) = Σ 12 2 ( 12 j=9 j α 2 = P θ= 1 9) = Σ 2(X j=9 ( 2+j j ) θ j (1 θ) 12 j =.075 ) θ j (1 θ) 3 =.0325 So at α =.05, two different decisions! Violates the Likelihood Principle, since L 1 (θ) L 2 (θ)!! Chapter 1: Approaches for Statistical Inference p. 16/19

62 Binomial vs. Negative Binomial p-values: α 1 = P θ= 1(X 9) = Σ 12 2 ( 12 j=9 j α 2 = P θ= 1 9) = Σ 2(X j=9 ( 2+j j ) θ j (1 θ) 12 j =.075 ) θ j (1 θ) 3 =.0325 So at α =.05, two different decisions! Violates the Likelihood Principle, since L 1 (θ) L 2 (θ)!! What happened? Besides the observed x = 9, we also took into account the more extreme X 10. Chapter 1: Approaches for Statistical Inference p. 16/19

63 Binomial vs. Negative Binomial p-values: α 1 = P θ= 1(X 9) = Σ 12 2 ( 12 j=9 j α 2 = P θ= 1 9) = Σ 2(X j=9 ( 2+j j ) θ j (1 θ) 12 j =.075 ) θ j (1 θ) 3 =.0325 So at α =.05, two different decisions! Violates the Likelihood Principle, since L 1 (θ) L 2 (θ)!! What happened? Besides the observed x = 9, we also took into account the more extreme X 10. Jeffreys (1961):...a hypothesis which may be true may be rejected because it has not predicted observable results which have not occurred. Chapter 1: Approaches for Statistical Inference p. 16/19

64 Binomial vs. Negative Binomial p-values: α 1 = P θ= 1(X 9) = Σ 12 2 ( 12 j=9 j α 2 = P θ= 1 9) = Σ 2(X j=9 ( 2+j j ) θ j (1 θ) 12 j =.075 ) θ j (1 θ) 3 =.0325 So at α =.05, two different decisions! Violates the Likelihood Principle, since L 1 (θ) L 2 (θ)!! What happened? Besides the observed x = 9, we also took into account the more extreme X 10. Jeffreys (1961):...a hypothesis which may be true may be rejected because it has not predicted observable results which have not occurred. In our example, the probability of the unpredicted and nonoccurring set X 10 has been used as evidence against H 0! Chapter 1: Approaches for Statistical Inference p. 16/19

65 Conditional (Bayesian) Perspective Always condition on data which has actually occurred; the long-run performance of a procedure is of (at most) secondary interest. Fix a prior distribution p(θ), and use Bayes Theorem (1763): p(θ x) p(x θ)p(θ) ( posterior likelihood prior ) Chapter 1: Approaches for Statistical Inference p. 17/19

66 Conditional (Bayesian) Perspective Always condition on data which has actually occurred; the long-run performance of a procedure is of (at most) secondary interest. Fix a prior distribution p(θ), and use Bayes Theorem (1763): p(θ x) p(x θ)p(θ) ( posterior likelihood prior ) Indeed, it often turns out that using the Bayesian formalism with relatively vague priors produces procedures which perform well using traditional frequentist criteria (e.g., low mean squared error over repeated sampling)! Chapter 1: Approaches for Statistical Inference p. 17/19

67 Conditional (Bayesian) Perspective Always condition on data which has actually occurred; the long-run performance of a procedure is of (at most) secondary interest. Fix a prior distribution p(θ), and use Bayes Theorem (1763): p(θ x) p(x θ)p(θ) ( posterior likelihood prior ) Indeed, it often turns out that using the Bayesian formalism with relatively vague priors produces procedures which perform well using traditional frequentist criteria (e.g., low mean squared error over repeated sampling)! several examples in Chapter 4 of the C&L text! Chapter 1: Approaches for Statistical Inference p. 17/19

68 Frequentist criticisms [from B. Efron (1986, Amer. Statist.), Why isn t everyone a Bayesian? ] Shouldn t condition on x (but this renews conflict with the Likelihood Principle) Chapter 1: Approaches for Statistical Inference p. 18/19

69 Frequentist criticisms [from B. Efron (1986, Amer. Statist.), Why isn t everyone a Bayesian? ] Shouldn t condition on x (but this renews conflict with the Likelihood Principle) Not easy/automatic (but computing keeps improving: MCMC methods, WinBUGS software...) Chapter 1: Approaches for Statistical Inference p. 18/19

70 Frequentist criticisms [from B. Efron (1986, Amer. Statist.), Why isn t everyone a Bayesian? ] Shouldn t condition on x (but this renews conflict with the Likelihood Principle) Not easy/automatic (but computing keeps improving: MCMC methods, WinBUGS software...) How to pick the prior p(θ): Two experimenters could get different answers with the same data! How to control influence of the prior? How to get objective results (say, for a court case, scientific report,...)? Chapter 1: Approaches for Statistical Inference p. 18/19

71 Frequentist criticisms [from B. Efron (1986, Amer. Statist.), Why isn t everyone a Bayesian? ] Shouldn t condition on x (but this renews conflict with the Likelihood Principle) Not easy/automatic (but computing keeps improving: MCMC methods, WinBUGS software...) How to pick the prior p(θ): Two experimenters could get different answers with the same data! How to control influence of the prior? How to get objective results (say, for a court case, scientific report,...)? Clearly this final criticism is the most serious! Chapter 1: Approaches for Statistical Inference p. 18/19

72 Bayesian Advantages in Inference Ability to formally incorporate prior information Chapter 1: Approaches for Statistical Inference p. 19/19

73 Bayesian Advantages in Inference Ability to formally incorporate prior information The reason for stopping experimentation does not affect the inference Chapter 1: Approaches for Statistical Inference p. 19/19

74 Bayesian Advantages in Inference Ability to formally incorporate prior information The reason for stopping experimentation does not affect the inference Answers are more easily interpretable by nonspecialists (e.g. confidence intervals) Chapter 1: Approaches for Statistical Inference p. 19/19

75 Bayesian Advantages in Inference Ability to formally incorporate prior information The reason for stopping experimentation does not affect the inference Answers are more easily interpretable by nonspecialists (e.g. confidence intervals) All analyses follow directly from the posterior; no separate theories of estimation, testing, multiple comparisons, etc. are needed Chapter 1: Approaches for Statistical Inference p. 19/19

76 Bayesian Advantages in Inference Ability to formally incorporate prior information The reason for stopping experimentation does not affect the inference Answers are more easily interpretable by nonspecialists (e.g. confidence intervals) All analyses follow directly from the posterior; no separate theories of estimation, testing, multiple comparisons, etc. are needed Any question can be directly answered (bioequivalence, multiple comparisons/hypotheses,...) Chapter 1: Approaches for Statistical Inference p. 19/19

77 Bayesian Advantages in Inference Ability to formally incorporate prior information The reason for stopping experimentation does not affect the inference Answers are more easily interpretable by nonspecialists (e.g. confidence intervals) All analyses follow directly from the posterior; no separate theories of estimation, testing, multiple comparisons, etc. are needed Any question can be directly answered (bioequivalence, multiple comparisons/hypotheses,...) Inferences are conditional on the actual data Chapter 1: Approaches for Statistical Inference p. 19/19

78 Bayesian Advantages in Inference Ability to formally incorporate prior information The reason for stopping experimentation does not affect the inference Answers are more easily interpretable by nonspecialists (e.g. confidence intervals) All analyses follow directly from the posterior; no separate theories of estimation, testing, multiple comparisons, etc. are needed Any question can be directly answered (bioequivalence, multiple comparisons/hypotheses,...) Inferences are conditional on the actual data Bayes procedures possess many optimality properties (e.g. consistent, impose parsimony in model choice, define the class of optimal frequentist procedures,...) Chapter 1: Approaches for Statistical Inference p. 19/19

Comparison of frequentist and Bayesian inference. Class 20, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom

Comparison of frequentist and Bayesian inference. Class 20, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom Comparison of frequentist and Bayesian inference. Class 20, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom 1 Learning Goals 1. Be able to explain the difference between the p-value and a posterior

More information

Likelihood: Frequentist vs Bayesian Reasoning

Likelihood: Frequentist vs Bayesian Reasoning "PRINCIPLES OF PHYLOGENETICS: ECOLOGY AND EVOLUTION" Integrative Biology 200B University of California, Berkeley Spring 2009 N Hallinan Likelihood: Frequentist vs Bayesian Reasoning Stochastic odels and

More information

Basics of Statistical Machine Learning

Basics of Statistical Machine Learning CS761 Spring 2013 Advanced Machine Learning Basics of Statistical Machine Learning Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu Modern machine learning is rooted in statistics. You will find many familiar

More information

Foundations of Statistics Frequentist and Bayesian

Foundations of Statistics Frequentist and Bayesian Mary Parker, http://www.austincc.edu/mparker/stat/nov04/ page 1 of 13 Foundations of Statistics Frequentist and Bayesian Statistics is the science of information gathering, especially when the information

More information

LAB : THE CHI-SQUARE TEST. Probability, Random Chance, and Genetics

LAB : THE CHI-SQUARE TEST. Probability, Random Chance, and Genetics Period Date LAB : THE CHI-SQUARE TEST Probability, Random Chance, and Genetics Why do we study random chance and probability at the beginning of a unit on genetics? Genetics is the study of inheritance,

More information

AP: LAB 8: THE CHI-SQUARE TEST. Probability, Random Chance, and Genetics

AP: LAB 8: THE CHI-SQUARE TEST. Probability, Random Chance, and Genetics Ms. Foglia Date AP: LAB 8: THE CHI-SQUARE TEST Probability, Random Chance, and Genetics Why do we study random chance and probability at the beginning of a unit on genetics? Genetics is the study of inheritance,

More information

People have thought about, and defined, probability in different ways. important to note the consequences of the definition:

People have thought about, and defined, probability in different ways. important to note the consequences of the definition: PROBABILITY AND LIKELIHOOD, A BRIEF INTRODUCTION IN SUPPORT OF A COURSE ON MOLECULAR EVOLUTION (BIOL 3046) Probability The subject of PROBABILITY is a branch of mathematics dedicated to building models

More information

A Few Basics of Probability

A Few Basics of Probability A Few Basics of Probability Philosophy 57 Spring, 2004 1 Introduction This handout distinguishes between inductive and deductive logic, and then introduces probability, a concept essential to the study

More information

STAT 35A HW2 Solutions

STAT 35A HW2 Solutions STAT 35A HW2 Solutions http://www.stat.ucla.edu/~dinov/courses_students.dir/09/spring/stat35.dir 1. A computer consulting firm presently has bids out on three projects. Let A i = { awarded project i },

More information

Objections to Bayesian statistics

Objections to Bayesian statistics Bayesian Analysis (2008) 3, Number 3, pp. 445 450 Objections to Bayesian statistics Andrew Gelman Abstract. Bayesian inference is one of the more controversial approaches to statistics. The fundamental

More information

The result of the bayesian analysis is the probability distribution of every possible hypothesis H, given one real data set D. This prestatistical approach to our problem was the standard approach of Laplace

More information

Non Parametric Inference

Non Parametric Inference Maura Department of Economics and Finance Università Tor Vergata Outline 1 2 3 Inverse distribution function Theorem: Let U be a uniform random variable on (0, 1). Let X be a continuous random variable

More information

HYPOTHESIS TESTING (ONE SAMPLE) - CHAPTER 7 1. used confidence intervals to answer questions such as...

HYPOTHESIS TESTING (ONE SAMPLE) - CHAPTER 7 1. used confidence intervals to answer questions such as... HYPOTHESIS TESTING (ONE SAMPLE) - CHAPTER 7 1 PREVIOUSLY used confidence intervals to answer questions such as... You know that 0.25% of women have red/green color blindness. You conduct a study of men

More information

Bayesian Updating with Discrete Priors Class 11, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom

Bayesian Updating with Discrete Priors Class 11, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom 1 Learning Goals Bayesian Updating with Discrete Priors Class 11, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom 1. Be able to apply Bayes theorem to compute probabilities. 2. Be able to identify

More information

Online 12 - Sections 9.1 and 9.2-Doug Ensley

Online 12 - Sections 9.1 and 9.2-Doug Ensley Student: Date: Instructor: Doug Ensley Course: MAT117 01 Applied Statistics - Ensley Assignment: Online 12 - Sections 9.1 and 9.2 1. Does a P-value of 0.001 give strong evidence or not especially strong

More information

Fairfield Public Schools

Fairfield Public Schools Mathematics Fairfield Public Schools AP Statistics AP Statistics BOE Approved 04/08/2014 1 AP STATISTICS Critical Areas of Focus AP Statistics is a rigorous course that offers advanced students an opportunity

More information

Normal distribution. ) 2 /2σ. 2π σ

Normal distribution. ) 2 /2σ. 2π σ Normal distribution The normal distribution is the most widely known and used of all distributions. Because the normal distribution approximates many natural phenomena so well, it has developed into a

More information

Week 3&4: Z tables and the Sampling Distribution of X

Week 3&4: Z tables and the Sampling Distribution of X Week 3&4: Z tables and the Sampling Distribution of X 2 / 36 The Standard Normal Distribution, or Z Distribution, is the distribution of a random variable, Z N(0, 1 2 ). The distribution of any other normal

More information

The normal approximation to the binomial

The normal approximation to the binomial The normal approximation to the binomial The binomial probability function is not useful for calculating probabilities when the number of trials n is large, as it involves multiplying a potentially very

More information

Chapter 4. Probability and Probability Distributions

Chapter 4. Probability and Probability Distributions Chapter 4. robability and robability Distributions Importance of Knowing robability To know whether a sample is not identical to the population from which it was selected, it is necessary to assess the

More information

1 Prior Probability and Posterior Probability

1 Prior Probability and Posterior Probability Math 541: Statistical Theory II Bayesian Approach to Parameter Estimation Lecturer: Songfeng Zheng 1 Prior Probability and Posterior Probability Consider now a problem of statistical inference in which

More information

HYPOTHESIS TESTING (ONE SAMPLE) - CHAPTER 7 1. used confidence intervals to answer questions such as...

HYPOTHESIS TESTING (ONE SAMPLE) - CHAPTER 7 1. used confidence intervals to answer questions such as... HYPOTHESIS TESTING (ONE SAMPLE) - CHAPTER 7 1 PREVIOUSLY used confidence intervals to answer questions such as... You know that 0.25% of women have red/green color blindness. You conduct a study of men

More information

The Binomial Probability Distribution

The Binomial Probability Distribution The Binomial Probability Distribution MATH 130, Elements of Statistics I J. Robert Buchanan Department of Mathematics Fall 2015 Objectives After this lesson we will be able to: determine whether a probability

More information

Lecture 9: Bayesian hypothesis testing

Lecture 9: Bayesian hypothesis testing Lecture 9: Bayesian hypothesis testing 5 November 27 In this lecture we ll learn about Bayesian hypothesis testing. 1 Introduction to Bayesian hypothesis testing Before we go into the details of Bayesian

More information

What is a P-value? Ronald A. Thisted, PhD Departments of Statistics and Health Studies The University of Chicago

What is a P-value? Ronald A. Thisted, PhD Departments of Statistics and Health Studies The University of Chicago What is a P-value? Ronald A. Thisted, PhD Departments of Statistics and Health Studies The University of Chicago 8 June 1998, Corrections 14 February 2010 Abstract Results favoring one treatment over another

More information

Sample Size and Power in Clinical Trials

Sample Size and Power in Clinical Trials Sample Size and Power in Clinical Trials Version 1.0 May 011 1. Power of a Test. Factors affecting Power 3. Required Sample Size RELATED ISSUES 1. Effect Size. Test Statistics 3. Variation 4. Significance

More information

Week 4: Standard Error and Confidence Intervals

Week 4: Standard Error and Confidence Intervals Health Sciences M.Sc. Programme Applied Biostatistics Week 4: Standard Error and Confidence Intervals Sampling Most research data come from subjects we think of as samples drawn from a larger population.

More information

Hypothesis testing - Steps

Hypothesis testing - Steps Hypothesis testing - Steps Steps to do a two-tailed test of the hypothesis that β 1 0: 1. Set up the hypotheses: H 0 : β 1 = 0 H a : β 1 0. 2. Compute the test statistic: t = b 1 0 Std. error of b 1 =

More information

C. The null hypothesis is not rejected when the alternative hypothesis is true. A. population parameters.

C. The null hypothesis is not rejected when the alternative hypothesis is true. A. population parameters. Sample Multiple Choice Questions for the material since Midterm 2. Sample questions from Midterms and 2 are also representative of questions that may appear on the final exam.. A randomly selected sample

More information

Introduction to Analysis of Variance (ANOVA) Limitations of the t-test

Introduction to Analysis of Variance (ANOVA) Limitations of the t-test Introduction to Analysis of Variance (ANOVA) The Structural Model, The Summary Table, and the One- Way ANOVA Limitations of the t-test Although the t-test is commonly used, it has limitations Can only

More information

University of Chicago Graduate School of Business. Business 41000: Business Statistics

University of Chicago Graduate School of Business. Business 41000: Business Statistics Name: University of Chicago Graduate School of Business Business 41000: Business Statistics Special Notes: 1. This is a closed-book exam. You may use an 8 11 piece of paper for the formulas. 2. Throughout

More information

E3: PROBABILITY AND STATISTICS lecture notes

E3: PROBABILITY AND STATISTICS lecture notes E3: PROBABILITY AND STATISTICS lecture notes 2 Contents 1 PROBABILITY THEORY 7 1.1 Experiments and random events............................ 7 1.2 Certain event. Impossible event............................

More information

Part 2: One-parameter models

Part 2: One-parameter models Part 2: One-parameter models Bernoilli/binomial models Return to iid Y 1,...,Y n Bin(1, θ). The sampling model/likelihood is p(y 1,...,y n θ) =θ P y i (1 θ) n P y i When combined with a prior p(θ), Bayes

More information

Binomial Sampling and the Binomial Distribution

Binomial Sampling and the Binomial Distribution Binomial Sampling and the Binomial Distribution Characterized by two mutually exclusive events." Examples: GENERAL: {success or failure} {on or off} {head or tail} {zero or one} BIOLOGY: {dead or alive}

More information

Model-based Synthesis. Tony O Hagan

Model-based Synthesis. Tony O Hagan Model-based Synthesis Tony O Hagan Stochastic models Synthesising evidence through a statistical model 2 Evidence Synthesis (Session 3), Helsinki, 28/10/11 Graphical modelling The kinds of models that

More information

Hypothesis Testing for Beginners

Hypothesis Testing for Beginners Hypothesis Testing for Beginners Michele Piffer LSE August, 2011 Michele Piffer (LSE) Hypothesis Testing for Beginners August, 2011 1 / 53 One year ago a friend asked me to put down some easy-to-read notes

More information

QUANTITATIVE METHODS BIOLOGY FINAL HONOUR SCHOOL NON-PARAMETRIC TESTS

QUANTITATIVE METHODS BIOLOGY FINAL HONOUR SCHOOL NON-PARAMETRIC TESTS QUANTITATIVE METHODS BIOLOGY FINAL HONOUR SCHOOL NON-PARAMETRIC TESTS This booklet contains lecture notes for the nonparametric work in the QM course. This booklet may be online at http://users.ox.ac.uk/~grafen/qmnotes/index.html.

More information

SCHOOL OF HEALTH AND HUMAN SCIENCES DON T FORGET TO RECODE YOUR MISSING VALUES

SCHOOL OF HEALTH AND HUMAN SCIENCES DON T FORGET TO RECODE YOUR MISSING VALUES SCHOOL OF HEALTH AND HUMAN SCIENCES Using SPSS Topics addressed today: 1. Differences between groups 2. Graphing Use the s4data.sav file for the first part of this session. DON T FORGET TO RECODE YOUR

More information

Two Correlated Proportions (McNemar Test)

Two Correlated Proportions (McNemar Test) Chapter 50 Two Correlated Proportions (Mcemar Test) Introduction This procedure computes confidence intervals and hypothesis tests for the comparison of the marginal frequencies of two factors (each with

More information

SENSITIVITY ANALYSIS AND INFERENCE. Lecture 12

SENSITIVITY ANALYSIS AND INFERENCE. Lecture 12 This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this

More information

Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS. Part 3: Discrete Uniform Distribution Binomial Distribution

Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS. Part 3: Discrete Uniform Distribution Binomial Distribution Chapter 3: DISCRETE RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS Part 3: Discrete Uniform Distribution Binomial Distribution Sections 3-5, 3-6 Special discrete random variable distributions we will cover

More information

A POPULATION MEAN, CONFIDENCE INTERVALS AND HYPOTHESIS TESTING

A POPULATION MEAN, CONFIDENCE INTERVALS AND HYPOTHESIS TESTING CHAPTER 5. A POPULATION MEAN, CONFIDENCE INTERVALS AND HYPOTHESIS TESTING 5.1 Concepts When a number of animals or plots are exposed to a certain treatment, we usually estimate the effect of the treatment

More information

Penalized regression: Introduction

Penalized regression: Introduction Penalized regression: Introduction Patrick Breheny August 30 Patrick Breheny BST 764: Applied Statistical Modeling 1/19 Maximum likelihood Much of 20th-century statistics dealt with maximum likelihood

More information

4. Continuous Random Variables, the Pareto and Normal Distributions

4. Continuous Random Variables, the Pareto and Normal Distributions 4. Continuous Random Variables, the Pareto and Normal Distributions A continuous random variable X can take any value in a given range (e.g. height, weight, age). The distribution of a continuous random

More information

MAS2317/3317. Introduction to Bayesian Statistics. More revision material

MAS2317/3317. Introduction to Bayesian Statistics. More revision material MAS2317/3317 Introduction to Bayesian Statistics More revision material Dr. Lee Fawcett, 2014 2015 1 Section A style questions 1. Describe briefly the frequency, classical and Bayesian interpretations

More information

An Introduction to Basic Statistics and Probability

An Introduction to Basic Statistics and Probability An Introduction to Basic Statistics and Probability Shenek Heyward NCSU An Introduction to Basic Statistics and Probability p. 1/4 Outline Basic probability concepts Conditional probability Discrete Random

More information

Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)

Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4) Summary of Formulas and Concepts Descriptive Statistics (Ch. 1-4) Definitions Population: The complete set of numerical information on a particular quantity in which an investigator is interested. We assume

More information

CHAPTER 2 Estimating Probabilities

CHAPTER 2 Estimating Probabilities CHAPTER 2 Estimating Probabilities Machine Learning Copyright c 2016. Tom M. Mitchell. All rights reserved. *DRAFT OF January 24, 2016* *PLEASE DO NOT DISTRIBUTE WITHOUT AUTHOR S PERMISSION* This is a

More information

Mathematical goals. Starting points. Materials required. Time needed

Mathematical goals. Starting points. Materials required. Time needed Level S2 of challenge: B/C S2 Mathematical goals Starting points Materials required Time needed Evaluating probability statements To help learners to: discuss and clarify some common misconceptions about

More information

Inclusion and Exclusion Criteria

Inclusion and Exclusion Criteria Inclusion and Exclusion Criteria Inclusion criteria = attributes of subjects that are essential for their selection to participate. Inclusion criteria function remove the influence of specific confounding

More information

Frequentist vs. Bayesian Statistics

Frequentist vs. Bayesian Statistics Bayes Theorem Frequentist vs. Bayesian Statistics Common situation in science: We have some data and we want to know the true hysical law describing it. We want to come u with a model that fits the data.

More information

Tutorial 5: Hypothesis Testing

Tutorial 5: Hypothesis Testing Tutorial 5: Hypothesis Testing Rob Nicholls nicholls@mrc-lmb.cam.ac.uk MRC LMB Statistics Course 2014 Contents 1 Introduction................................ 1 2 Testing distributional assumptions....................

More information

IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem

IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem Time on my hands: Coin tosses. Problem Formulation: Suppose that I have

More information

Bayesian probability theory

Bayesian probability theory Bayesian probability theory Bruno A. Olshausen arch 1, 2004 Abstract Bayesian probability theory provides a mathematical framework for peforming inference, or reasoning, using probability. The foundations

More information

Paper PO06. Randomization in Clinical Trial Studies

Paper PO06. Randomization in Clinical Trial Studies Paper PO06 Randomization in Clinical Trial Studies David Shen, WCI, Inc. Zaizai Lu, AstraZeneca Pharmaceuticals ABSTRACT Randomization is of central importance in clinical trials. It prevents selection

More information

P (B) In statistics, the Bayes theorem is often used in the following way: P (Data Unknown)P (Unknown) P (Data)

P (B) In statistics, the Bayes theorem is often used in the following way: P (Data Unknown)P (Unknown) P (Data) 22S:101 Biostatistics: J. Huang 1 Bayes Theorem For two events A and B, if we know the conditional probability P (B A) and the probability P (A), then the Bayes theorem tells that we can compute the conditional

More information

p ˆ (sample mean and sample

p ˆ (sample mean and sample Chapter 6: Confidence Intervals and Hypothesis Testing When analyzing data, we can t just accept the sample mean or sample proportion as the official mean or proportion. When we estimate the statistics

More information

Curriculum Map Statistics and Probability Honors (348) Saugus High School Saugus Public Schools 2009-2010

Curriculum Map Statistics and Probability Honors (348) Saugus High School Saugus Public Schools 2009-2010 Curriculum Map Statistics and Probability Honors (348) Saugus High School Saugus Public Schools 2009-2010 Week 1 Week 2 14.0 Students organize and describe distributions of data by using a number of different

More information

This chapter discusses some of the basic concepts in inferential statistics.

This chapter discusses some of the basic concepts in inferential statistics. Research Skills for Psychology Majors: Everything You Need to Know to Get Started Inferential Statistics: Basic Concepts This chapter discusses some of the basic concepts in inferential statistics. Details

More information

CONTENTS OF DAY 2. II. Why Random Sampling is Important 9 A myth, an urban legend, and the real reason NOTES FOR SUMMER STATISTICS INSTITUTE COURSE

CONTENTS OF DAY 2. II. Why Random Sampling is Important 9 A myth, an urban legend, and the real reason NOTES FOR SUMMER STATISTICS INSTITUTE COURSE 1 2 CONTENTS OF DAY 2 I. More Precise Definition of Simple Random Sample 3 Connection with independent random variables 3 Problems with small populations 8 II. Why Random Sampling is Important 9 A myth,

More information

ST 371 (IV): Discrete Random Variables

ST 371 (IV): Discrete Random Variables ST 371 (IV): Discrete Random Variables 1 Random Variables A random variable (rv) is a function that is defined on the sample space of the experiment and that assigns a numerical variable to each possible

More information

Topic #6: Hypothesis. Usage

Topic #6: Hypothesis. Usage Topic #6: Hypothesis A hypothesis is a suggested explanation of a phenomenon or reasoned proposal suggesting a possible correlation between multiple phenomena. The term derives from the ancient Greek,

More information

Bayesian Tutorial (Sheet Updated 20 March)

Bayesian Tutorial (Sheet Updated 20 March) Bayesian Tutorial (Sheet Updated 20 March) Practice Questions (for discussing in Class) Week starting 21 March 2016 1. What is the probability that the total of two dice will be greater than 8, given that

More information

Characteristics of Binomial Distributions

Characteristics of Binomial Distributions Lesson2 Characteristics of Binomial Distributions In the last lesson, you constructed several binomial distributions, observed their shapes, and estimated their means and standard deviations. In Investigation

More information

Correlational Research

Correlational Research Correlational Research Chapter Fifteen Correlational Research Chapter Fifteen Bring folder of readings The Nature of Correlational Research Correlational Research is also known as Associational Research.

More information

Basic Bayesian Methods

Basic Bayesian Methods 6 Basic Bayesian Methods Mark E. Glickman and David A. van Dyk Summary In this chapter, we introduce the basics of Bayesian data analysis. The key ingredients to a Bayesian analysis are the likelihood

More information

Book Review of Rosenhouse, The Monty Hall Problem. Leslie Burkholder 1

Book Review of Rosenhouse, The Monty Hall Problem. Leslie Burkholder 1 Book Review of Rosenhouse, The Monty Hall Problem Leslie Burkholder 1 The Monty Hall Problem, Jason Rosenhouse, New York, Oxford University Press, 2009, xii, 195 pp, US $24.95, ISBN 978-0-19-5#6789-8 (Source

More information

Solutions: Problems for Chapter 3. Solutions: Problems for Chapter 3

Solutions: Problems for Chapter 3. Solutions: Problems for Chapter 3 Problem A: You are dealt five cards from a standard deck. Are you more likely to be dealt two pairs or three of a kind? experiment: choose 5 cards at random from a standard deck Ω = {5-combinations of

More information

Independent samples t-test. Dr. Tom Pierce Radford University

Independent samples t-test. Dr. Tom Pierce Radford University Independent samples t-test Dr. Tom Pierce Radford University The logic behind drawing causal conclusions from experiments The sampling distribution of the difference between means The standard error of

More information

The sample space for a pair of die rolls is the set. The sample space for a random number between 0 and 1 is the interval [0, 1].

The sample space for a pair of die rolls is the set. The sample space for a random number between 0 and 1 is the interval [0, 1]. Probability Theory Probability Spaces and Events Consider a random experiment with several possible outcomes. For example, we might roll a pair of dice, flip a coin three times, or choose a random real

More information

Section 1.3 P 1 = 1 2. = 1 4 2 8. P n = 1 P 3 = Continuing in this fashion, it should seem reasonable that, for any n = 1, 2, 3,..., = 1 2 4.

Section 1.3 P 1 = 1 2. = 1 4 2 8. P n = 1 P 3 = Continuing in this fashion, it should seem reasonable that, for any n = 1, 2, 3,..., = 1 2 4. Difference Equations to Differential Equations Section. The Sum of a Sequence This section considers the problem of adding together the terms of a sequence. Of course, this is a problem only if more than

More information

Introduction to. Hypothesis Testing CHAPTER LEARNING OBJECTIVES. 1 Identify the four steps of hypothesis testing.

Introduction to. Hypothesis Testing CHAPTER LEARNING OBJECTIVES. 1 Identify the four steps of hypothesis testing. Introduction to Hypothesis Testing CHAPTER 8 LEARNING OBJECTIVES After reading this chapter, you should be able to: 1 Identify the four steps of hypothesis testing. 2 Define null hypothesis, alternative

More information

Bayesian Statistical Analysis in Medical Research

Bayesian Statistical Analysis in Medical Research Bayesian Statistical Analysis in Medical Research David Draper Department of Applied Mathematics and Statistics University of California, Santa Cruz draper@ams.ucsc.edu www.ams.ucsc.edu/ draper ROLE Steering

More information

Business Statistics 41000: Probability 1

Business Statistics 41000: Probability 1 Business Statistics 41000: Probability 1 Drew D. Creal University of Chicago, Booth School of Business Week 3: January 24 and 25, 2014 1 Class information Drew D. Creal Email: dcreal@chicagobooth.edu Office:

More information

Chicago Booth BUSINESS STATISTICS 41000 Final Exam Fall 2011

Chicago Booth BUSINESS STATISTICS 41000 Final Exam Fall 2011 Chicago Booth BUSINESS STATISTICS 41000 Final Exam Fall 2011 Name: Section: I pledge my honor that I have not violated the Honor Code Signature: This exam has 34 pages. You have 3 hours to complete this

More information

Chapter 8 Hypothesis Testing Chapter 8 Hypothesis Testing 8-1 Overview 8-2 Basics of Hypothesis Testing

Chapter 8 Hypothesis Testing Chapter 8 Hypothesis Testing 8-1 Overview 8-2 Basics of Hypothesis Testing Chapter 8 Hypothesis Testing 1 Chapter 8 Hypothesis Testing 8-1 Overview 8-2 Basics of Hypothesis Testing 8-3 Testing a Claim About a Proportion 8-5 Testing a Claim About a Mean: s Not Known 8-6 Testing

More information

The Null Hypothesis. Geoffrey R. Loftus University of Washington

The Null Hypothesis. Geoffrey R. Loftus University of Washington The Null Hypothesis Geoffrey R. Loftus University of Washington Send correspondence to: Geoffrey R. Loftus Department of Psychology, Box 351525 University of Washington Seattle, WA 98195-1525 gloftus@u.washington.edu

More information

3.4. The Binomial Probability Distribution. Copyright Cengage Learning. All rights reserved.

3.4. The Binomial Probability Distribution. Copyright Cengage Learning. All rights reserved. 3.4 The Binomial Probability Distribution Copyright Cengage Learning. All rights reserved. The Binomial Probability Distribution There are many experiments that conform either exactly or approximately

More information

Mind on Statistics. Chapter 12

Mind on Statistics. Chapter 12 Mind on Statistics Chapter 12 Sections 12.1 Questions 1 to 6: For each statement, determine if the statement is a typical null hypothesis (H 0 ) or alternative hypothesis (H a ). 1. There is no difference

More information

Chapter 6: Probability

Chapter 6: Probability Chapter 6: Probability In a more mathematically oriented statistics course, you would spend a lot of time talking about colored balls in urns. We will skip over such detailed examinations of probability,

More information

Statistics 104: Section 6!

Statistics 104: Section 6! Page 1 Statistics 104: Section 6! TF: Deirdre (say: Dear-dra) Bloome Email: dbloome@fas.harvard.edu Section Times Thursday 2pm-3pm in SC 109, Thursday 5pm-6pm in SC 705 Office Hours: Thursday 6pm-7pm SC

More information

Conditional Probability, Hypothesis Testing, and the Monty Hall Problem

Conditional Probability, Hypothesis Testing, and the Monty Hall Problem Conditional Probability, Hypothesis Testing, and the Monty Hall Problem Ernie Croot September 17, 2008 On more than one occasion I have heard the comment Probability does not exist in the real world, and

More information

Course Syllabus MATH 110 Introduction to Statistics 3 credits

Course Syllabus MATH 110 Introduction to Statistics 3 credits Course Syllabus MATH 110 Introduction to Statistics 3 credits Prerequisites: Algebra proficiency is required, as demonstrated by successful completion of high school algebra, by completion of a college

More information

5.1 Identifying the Target Parameter

5.1 Identifying the Target Parameter University of California, Davis Department of Statistics Summer Session II Statistics 13 August 20, 2012 Date of latest update: August 20 Lecture 5: Estimation with Confidence intervals 5.1 Identifying

More information

The Importance of Statistics Education

The Importance of Statistics Education The Importance of Statistics Education Professor Jessica Utts Department of Statistics University of California, Irvine http://www.ics.uci.edu/~jutts jutts@uci.edu Outline of Talk What is Statistics? Four

More information

Hypothesis testing. c 2014, Jeffrey S. Simonoff 1

Hypothesis testing. c 2014, Jeffrey S. Simonoff 1 Hypothesis testing So far, we ve talked about inference from the point of estimation. We ve tried to answer questions like What is a good estimate for a typical value? or How much variability is there

More information

Introduction to Hypothesis Testing

Introduction to Hypothesis Testing I. Terms, Concepts. Introduction to Hypothesis Testing A. In general, we do not know the true value of population parameters - they must be estimated. However, we do have hypotheses about what the true

More information

Stat 411/511 THE RANDOMIZATION TEST. Charlotte Wickham. stat511.cwick.co.nz. Oct 16 2015

Stat 411/511 THE RANDOMIZATION TEST. Charlotte Wickham. stat511.cwick.co.nz. Oct 16 2015 Stat 411/511 THE RANDOMIZATION TEST Oct 16 2015 Charlotte Wickham stat511.cwick.co.nz Today Review randomization model Conduct randomization test What about CIs? Using a t-distribution as an approximation

More information

Lesson 9 Hypothesis Testing

Lesson 9 Hypothesis Testing Lesson 9 Hypothesis Testing Outline Logic for Hypothesis Testing Critical Value Alpha (α) -level.05 -level.01 One-Tail versus Two-Tail Tests -critical values for both alpha levels Logic for Hypothesis

More information

Bayesian inference for population prediction of individuals without health insurance in Florida

Bayesian inference for population prediction of individuals without health insurance in Florida Bayesian inference for population prediction of individuals without health insurance in Florida Neung Soo Ha 1 1 NISS 1 / 24 Outline Motivation Description of the Behavioral Risk Factor Surveillance System,

More information

99.37, 99.38, 99.38, 99.39, 99.39, 99.39, 99.39, 99.40, 99.41, 99.42 cm

99.37, 99.38, 99.38, 99.39, 99.39, 99.39, 99.39, 99.40, 99.41, 99.42 cm Error Analysis and the Gaussian Distribution In experimental science theory lives or dies based on the results of experimental evidence and thus the analysis of this evidence is a critical part of the

More information

Chapter 5. Random variables

Chapter 5. Random variables Random variables random variable numerical variable whose value is the outcome of some probabilistic experiment; we use uppercase letters, like X, to denote such a variable and lowercase letters, like

More information

Chapter 7 Section 7.1: Inference for the Mean of a Population

Chapter 7 Section 7.1: Inference for the Mean of a Population Chapter 7 Section 7.1: Inference for the Mean of a Population Now let s look at a similar situation Take an SRS of size n Normal Population : N(, ). Both and are unknown parameters. Unlike what we used

More information

Lecture 8 The Subjective Theory of Betting on Theories

Lecture 8 The Subjective Theory of Betting on Theories Lecture 8 The Subjective Theory of Betting on Theories Patrick Maher Philosophy 517 Spring 2007 Introduction The subjective theory of probability holds that the laws of probability are laws that rational

More information

TUTORIAL on ICH E9 and Other Statistical Regulatory Guidance. Session 1: ICH E9 and E10. PSI Conference, May 2011

TUTORIAL on ICH E9 and Other Statistical Regulatory Guidance. Session 1: ICH E9 and E10. PSI Conference, May 2011 TUTORIAL on ICH E9 and Other Statistical Regulatory Guidance Session 1: PSI Conference, May 2011 Kerry Gordon, Quintiles 1 E9, and how to locate it 2 ICH E9 Statistical Principles for Clinical Trials (Issued

More information

Likelihood Approaches for Trial Designs in Early Phase Oncology

Likelihood Approaches for Trial Designs in Early Phase Oncology Likelihood Approaches for Trial Designs in Early Phase Oncology Clinical Trials Elizabeth Garrett-Mayer, PhD Cody Chiuzan, PhD Hollings Cancer Center Department of Public Health Sciences Medical University

More information

Having a coin come up heads or tails is a variable on a nominal scale. Heads is a different category from tails.

Having a coin come up heads or tails is a variable on a nominal scale. Heads is a different category from tails. Chi-square Goodness of Fit Test The chi-square test is designed to test differences whether one frequency is different from another frequency. The chi-square test is designed for use with data on a nominal

More information

Section 5 Part 2. Probability Distributions for Discrete Random Variables

Section 5 Part 2. Probability Distributions for Discrete Random Variables Section 5 Part 2 Probability Distributions for Discrete Random Variables Review and Overview So far we ve covered the following probability and probability distribution topics Probability rules Probability

More information

Two-sample inference: Continuous data

Two-sample inference: Continuous data Two-sample inference: Continuous data Patrick Breheny April 5 Patrick Breheny STA 580: Biostatistics I 1/32 Introduction Our next two lectures will deal with two-sample inference for continuous data As

More information