Multiple-Comparison Procedures
|
|
|
- Gervase Turner
- 9 years ago
- Views:
Transcription
1 Multiple-Comparison Procedures References A good review of many methods for both parametric and nonparametric multiple comparisons, planned and unplanned, and with some discussion of the philosophical as well as practical issues in their use, is: Day, R. W., and G. P. Quinn (1989) Comparisons of treatments after an analysis of variance in ecology, Ecological Monographs 59: Also, the GLM chapter of the SAS manual discusses the various error rates controlled by different procedures, gives formulas, and concludes with reasonable recommendations. For a contrary perspective on the value of controlling the overall significance level for a set of inferences (and interesting discussion of a couple of other issues in statistics), see: Stewart-Oaten, A. (1995) Rules and judgements in statistics: Three examples, Ecology 76: General usage The purpose of most multiple-comparisons procedures is to control the overall significance level for some set of inferences performed as a follow-up to ANOVA. This overall significance level or error rate is the probability, conditional on all the null hypotheses being tested being true, of rejecting at least one of them, or equivalently, of having at least one confidence interval not include the true value. Overview of methods There are many methods for multiple-comparisons. Most are for pairwise comparisons of group means, to determine which are significantly different from which others. Other methods, though, are for more specialized purposes (e.g. comparing each of several treatments to a control) or allow testing of more-general hypotheses contrasting sets of group means (as is often done in preplanned contrasts ). The various methods differ in how well they properly control the overall significance level and in their relative power; some, such as the popular Duncan s multiple range test do not control the overall significance level. The ones described in this handout all adequately control overall significance and are either easy to use or powerful. They are: Bonferroni extremely general and simple, but often not powerful Tukey s the best for all-possible pairwise comparisons when sample sizes are unequal or confidence intervals are needed; very good even with equal samples sizes without confidence intervals stepdown the most powerful for all possible pairwise comparisons when sample sizes are equal Dunnett s for comparing one sample ( control ) to each of the others, but not comparing the others to each other.
2 MCB compares each mean to the best (largest or smallest, as you specify) of the other means. Scheffé s for unplanned contrasts among sets of means When more than one method is applicable, and it is not clear which is more efficient (i.e. will give the narrower CI, for given α), it is legitimate to try all the applicable methods and use the one which proves most efficient for the given data. Notation In the following α = the overall significance level for the set of inferences, α* = the significance level for a single one of those inferences, c = the number of inferences in the set, I = the number of samples, and N = the total sample size. Note that when all possible pairwise comparisons are made among I means, there are c = I(I-1)/2 comparisons. Bonferroni This approach can be thought of as alpha-splitting. If c inferences (tests or confidence intervals) are each made at some level α*, the maximum possible overall error rate is cα*. Therefore you simply set the level for each separate inference, α*, equal to α/c, where α is your desired overall significance level. Equivalently, and perhaps preferably, when the inferences are tests of hypotheses you can compute a P-value as usual for each test, and multiply it by c. Note that this method is completely general: it applies to any set of c inferences, not only to multiple comparisons following ANOVA. Advantages and disadvantages The main advantage of this approach is that it is very easy, as well as very widely applicable. The main disadvantage is that it often is unnecessarily conservative (weak): α* is smaller than it needs to be. Procedures Computer packages will apply these procedures to all possible pairwise comparisons; if the number of means is I, the total number of possible pairwise comparisons is c = I(I-1)/2. If you only want to compare some of the means, you can do so and define c accordingly. By doing this you do not pay as much of a multiple-comparisons penalty as if you did all possible comparisons. However, you must have decided which subset of comparisons to make before seeing the data!! Otherwise you implicitly did compare all means and must adjust the analysis accordingly. Multiple-Comparison Procedures 2 of 9
3 If your software does not do Bonferroni comparisons, or if you want to coerce the software to correct for only a subset of all pairwise comparisons (as discussed in the preceding paragraph), you may be able to use Fisher s LCD comparisons; these are pairwise comparisons with no correction to control the overall significance level. You therefore would specify the appropriate α* for each comparison, based on the number of comparisons. If doing Bonferroni comparisons by hand you can calculate either two-sample t statistics (or CIs) for each pair of means, or pairwise contrasts between means; the latter approach uses the ANOVA MSE and thus shares the ANOVA assumption of equal variances, while the former approach does not require this assumption. You then use the t critical values for α*/2 in the tests or CIs. (Since α*/2 probably will not appear in a table of t critical values, you may need to interpolate if doing this by hand). Tukey s ( honestly significant difference or HSD ) This approach is specifically for comparing group means in an ANOVA setting. It is based on the distribution of q, the studentized range. The studentized range with k and r degrees of freedom is the range (i.e. maximum minimum) of a set of k independent observations from some normal distribution, divided by an independent estimate (with r degrees-of-freedom) of the standard deviation of that normal distribution. Many texts have tables of this distribution. If there are I samples, all populations means are the same (the complete null hypothesis is true), x L and x S are the largest and smallest sample means, and n L and n S are the respective sample sizes, then ( ) x L x S MSE n L n S will follow the studentized-range distribution with I and N - I degrees of freedom (N is the total sample size). Critical values for q then would be appropriate for comparing x L and x S. Although other pairs of means do not actually represent the range of the observed sample of means (they will differ by less than x L - x S ), q critical values also are used for comparing them; this results in a conservative procedure. A (1-α) CI for (μ i - μ j ) is ( x i x j ) ± q IN, I, 1 α μ i and μ j are significantly different at level α if x i x j q IN, I, 1 α MSE n i n j MSE n i n j Multiple-Comparison Procedures 3 of 9
4 Advantages and disadvantages When doing all pairwise comparisons, this method is considered the best available when confidence intervals are needed or sample sizes are not equal. When samples sizes are equal and confidence intervals are not needed Tukey s test is slightly less powerful than the stepdown procedures, but if they are not available Tukey s is the next-best choice, and unless the number of groups is large, the loss in power will be slight. Procedures Any worthwhile statistical software will provide Tukey s comparisons. Stepdown procedures As noted above, Tukey s procedure treats every pair of sample means as if they are the mostdifferent pair: every pairwise difference is compared to the distribution expected for the range of I means. The stepdown procedures are modifications of Tukey s procedure which take into account that all but one of the comparisons are less different than the range. In essence, they work like this: 1. Compare the largest and smallest sample means, x L and, to the q distribution for comparing I means. If this comparison is not significant, stop; otherwise, continue. 2. Compare the next most different pair of means (either the largest and the second-smallest or the second-largest and the smallest) to the q distribution for comparing I - 1 means. This is appropriate since one of the two most-extreme means is excluded from the set of means of which this second comparison represents the range. If this comparison is not significant, stop; otherwise, continue. 3. Continue comparing successively less-different pairs of means to q distributions with the number of means parameter successively smaller to represent the smaller set of means being considered in choosing each pair. 4. Etc. until a comparison is not significant or all comparisons have been made. Advantages and disadvantages When doing all pairwise comparisons, this approach is the best available when confidence intervals are not needed and sample sizes are equal. Procedures There actually are numerous implementations of the general step-down procedure described above. Which if any are available will depend on the software being used. SAS provides the Ryan-Einot-Gabriel-Welsch multiple-range test in the one-way ANOVA procedure within the Analyst application and presumably also within the Enterprise Guide interface (also available as the REGWQ and REGWF options to PROC GLM). x S Multiple-Comparison Procedures 4 of 9
5 Dunnett s procedure for comparing treatments with a control When you do not wish to make all pairwise comparisons, but rather only to compare one of the groups (the controls, usually) with each other group, this procedure reduces the multiple-comparisons price you otherwise would pay, while taking into account that all these comparisons are correlated since they all use the same control data. You can also specify one-sided tests, which might be appropriate here but are not in general multiple comparisons. Unfortunately, the table of the critical values needed for this procedure is not widely available, so you probably will have to rely on a computer package. Advantages and disadvantages When it is appropriate when you really are interested only in comparisons of one group to each of the others this approach is more powerful than methods performing all possible pairwise comparisons, and therefore is recommended. Its disadvantage is simply that it does not compare the other groups to each other at all. Hsu s multiple comparisons with best (MCB) Hsu s MCB does comparisons between each sample mean and the best of all the other means, where you specify that best means either largest or smallest. In essence it is a modification of Dunnett s method, allowing it to be applied when you do not know in advance which group you want to compare all the others to. Its purpose is as the name suggests: to select which group(s) is/are the best: not significantly different from each other but significantly better than the others. Advantages and disadvantages Since this approach does fewer comparisons than do methods (Tukey s, stepdown) designed for doing all pairwise comparisons, it does not have to make as great a reduction in individual significance levels and therefore will be more powerful. Conversely, it makes more comparisons than Dunnett s and therefore will be less powerful. It thus is the best method available when it truly is appropriate to your purpose. Scheffé s method for all possible contrasts If you want to consider post hoc comparisons other than just pairwise comparisons (i.e. unplanned contrasts), Scheffé s method can be used to control the overall confidence level. The best way to use this procedure is to calculate confidence intervals for the contrasts of interest, just as for preplanned contrasts but using (in place of the t* critical value) the quantity S, where S 2 = ( I 1) F 1 α; I 1 ; N I i.e. use (I-1) times the square root of the appropriate F critical value. Multiple-Comparison Procedures 5 of 9
6 Advantages and disadvantages This procedure is extremely conservative since it controls the overall significance for any possible contrast or set of contrasts, even when suggested by the data. It therefore should not be used if only pairwise comparisons are wanted. Furthermore, if the number of contrasts being considered is not considerably greater than the number of groups (i.e. I) and the contrasts were not suggested by the data, a Bonferroni correction probably will be more powerful than Scheffé s. If the contrasts were suggested by the data, however, Scheffé s should be used rather than Bonferroni, since all possible contrasts were implicitly considered (or at least you were willing to consider them). Multiple-Comparison Procedures 6 of 9
7 Computing procedures Minitab Minitab provides only Tukey s, Dunnett s and Hsu s MCB comparisons; it does not provide any step-down methods. Multiple comparisons procedures are options of the oneway ANOVA procedures (Stat ANOVA Oneway and Stat ANOVA Oneway (Unstacked) ), as well as for some of the multifactor ANOVA procedures. In the One-way Analysis of Variance window, after selecting the response variable and the factor, click on the Comparisons button. Then select the procedure(s) desired; choices are Tukey s, Fisher s, Dunnett s, and Hsu s. Tukey s, Dunnett s, and Hsu s procedures are as described above, and each allows the family (i.e. overall) error rate to be specified. For Dunnett s, which level of the factor is to be considered the control must be specified (if values are text rather than numerical, the control value must be put in quotes). The Fisher s option simply does pairwise comparisons with no attempt to control the overall significance level. If done only when the main ANOVA is significant, this is called Fisher s protected Least Significant Difference (LSD) procedure ( protected by the significance of the main ANOVA). As noted in the section on the Bonferroni method, this procedure also can be used to carry out Bonferroni comparisons by specifying the appropriate individual error rate (i.e. α* = α/c where α is the desired overall significance and c is the total number of comparisons desired); this would only be useful if for some reason only a subset (chosen in advance) of the pairwise comparisons was of interest so that Tukey s procedure would be too conservative. Results: Results are presented as confidence intervals for pairwise differences. After a heading saying what kind of comparisons they are and at that confidence level, there is a line stating the testwise confidence level corresponding to the chosen family-wise confidence level. Then come the actual results, in the form of a series of tables, each comparing one group to each of the subsequent groups. The first table compares the first group to each of the others, the second table compares the second group to all remaining groups (i.e. all but the first), etc. (Group ordering is alphabetical by the group labels.) Each table is labelled factor = value subtracted from: where the name of the grouping variable (the ANOVA factor) takes the place of factor and the label of the reference group for that table takes the place of value. The rows in the table then are labelled (in the column headed factor ) by the label of the group being compared to that table s reference group. The pairwise comparisons in each row are given as the estimated difference in population means (estimated simply by the difference in sample means), listed in the column headed Center, surrounded by the CI for the difference in population means, listed in the columns headed Lower and Upper, followed by a crude chart of these estimates. Multiple-Comparison Procedures 7 of 9
8 Tukey 95% Simultaneous Confidence Intervals All Pairwise Comparisons among Levels of source Individual confidence level = 98.06% source = Kauai subtracted from: source Lower Center Upper kure ( * ) midway ( * ) source = kure subtracted from: source Lower Center Upper midway ( * ) S-Plus S-Plus provides several methods, including Tukey s, Dunnett s, and a special simulationbased method. It does not provide Hsu s MCB or any step-down methods. Multiple comparisons can be done using the general ANOVA procedure invoked by Statistics ANOVA Fixed Effects (The simpler method invoked by Statistics Compare Samples k Samples Oneway ANOVA does not do comparisons.) Comparisons are requested on the Compare tab of the ANOVA dialog. The only item in this dialog that must be specified is the Variable - Levels Of:, which must be the factor in the ANOVA model. The default options generally are appropriate. Options: The Comparison Type part of the dialog gives a choice among mca (mean comparisons among all groups; this is the default), mcc (mean comparisons with control), and none, which gives CIs for the level means rather than pairwise comparisons. The Method part of the dialog gives a choice among best, best.fast, Bonferroni, Dunnett, Fisher.lsd, Scheffe, Sidak, Simulation, and Tukey. Most of these are described in the main part of this handout. The others are: Multiple-Comparison Procedures 8 of 9
9 best: whichever of the suite of methods is most powerful, best.fast: whichever of the suite of methods other than Simulation is most powerful, Fisher.lsd: as described in passing elsewhere in this handout, these are pairwise contrasts with no correction for multiple testing, Sidak: similar to but slightly more powerful than Bonferroni, and Simulation (see below). The default method is best.fast. When the Comparison Type is mca the best.fast option generally will use the Tukey method for two-ended CIs and the Sidak method if only upper or lower bounds are requested. When the Comparison Type is mcc Dunnett s method generally will be used. The Simulation method uses Monte Carlo simulation of the sampling distribution of the set of sample means assuming the null hypothesis of equal population means. This distribution will depend on will details of the design (including sample sizes) in ways that the standard methods, using tabled distributions, cannot fully address. The simulation method therefore will generally be more powerful than other valid methods, and sometimes will be much more powerful. Its disadvantage is that it can take a lot of computer time with large data sets. Output: The output (see below) is in the form of the difference, its SE, and the resulting CI, for each of the pairs of differences. Significant comparisons are flagged. If the Plot Intervals option is checked, a graph showing all the pairwise CIs is produced, as shown below after the text output. 95 % simultaneous confidence intervals for specified linear combinations, by the Tukey method critical point: response variable: beakwid intervals excluding 0 are flagged by '****' Estimate Std.Error Lower Bound Upper Bound Laysan-North **** Laysan-SouthEast **** North-SouthEast Laysan-North Laysan-SouthEast North-SouthEast ( ( ) ) ( ) simultaneous 95 % confidence limits, Tukey method response variable: beakwid Multiple-Comparison Procedures 9 of 9
INTERPRETING THE ONE-WAY ANALYSIS OF VARIANCE (ANOVA)
INTERPRETING THE ONE-WAY ANALYSIS OF VARIANCE (ANOVA) As with other parametric statistics, we begin the one-way ANOVA with a test of the underlying assumptions. Our first assumption is the assumption of
LAB 4 INSTRUCTIONS CONFIDENCE INTERVALS AND HYPOTHESIS TESTING
LAB 4 INSTRUCTIONS CONFIDENCE INTERVALS AND HYPOTHESIS TESTING In this lab you will explore the concept of a confidence interval and hypothesis testing through a simulation problem in engineering setting.
Section 13, Part 1 ANOVA. Analysis Of Variance
Section 13, Part 1 ANOVA Analysis Of Variance Course Overview So far in this course we ve covered: Descriptive statistics Summary statistics Tables and Graphs Probability Probability Rules Probability
One-Way ANOVA using SPSS 11.0. SPSS ANOVA procedures found in the Compare Means analyses. Specifically, we demonstrate
1 One-Way ANOVA using SPSS 11.0 This section covers steps for testing the difference between three or more group means using the SPSS ANOVA procedures found in the Compare Means analyses. Specifically,
Chapter 5 Analysis of variance SPSS Analysis of variance
Chapter 5 Analysis of variance SPSS Analysis of variance Data file used: gss.sav How to get there: Analyze Compare Means One-way ANOVA To test the null hypothesis that several population means are equal,
13: Additional ANOVA Topics. Post hoc Comparisons
13: Additional ANOVA Topics Post hoc Comparisons ANOVA Assumptions Assessing Group Variances When Distributional Assumptions are Severely Violated Kruskal-Wallis Test Post hoc Comparisons In the prior
Simple Tricks for Using SPSS for Windows
Simple Tricks for Using SPSS for Windows Chapter 14. Follow-up Tests for the Two-Way Factorial ANOVA The Interaction is Not Significant If you have performed a two-way ANOVA using the General Linear Model,
How To Check For Differences In The One Way Anova
MINITAB ASSISTANT WHITE PAPER This paper explains the research conducted by Minitab statisticians to develop the methods and data checks used in the Assistant in Minitab 17 Statistical Software. One-Way
SPSS Tests for Versions 9 to 13
SPSS Tests for Versions 9 to 13 Chapter 2 Descriptive Statistic (including median) Choose Analyze Descriptive statistics Frequencies... Click on variable(s) then press to move to into Variable(s): list
NCSS Statistical Software
Chapter 06 Introduction This procedure provides several reports for the comparison of two distributions, including confidence intervals for the difference in means, two-sample t-tests, the z-test, the
Two-Sample T-Tests Assuming Equal Variance (Enter Means)
Chapter 4 Two-Sample T-Tests Assuming Equal Variance (Enter Means) Introduction This procedure provides sample size and power calculations for one- or two-sided two-sample t-tests when the variances of
Analysis of Variance. MINITAB User s Guide 2 3-1
3 Analysis of Variance Analysis of Variance Overview, 3-2 One-Way Analysis of Variance, 3-5 Two-Way Analysis of Variance, 3-11 Analysis of Means, 3-13 Overview of Balanced ANOVA and GLM, 3-18 Balanced
8. Comparing Means Using One Way ANOVA
8. Comparing Means Using One Way ANOVA Objectives Calculate a one-way analysis of variance Run various multiple comparisons Calculate measures of effect size A One Way ANOVA is an analysis of variance
UNDERSTANDING THE TWO-WAY ANOVA
UNDERSTANDING THE e have seen how the one-way ANOVA can be used to compare two or more sample means in studies involving a single independent variable. This can be extended to two independent variables
Chapter 7. One-way ANOVA
Chapter 7 One-way ANOVA One-way ANOVA examines equality of population means for a quantitative outcome and a single categorical explanatory variable with any number of levels. The t-test of Chapter 6 looks
SPSS Explore procedure
SPSS Explore procedure One useful function in SPSS is the Explore procedure, which will produce histograms, boxplots, stem-and-leaf plots and extensive descriptive statistics. To run the Explore procedure,
Two-Sample T-Tests Allowing Unequal Variance (Enter Difference)
Chapter 45 Two-Sample T-Tests Allowing Unequal Variance (Enter Difference) Introduction This procedure provides sample size and power calculations for one- or two-sided two-sample t-tests when no assumption
ANOVA ANOVA. Two-Way ANOVA. One-Way ANOVA. When to use ANOVA ANOVA. Analysis of Variance. Chapter 16. A procedure for comparing more than two groups
ANOVA ANOVA Analysis of Variance Chapter 6 A procedure for comparing more than two groups independent variable: smoking status non-smoking one pack a day > two packs a day dependent variable: number of
Data Analysis Tools. Tools for Summarizing Data
Data Analysis Tools This section of the notes is meant to introduce you to many of the tools that are provided by Excel under the Tools/Data Analysis menu item. If your computer does not have that tool
Analysis of Variance ANOVA
Analysis of Variance ANOVA Overview We ve used the t -test to compare the means from two independent groups. Now we ve come to the final topic of the course: how to compare means from more than two populations.
Recall this chart that showed how most of our course would be organized:
Chapter 4 One-Way ANOVA Recall this chart that showed how most of our course would be organized: Explanatory Variable(s) Response Variable Methods Categorical Categorical Contingency Tables Categorical
Projects Involving Statistics (& SPSS)
Projects Involving Statistics (& SPSS) Academic Skills Advice Starting a project which involves using statistics can feel confusing as there seems to be many different things you can do (charts, graphs,
Multivariate Analysis of Variance. The general purpose of multivariate analysis of variance (MANOVA) is to determine
2 - Manova 4.3.05 25 Multivariate Analysis of Variance What Multivariate Analysis of Variance is The general purpose of multivariate analysis of variance (MANOVA) is to determine whether multiple levels
How To Test For Significance On A Data Set
Non-Parametric Univariate Tests: 1 Sample Sign Test 1 1 SAMPLE SIGN TEST A non-parametric equivalent of the 1 SAMPLE T-TEST. ASSUMPTIONS: Data is non-normally distributed, even after log transforming.
ABSORBENCY OF PAPER TOWELS
ABSORBENCY OF PAPER TOWELS 15. Brief Version of the Case Study 15.1 Problem Formulation 15.2 Selection of Factors 15.3 Obtaining Random Samples of Paper Towels 15.4 How will the Absorbency be measured?
SCHOOL OF HEALTH AND HUMAN SCIENCES DON T FORGET TO RECODE YOUR MISSING VALUES
SCHOOL OF HEALTH AND HUMAN SCIENCES Using SPSS Topics addressed today: 1. Differences between groups 2. Graphing Use the s4data.sav file for the first part of this session. DON T FORGET TO RECODE YOUR
THE FIRST SET OF EXAMPLES USE SUMMARY DATA... EXAMPLE 7.2, PAGE 227 DESCRIBES A PROBLEM AND A HYPOTHESIS TEST IS PERFORMED IN EXAMPLE 7.
THERE ARE TWO WAYS TO DO HYPOTHESIS TESTING WITH STATCRUNCH: WITH SUMMARY DATA (AS IN EXAMPLE 7.17, PAGE 236, IN ROSNER); WITH THE ORIGINAL DATA (AS IN EXAMPLE 8.5, PAGE 301 IN ROSNER THAT USES DATA FROM
NCSS Statistical Software
Chapter 06 Introduction This procedure provides several reports for the comparison of two distributions, including confidence intervals for the difference in means, two-sample t-tests, the z-test, the
Descriptive Statistics
Descriptive Statistics Primer Descriptive statistics Central tendency Variation Relative position Relationships Calculating descriptive statistics Descriptive Statistics Purpose to describe or summarize
Linear Models in STATA and ANOVA
Session 4 Linear Models in STATA and ANOVA Page Strengths of Linear Relationships 4-2 A Note on Non-Linear Relationships 4-4 Multiple Linear Regression 4-5 Removal of Variables 4-8 Independent Samples
CHAPTER 11 CHI-SQUARE AND F DISTRIBUTIONS
CHAPTER 11 CHI-SQUARE AND F DISTRIBUTIONS CHI-SQUARE TESTS OF INDEPENDENCE (SECTION 11.1 OF UNDERSTANDABLE STATISTICS) In chi-square tests of independence we use the hypotheses. H0: The variables are independent
One-Way Analysis of Variance (ANOVA) Example Problem
One-Way Analysis of Variance (ANOVA) Example Problem Introduction Analysis of Variance (ANOVA) is a hypothesis-testing technique used to test the equality of two or more population (or treatment) means
Hypothesis testing - Steps
Hypothesis testing - Steps Steps to do a two-tailed test of the hypothesis that β 1 0: 1. Set up the hypotheses: H 0 : β 1 = 0 H a : β 1 0. 2. Compute the test statistic: t = b 1 0 Std. error of b 1 =
Odds ratio, Odds ratio test for independence, chi-squared statistic.
Odds ratio, Odds ratio test for independence, chi-squared statistic. Announcements: Assignment 5 is live on webpage. Due Wed Aug 1 at 4:30pm. (9 days, 1 hour, 58.5 minutes ) Final exam is Aug 9. Review
STATISTICA Formula Guide: Logistic Regression. Table of Contents
: Table of Contents... 1 Overview of Model... 1 Dispersion... 2 Parameterization... 3 Sigma-Restricted Model... 3 Overparameterized Model... 4 Reference Coding... 4 Model Summary (Summary Tab)... 5 Summary
Chapter 23. Inferences for Regression
Chapter 23. Inferences for Regression Topics covered in this chapter: Simple Linear Regression Simple Linear Regression Example 23.1: Crying and IQ The Problem: Infants who cry easily may be more easily
CS 147: Computer Systems Performance Analysis
CS 147: Computer Systems Performance Analysis One-Factor Experiments CS 147: Computer Systems Performance Analysis One-Factor Experiments 1 / 42 Overview Introduction Overview Overview Introduction Finding
Analysis of Data. Organizing Data Files in SPSS. Descriptive Statistics
Analysis of Data Claudia J. Stanny PSY 67 Research Design Organizing Data Files in SPSS All data for one subject entered on the same line Identification data Between-subjects manipulations: variable to
Two-Way ANOVA tests. I. Definition and Applications...2. II. Two-Way ANOVA prerequisites...2. III. How to use the Two-Way ANOVA tool?...
Two-Way ANOVA tests Contents at a glance I. Definition and Applications...2 II. Two-Way ANOVA prerequisites...2 III. How to use the Two-Way ANOVA tool?...3 A. Parametric test, assume variances equal....4
Lesson 1: Comparison of Population Means Part c: Comparison of Two- Means
Lesson : Comparison of Population Means Part c: Comparison of Two- Means Welcome to lesson c. This third lesson of lesson will discuss hypothesis testing for two independent means. Steps in Hypothesis
Good luck! BUSINESS STATISTICS FINAL EXAM INSTRUCTIONS. Name:
Glo bal Leadership M BA BUSINESS STATISTICS FINAL EXAM Name: INSTRUCTIONS 1. Do not open this exam until instructed to do so. 2. Be sure to fill in your name before starting the exam. 3. You have two hours
A Basic Guide to Analyzing Individual Scores Data with SPSS
A Basic Guide to Analyzing Individual Scores Data with SPSS Step 1. Clean the data file Open the Excel file with your data. You may get the following message: If you get this message, click yes. Delete
1. What is the critical value for this 95% confidence interval? CV = z.025 = invnorm(0.025) = 1.96
1 Final Review 2 Review 2.1 CI 1-propZint Scenario 1 A TV manufacturer claims in its warranty brochure that in the past not more than 10 percent of its TV sets needed any repair during the first two years
QUANTITATIVE METHODS BIOLOGY FINAL HONOUR SCHOOL NON-PARAMETRIC TESTS
QUANTITATIVE METHODS BIOLOGY FINAL HONOUR SCHOOL NON-PARAMETRIC TESTS This booklet contains lecture notes for the nonparametric work in the QM course. This booklet may be online at http://users.ox.ac.uk/~grafen/qmnotes/index.html.
CHAPTER 13. Experimental Design and Analysis of Variance
CHAPTER 13 Experimental Design and Analysis of Variance CONTENTS STATISTICS IN PRACTICE: BURKE MARKETING SERVICES, INC. 13.1 AN INTRODUCTION TO EXPERIMENTAL DESIGN AND ANALYSIS OF VARIANCE Data Collection
Part 3. Comparing Groups. Chapter 7 Comparing Paired Groups 189. Chapter 8 Comparing Two Independent Groups 217
Part 3 Comparing Groups Chapter 7 Comparing Paired Groups 189 Chapter 8 Comparing Two Independent Groups 217 Chapter 9 Comparing More Than Two Groups 257 188 Elementary Statistics Using SAS Chapter 7 Comparing
Simple Regression Theory II 2010 Samuel L. Baker
SIMPLE REGRESSION THEORY II 1 Simple Regression Theory II 2010 Samuel L. Baker Assessing how good the regression equation is likely to be Assignment 1A gets into drawing inferences about how close the
t Tests in Excel The Excel Statistical Master By Mark Harmon Copyright 2011 Mark Harmon
t-tests in Excel By Mark Harmon Copyright 2011 Mark Harmon No part of this publication may be reproduced or distributed without the express permission of the author. [email protected] www.excelmasterseries.com
Two Related Samples t Test
Two Related Samples t Test In this example 1 students saw five pictures of attractive people and five pictures of unattractive people. For each picture, the students rated the friendliness of the person
Association Between Variables
Contents 11 Association Between Variables 767 11.1 Introduction............................ 767 11.1.1 Measure of Association................. 768 11.1.2 Chapter Summary.................... 769 11.2 Chi
IBM SPSS Statistics 20 Part 4: Chi-Square and ANOVA
CALIFORNIA STATE UNIVERSITY, LOS ANGELES INFORMATION TECHNOLOGY SERVICES IBM SPSS Statistics 20 Part 4: Chi-Square and ANOVA Summer 2013, Version 2.0 Table of Contents Introduction...2 Downloading the
Variables Control Charts
MINITAB ASSISTANT WHITE PAPER This paper explains the research conducted by Minitab statisticians to develop the methods and data checks used in the Assistant in Minitab 17 Statistical Software. Variables
2 Sample t-test (unequal sample sizes and unequal variances)
Variations of the t-test: Sample tail Sample t-test (unequal sample sizes and unequal variances) Like the last example, below we have ceramic sherd thickness measurements (in cm) of two samples representing
EPS 625 INTERMEDIATE STATISTICS FRIEDMAN TEST
EPS 625 INTERMEDIATE STATISTICS The Friedman test is an extension of the Wilcoxon test. The Wilcoxon test can be applied to repeated-measures data if participants are assessed on two occasions or conditions
Estimation of σ 2, the variance of ɛ
Estimation of σ 2, the variance of ɛ The variance of the errors σ 2 indicates how much observations deviate from the fitted surface. If σ 2 is small, parameters β 0, β 1,..., β k will be reliably estimated
HYPOTHESIS TESTING: CONFIDENCE INTERVALS, T-TESTS, ANOVAS, AND REGRESSION
HYPOTHESIS TESTING: CONFIDENCE INTERVALS, T-TESTS, ANOVAS, AND REGRESSION HOD 2990 10 November 2010 Lecture Background This is a lightning speed summary of introductory statistical methods for senior undergraduate
Comparing Means in Two Populations
Comparing Means in Two Populations Overview The previous section discussed hypothesis testing when sampling from a single population (either a single mean or two means from the same population). Now we
Class 19: Two Way Tables, Conditional Distributions, Chi-Square (Text: Sections 2.5; 9.1)
Spring 204 Class 9: Two Way Tables, Conditional Distributions, Chi-Square (Text: Sections 2.5; 9.) Big Picture: More than Two Samples In Chapter 7: We looked at quantitative variables and compared the
Study Guide for the Final Exam
Study Guide for the Final Exam When studying, remember that the computational portion of the exam will only involve new material (covered after the second midterm), that material from Exam 1 will make
Biostatistics: DESCRIPTIVE STATISTICS: 2, VARIABILITY
Biostatistics: DESCRIPTIVE STATISTICS: 2, VARIABILITY 1. Introduction Besides arriving at an appropriate expression of an average or consensus value for observations of a population, it is important to
Confidence Intervals for Cp
Chapter 296 Confidence Intervals for Cp Introduction This routine calculates the sample size needed to obtain a specified width of a Cp confidence interval at a stated confidence level. Cp is a process
Sample Size and Power in Clinical Trials
Sample Size and Power in Clinical Trials Version 1.0 May 011 1. Power of a Test. Factors affecting Power 3. Required Sample Size RELATED ISSUES 1. Effect Size. Test Statistics 3. Variation 4. Significance
Minitab Tutorials for Design and Analysis of Experiments. Table of Contents
Table of Contents Introduction to Minitab...2 Example 1 One-Way ANOVA...3 Determining Sample Size in One-way ANOVA...8 Example 2 Two-factor Factorial Design...9 Example 3: Randomized Complete Block Design...14
CHAPTER 14 NONPARAMETRIC TESTS
CHAPTER 14 NONPARAMETRIC TESTS Everything that we have done up until now in statistics has relied heavily on one major fact: that our data is normally distributed. We have been able to make inferences
Regression step-by-step using Microsoft Excel
Step 1: Regression step-by-step using Microsoft Excel Notes prepared by Pamela Peterson Drake, James Madison University Type the data into the spreadsheet The example used throughout this How to is a regression
Examining Differences (Comparing Groups) using SPSS Inferential statistics (Part I) Dwayne Devonish
Examining Differences (Comparing Groups) using SPSS Inferential statistics (Part I) Dwayne Devonish Statistics Statistics are quantitative methods of describing, analysing, and drawing inferences (conclusions)
Hypothesis testing. c 2014, Jeffrey S. Simonoff 1
Hypothesis testing So far, we ve talked about inference from the point of estimation. We ve tried to answer questions like What is a good estimate for a typical value? or How much variability is there
3.4 Statistical inference for 2 populations based on two samples
3.4 Statistical inference for 2 populations based on two samples Tests for a difference between two population means The first sample will be denoted as X 1, X 2,..., X m. The second sample will be denoted
Chapter 7. Comparing Means in SPSS (t-tests) Compare Means analyses. Specifically, we demonstrate procedures for running Dependent-Sample (or
1 Chapter 7 Comparing Means in SPSS (t-tests) This section covers procedures for testing the differences between two means using the SPSS Compare Means analyses. Specifically, we demonstrate procedures
Chapter 2 Probability Topics SPSS T tests
Chapter 2 Probability Topics SPSS T tests Data file used: gss.sav In the lecture about chapter 2, only the One-Sample T test has been explained. In this handout, we also give the SPSS methods to perform
KSTAT MINI-MANUAL. Decision Sciences 434 Kellogg Graduate School of Management
KSTAT MINI-MANUAL Decision Sciences 434 Kellogg Graduate School of Management Kstat is a set of macros added to Excel and it will enable you to do the statistics required for this course very easily. To
Lin s Concordance Correlation Coefficient
NSS Statistical Software NSS.com hapter 30 Lin s oncordance orrelation oefficient Introduction This procedure calculates Lin s concordance correlation coefficient ( ) from a set of bivariate data. The
Testing Group Differences using T-tests, ANOVA, and Nonparametric Measures
Testing Group Differences using T-tests, ANOVA, and Nonparametric Measures Jamie DeCoster Department of Psychology University of Alabama 348 Gordon Palmer Hall Box 870348 Tuscaloosa, AL 35487-0348 Phone:
NCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( )
Chapter 340 Principal Components Regression Introduction is a technique for analyzing multiple regression data that suffer from multicollinearity. When multicollinearity occurs, least squares estimates
Confidence Intervals for One Standard Deviation Using Standard Deviation
Chapter 640 Confidence Intervals for One Standard Deviation Using Standard Deviation Introduction This routine calculates the sample size necessary to achieve a specified interval width or distance from
General Method: Difference of Means. 3. Calculate df: either Welch-Satterthwaite formula or simpler df = min(n 1, n 2 ) 1.
General Method: Difference of Means 1. Calculate x 1, x 2, SE 1, SE 2. 2. Combined SE = SE1 2 + SE2 2. ASSUMES INDEPENDENT SAMPLES. 3. Calculate df: either Welch-Satterthwaite formula or simpler df = min(n
SPSS and AMOS. Miss Brenda Lee 2:00p.m. 6:00p.m. 24 th July, 2015 The Open University of Hong Kong
Seminar on Quantitative Data Analysis: SPSS and AMOS Miss Brenda Lee 2:00p.m. 6:00p.m. 24 th July, 2015 The Open University of Hong Kong SBAS (Hong Kong) Ltd. All Rights Reserved. 1 Agenda MANOVA, Repeated
Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)
Summary of Formulas and Concepts Descriptive Statistics (Ch. 1-4) Definitions Population: The complete set of numerical information on a particular quantity in which an investigator is interested. We assume
Stepwise Regression. Chapter 311. Introduction. Variable Selection Procedures. Forward (Step-Up) Selection
Chapter 311 Introduction Often, theory and experience give only general direction as to which of a pool of candidate variables (including transformed variables) should be included in the regression model.
Skewed Data and Non-parametric Methods
0 2 4 6 8 10 12 14 Skewed Data and Non-parametric Methods Comparing two groups: t-test assumes data are: 1. Normally distributed, and 2. both samples have the same SD (i.e. one sample is simply shifted
1.5 Oneway Analysis of Variance
Statistics: Rosie Cornish. 200. 1.5 Oneway Analysis of Variance 1 Introduction Oneway analysis of variance (ANOVA) is used to compare several means. This method is often used in scientific or medical experiments
Once saved, if the file was zipped you will need to unzip it. For the files that I will be posting you need to change the preferences.
1 Commands in JMP and Statcrunch Below are a set of commands in JMP and Statcrunch which facilitate a basic statistical analysis. The first part concerns commands in JMP, the second part is for analysis
The Dummy s Guide to Data Analysis Using SPSS
The Dummy s Guide to Data Analysis Using SPSS Mathematics 57 Scripps College Amy Gamble April, 2001 Amy Gamble 4/30/01 All Rights Rerserved TABLE OF CONTENTS PAGE Helpful Hints for All Tests...1 Tests
Confidence Intervals for Cpk
Chapter 297 Confidence Intervals for Cpk Introduction This routine calculates the sample size needed to obtain a specified width of a Cpk confidence interval at a stated confidence level. Cpk is a process
1 Basic ANOVA concepts
Math 143 ANOVA 1 Analysis of Variance (ANOVA) Recall, when we wanted to compare two population means, we used the 2-sample t procedures. Now let s expand this to compare k 3 population means. As with the
Statistiek II. John Nerbonne. October 1, 2010. Dept of Information Science [email protected]
Dept of Information Science [email protected] October 1, 2010 Course outline 1 One-way ANOVA. 2 Factorial ANOVA. 3 Repeated measures ANOVA. 4 Correlation and regression. 5 Multiple regression. 6 Logistic
An introduction to IBM SPSS Statistics
An introduction to IBM SPSS Statistics Contents 1 Introduction... 1 2 Entering your data... 2 3 Preparing your data for analysis... 10 4 Exploring your data: univariate analysis... 14 5 Generating descriptive
Bill Burton Albert Einstein College of Medicine [email protected] April 28, 2014 EERS: Managing the Tension Between Rigor and Resources 1
Bill Burton Albert Einstein College of Medicine [email protected] April 28, 2014 EERS: Managing the Tension Between Rigor and Resources 1 Calculate counts, means, and standard deviations Produce
DDBA 8438: Introduction to Hypothesis Testing Video Podcast Transcript
DDBA 8438: Introduction to Hypothesis Testing Video Podcast Transcript JENNIFER ANN MORROW: Welcome to "Introduction to Hypothesis Testing." My name is Dr. Jennifer Ann Morrow. In today's demonstration,
Point Biserial Correlation Tests
Chapter 807 Point Biserial Correlation Tests Introduction The point biserial correlation coefficient (ρ in this chapter) is the product-moment correlation calculated between a continuous random variable
Unit 31 A Hypothesis Test about Correlation and Slope in a Simple Linear Regression
Unit 31 A Hypothesis Test about Correlation and Slope in a Simple Linear Regression Objectives: To perform a hypothesis test concerning the slope of a least squares line To recognize that testing for a
Concepts of Experimental Design
Design Institute for Six Sigma A SAS White Paper Table of Contents Introduction...1 Basic Concepts... 1 Designing an Experiment... 2 Write Down Research Problem and Questions... 2 Define Population...
Directions for using SPSS
Directions for using SPSS Table of Contents Connecting and Working with Files 1. Accessing SPSS... 2 2. Transferring Files to N:\drive or your computer... 3 3. Importing Data from Another File Format...
Using Excel s Analysis ToolPak Add-In
Using Excel s Analysis ToolPak Add-In S. Christian Albright, September 2013 Introduction This document illustrates the use of Excel s Analysis ToolPak add-in for data analysis. The document is aimed at
Introduction to Analysis of Variance (ANOVA) Limitations of the t-test
Introduction to Analysis of Variance (ANOVA) The Structural Model, The Summary Table, and the One- Way ANOVA Limitations of the t-test Although the t-test is commonly used, it has limitations Can only
Consider a study in which. How many subjects? The importance of sample size calculations. An insignificant effect: two possibilities.
Consider a study in which How many subjects? The importance of sample size calculations Office of Research Protections Brown Bag Series KB Boomer, Ph.D. Director, [email protected] A researcher conducts
MBA 611 STATISTICS AND QUANTITATIVE METHODS
MBA 611 STATISTICS AND QUANTITATIVE METHODS Part I. Review of Basic Statistics (Chapters 1-11) A. Introduction (Chapter 1) Uncertainty: Decisions are often based on incomplete information from uncertain
Confidence Intervals for Spearman s Rank Correlation
Chapter 808 Confidence Intervals for Spearman s Rank Correlation Introduction This routine calculates the sample size needed to obtain a specified width of Spearman s rank correlation coefficient confidence
