Come scegliere un test statistico



Similar documents
Analyzing Data with GraphPad Prism

Projects Involving Statistics (& SPSS)

QUANTITATIVE METHODS BIOLOGY FINAL HONOUR SCHOOL NON-PARAMETRIC TESTS

The InStat guide to choosing and interpreting statistical tests

SPSS Tests for Versions 9 to 13

SPSS Explore procedure

SCHOOL OF HEALTH AND HUMAN SCIENCES DON T FORGET TO RECODE YOUR MISSING VALUES

Version 4.0. Statistics Guide. Statistical analyses for laboratory and clinical researchers. Harvey Motulsky

Analyzing Research Data Using Excel

Rank-Based Non-Parametric Tests

Overview of Non-Parametric Statistics PRESENTER: ELAINE EISENBEISZ OWNER AND PRINCIPAL, OMEGA STATISTICS

Statistics for Sports Medicine

Chapter 12 Nonparametric Tests. Chapter Table of Contents

The Dummy s Guide to Data Analysis Using SPSS

Study Guide for the Final Exam

Version 5.0. Statistics Guide. Harvey Motulsky President, GraphPad Software Inc GraphPad Software, inc. All rights reserved.

EPS 625 INTERMEDIATE STATISTICS FRIEDMAN TEST

Nonparametric statistics and model selection

Parametric and Nonparametric: Demystifying the Terms

Statistical tests for SPSS

CHAPTER 14 NONPARAMETRIC TESTS

Difference tests (2): nonparametric

Descriptive Statistics

Data Analysis, Research Study Design and the IRB

Using Excel for inferential statistics

INTERPRETING THE ONE-WAY ANALYSIS OF VARIANCE (ANOVA)

THE KRUSKAL WALLLIS TEST

UNIVERSITY OF NAIROBI

Analyzing Data with GraphPad Prism

Independent samples t-test. Dr. Tom Pierce Radford University

Biostatistics: Types of Data Analysis

Simple Predictive Analytics Curtis Seare

Comparing Means in Two Populations

Nonparametric tests these test hypotheses that are not statements about population parameters (e.g.,

II. DISTRIBUTIONS distribution normal distribution. standard scores

business statistics using Excel OXFORD UNIVERSITY PRESS Glyn Davis & Branko Pecar

X X X a) perfect linear correlation b) no correlation c) positive correlation (r = 1) (r = 0) (0 < r < 1)

Chapter G08 Nonparametric Statistics

StatCrunch and Nonparametric Statistics

Analysis of Variance ANOVA

Nonparametric Two-Sample Tests. Nonparametric Tests. Sign Test

Nonparametric Statistics

Post-hoc comparisons & two-way analysis of variance. Two-way ANOVA, II. Post-hoc testing for main effects. Post-hoc testing 9.

Types of Data, Descriptive Statistics, and Statistical Tests for Nominal Data. Patrick F. Smith, Pharm.D. University at Buffalo Buffalo, New York

Stat 5102 Notes: Nonparametric Tests and. confidence interval

The correlation coefficient

SAS/STAT. 9.2 User s Guide. Introduction to. Nonparametric Analysis. (Book Excerpt) SAS Documentation

Case Study in Data Analysis Does a drug prevent cardiomegaly in heart failure?

CALCULATIONS & STATISTICS

Recall this chart that showed how most of our course would be organized:

DATA ANALYSIS. QEM Network HBCU-UP Fundamentals of Education Research Workshop Gerunda B. Hughes, Ph.D. Howard University

NAG C Library Chapter Introduction. g08 Nonparametric Statistics

Research Methods & Experimental Design

SPSS ADVANCED ANALYSIS WENDIANN SETHI SPRING 2011

DATA INTERPRETATION AND STATISTICS

NCSS Statistical Software

13: Additional ANOVA Topics. Post hoc Comparisons

STATISTICAL ANALYSIS WITH EXCEL COURSE OUTLINE

Testing for differences I exercises with SPSS

training programme in pharmaceutical medicine Clinical Data Management and Analysis

Statistics. One-two sided test, Parametric and non-parametric test statistics: one group, two groups, and more than two groups samples

MEASURES OF LOCATION AND SPREAD

Intro to Parametric & Nonparametric Statistics

Introduction to Analysis of Variance (ANOVA) Limitations of the t-test

Simple Regression Theory II 2010 Samuel L. Baker

Additional sources Compilation of sources:

Introduction to Statistics with GraphPad Prism (5.01) Version 1.1

Bowerman, O'Connell, Aitken Schermer, & Adcock, Business Statistics in Practice, Canadian edition

Point Biserial Correlation Tests

Tutorial 5: Hypothesis Testing

STA-201-TE. 5. Measures of relationship: correlation (5%) Correlation coefficient; Pearson r; correlation and causation; proportion of common variance

DESCRIPTIVE STATISTICS. The purpose of statistics is to condense raw data to make it easier to answer specific questions; test hypotheses.

Business Statistics. Successful completion of Introductory and/or Intermediate Algebra courses is recommended before taking Business Statistics.

Week 4: Standard Error and Confidence Intervals

THE FIRST SET OF EXAMPLES USE SUMMARY DATA... EXAMPLE 7.2, PAGE 227 DESCRIBES A PROBLEM AND A HYPOTHESIS TEST IS PERFORMED IN EXAMPLE 7.

Deciding which statistical test to use:

LAB 4 INSTRUCTIONS CONFIDENCE INTERVALS AND HYPOTHESIS TESTING

Skewed Data and Non-parametric Methods

Analysing Questionnaires using Minitab (for SPSS queries contact -)

Principles of Hypothesis Testing for Public Health

6.4 Normal Distribution

Nonparametric Statistics

statistics Chi-square tests and nonparametric Summary sheet from last time: Hypothesis testing Summary sheet from last time: Confidence intervals

T-TESTS: There are two versions of the t-test:

Statistics in Medicine Research Lecture Series CSMC Fall 2014

CHAPTER 15 NOMINAL MEASURES OF CORRELATION: PHI, THE CONTINGENCY COEFFICIENT, AND CRAMER'S V

Normality Testing in Excel

Session 7 Bivariate Data and Analysis

1.5 Oneway Analysis of Variance

NONPARAMETRIC STATISTICS 1. depend on assumptions about the underlying distribution of the data (or on the Central Limit Theorem)

HYPOTHESIS TESTING: CONFIDENCE INTERVALS, T-TESTS, ANOVAS, AND REGRESSION

Data Analysis Tools. Tools for Summarizing Data

An introduction to IBM SPSS Statistics

Once saved, if the file was zipped you will need to unzip it. For the files that I will be posting you need to change the preferences.

The Statistics Tutor s Quick Guide to

Nonparametric Tests. Chi-Square Test for Independence

CONTINGENCY TABLES ARE NOT ALL THE SAME David C. Howell University of Vermont

Introduction to Minitab and basic commands. Manipulating data in Minitab Describing data; calculating statistics; transformation.

Course Text. Required Computing Software. Course Description. Course Objectives. StraighterLine. Business Statistics

Transcription:

Come scegliere un test statistico Estratto dal Capitolo 37 of Intuitive Biostatistics (ISBN 0-19-508607-4) by Harvey Motulsky. Copyright 1995 by Oxfd University Press Inc. (disponibile in Iinternet) Table 37.1. Selecting a statistical test Goal Describe one group Compare one group to a hypothetical value Compare two unpaired groups Compare two paired groups Compare three me unmatched groups Compare three me matched groups Quantify association between two variables Predict value from another measured variable Predict value from several measured binomial variables Type of Data Measurement (from Gaussian Population) Mean, SD One-sample t test Unpaired t test Rank, Sce, Measurement (from Non- Gaussian Population) Median, interquartile range Wilcoxon test Binomial (Two Possible Outcomes) Proption Chi-square Binomial test ** Mann-Whitney test Fisher's test (chi-square f large samples) Survival Time Kaplan Meier survival curve Log-rank test Mantel- Haenszel* Paired t test Wilcoxon test McNemar's test Conditional proptional hazards One-way ANOVA Repeatedmeasures ANOVA Pearson crelation Simple linear regression Nonlinear regression Multiple linear Multiple nonlinear * Kruskal-Wallis test Chi-square test Cox proptional hazard * Friedman test Cochrane Q** Conditional proptional hazards * Spearman crelation Nonparametric * Contingency coefficients** Simple logistic Multiple logistic Cox proptional hazard Cox proptional hazard

REVIEW OF NONPARAMETRIC TESTS Choosing the right test to compare measurements is a bit tricky, as you must choose between two families of tests: parametric and nonparametric. Many -statistical test are based upon the assumption that the data are sampled from a Gaussian distribution. These tests are referred to as parametric tests. Commonly used parametric tests are listed in the first column of the table and include the t test and analysis of variance. Tests that do not make assumptions about the population distribution are referred to as nonparametrictests. You've already learned a bit about nonparametric tests in previous chapters. All commonly used nonparametric tests rank the outcome variable from low to high and then analyze the ranks. These tests are listed in the second column of the table and include the Wilcoxon, Mann-Whitney test, and Kruskal-Wallis tests. These tests are also called distribution-free tests. CHOOSING BETWEEN PARAMETRIC AND NONPARAMETRIC TESTS: THE EASY CASES Choosing between parametric and nonparametric tests is sometimes easy. You should definitely choose a parametric test if you are sure that your data are sampled from a population that follows a Gaussian distribution (at least approximately). You should definitely select a nonparametric test in three situations: The outcome is a rank a sce and the population is clearly not Gaussian. Examples include class ranking of students, the Apgar sce f the health of newbn babies (measured on a scale of 0 to IO and where all sces are integers), the visual analogue sce f pain (measured on a continuous scale where 0 is no pain and 10 is unbearable pain), and the star scale commonly used by movie and restaurant critics (* is OK, ***** is fantastic). Some values are "off the scale," that is, too high too low to measure. Even if the population is Gaussian, it is impossible to analyze such data with a parametric test since you don't know all of the values. Using a nonparametric test with these data is simple. Assign values too low to measure an arbitrary very low value and assign values too high to measure an arbitrary very high value. Then perfm a nonparametric test. Since the nonparametric test only knows about the relative ranks of the values, it won't matter that you didn't know all the values exactly. The data ire measurements, and you are sure that the population is not distributed in a Gaussian manner. If the data are not sampled from a Gaussian distribution, consider whether you can transfmed the values to make the distribution become Gaussian. F example, you might take the logarithm reciprocal of all values. There are often biological chemical reasons (as well as statistical ones) f perfming a particular transfm. CHOOSING BETWEEN PARAMETRIC AND NONPARAMETRIC TESTS: THE HARD CASES It is not always easy to decide whether a sample comes from a Gaussian population. Consider these points: If you collect many data points (over a hundred so), you can look at the distribution of data and it will be fairly obvious whether the distribution is approximately bell shaped. A fmal statistical test (Kolmogov-Smirnoff test, not explained in this book) can be used to test whether the distribution of the data differs significantly from a Gaussian distribution. With few data points, it is difficult to tell whether the data are Gaussian by inspection, and the fmal test

has little power to discriminate between Gaussian and non-gaussian distributions. You should look at previous data as well. Remember, what matters is the distribution of the overall population, not the distribution of your sample. In deciding whether a population is Gaussian, look at all available data, not just data in the current experiment. Consider the source of scatter. When the scatter comes from the sum of numerous sources (with no one source contributing most of the scatter), you expect to find a roughly Gaussian distribution. When in doubt, some people choose a parametric test (because they aren't sure the Gaussian assumption is violated), and others choose a nonparametric test (because they aren't sure the Gaussian assumption is met). CHOOSING BETWEEN PARAMETRIC AND NONPARAMETRIC TESTS: DOES IT MATTER? Does it matter whether you choose a parametric nonparametric test? The answer depends on sample size. There are four cases to think about: Large sample. What happens when you use a parametric test with data from a nongaussian population? The central limit theem (discussed in Chapter 5) ensures that parametric tests wk well with large samples even if the population is non-gaussian. In other wds, parametric tests are robust to deviations from Gaussian distributions, so long as the samples are large. The snag is that it is impossible to say how large is large enough, as it depends on the nature of the particular non-gaussian distribution. Unless the population distribution is really weird, you are probably safe choosing a parametric test when there are at least two dozen data points in each group. Large sample. What happens when you use a nonparametric test with data from a Gaussian population? Nonparametric tests wk well with large samples from Gaussian populations. The P values tend to be a bit too large, but the discrepancy is small. In other wds, nonparametric tests are only slightly less powerful than parametric tests with large samples. Small samples. What happens when you use a parametric test with data from nongaussian populations? You can't rely on the central limit theem, so the P value may be inaccurate. Small samples. When you use a nonparametric test with data from a Gaussian population, the P values tend to be too high. The nonparametric tests lack statistical power with small samples. Thus, large data sets present no problems. It is usually easy to tell if the data come from a Gaussian population, but it doesn't really matter because the nonparametric tests are so powerful and the parametric tests are so robust. Small data sets present a dilemma. It is difficult to tell if the data come from a Gaussian population, but it matters a lot. The nonparametric tests are not powerful and the parametric tests are not robust. ONE- OR TWO-SIDED P VALUE? With many tests, you must choose whether you wish to calculate a one- two-sided P value (same as one- two-tailed P value). The difference between one- and two-sided P values was discussed in Chapter 10. Let's review the difference in the context of a t test. The P value is calculated f the null hypothesis that the two population means are equal, and any discrepancy between the two sample means is due to chance. If this null hypothesis is true, the one-sided P value is the probability that two sample means would differ as much as was observed ( further) in the direction specified by the hypothesis just by chance, even though the means of the overall populations are actually equal. The two-sided P value also includes the probability that the sample means would differ that much in the opposite direction (i.e., the other group has the larger mean). The two-sided P value is twice the onesided P value.

A one-sided P value is appropriate when you can state with certainty (and befe collecting any data) that there either will be no difference between the means that the difference will go in a direction you can specify in advance (i.e., you have specified which group will have the larger mean). If you cannot specify the direction of any difference befe collecting data, then a two-sided P value is me appropriate. If in doubt, select a two-sided P value. If you select a one-sided test, you should do so befe collecting any data and you need to state the direction of your experimental hypothesis. If the data go the other way, you must be willing to attribute that difference ( association crelation) to chance, no matter how striking the data. If you would be intrigued, even a little, by data that goes in the "wrong" direction, then you should use a two-sided P value. F reasons discussed in Chapter 10, I recommend that you always calculate a two-sided P value. PAIRED OR UNPAIRED TEST? When comparing two groups, you need to decide whether to use a paired test. When comparing three me groups, the term paired is not apt and the term repeated measures is used instead. Use an unpaired test to compare groups when the individual values are not paired matched with one another. Select a paired repeated-measures test when values represent repeated measurements on one subject (befe and after an intervention) measurements on matched subjects. The paired repeated-measures tests are also appropriate f repeated labaty experiments run at different times, each with its own control. You should select a paired test when values in one group are me closely crelated with a specific value in the other group than with random values in the other group. It is only appropriate to select a paired test when the subjects were matched paired befe the data were collected. You cannot base the pairing on the data you are analyzing. FISHER'S TEST OR THE CHI-SQUARE TEST? When analyzing contingency tables with two rows and two columns, you can use either Fisher's exact test the chi-square test. The Fisher's test is the best choice as it always gives the exact P value. The chi-square test is simpler to calculate but yields only an approximate P value. If a computer is doing the calculations, you should choose Fisher's test unless you prefer the familiarity of the chi-square test. You should definitely avoid the chi-square test when the numbers in the contingency table are very small (any number less than about six). When the numbers are larger, the P values repted by the chisquare and Fisher's test will he very similar. The chi-square test calculates approximate P values, and the Yates' continuity crection is designed to make the approximation better. Without the Yates' crection, the P values are too low. However, the crection goes too far, and the resulting P value is too high. Statisticians give different recommendations regarding Yates' crection. With large sample sizes, the Yates' crection makes little difference. If you select Fisher's test, the P value is exact and Yates' crection is not needed and is not available. REGRESSION OR CORRELATION? Linear regression and crelation are similar and easily confused. In some situations it makes sense to perfm both calculations. Calculate linear crelation if you measured both X and Y in each subject

and wish to quantity how well they are associated. Select the Pearson (parametric) crelation coefficient if you can assume that both X and Y are sampled from Gaussian populations. Otherwise choose the Spearman nonparametric crelation coefficient. Don't calculate the crelation coefficient ( its confidence interval) if you manipulated the X variable. Calculate linear regressions only if one of the variables (X) is likely to precede cause the other variable (Y). Definitely choose linear regression if you manipulated the X variable. It makes a big difference which variable is called X and which is called Y, as linear regression calculations are not symmetrical with respect to X and Y. If you swap the two variables, you will obtain a different regression line. In contrast, linear crelation calculations are symmetrical with respect to X and Y. If you swap the labels X and Y, you will still get the same crelation coefficient.