THE KRUSKAL WALLLIS TEST



Similar documents
INTERPRETING THE ONE-WAY ANALYSIS OF VARIANCE (ANOVA)

EPS 625 INTERMEDIATE STATISTICS FRIEDMAN TEST

UNDERSTANDING THE INDEPENDENT-SAMPLES t TEST

1 Nonparametric Statistics

SCHOOL OF HEALTH AND HUMAN SCIENCES DON T FORGET TO RECODE YOUR MISSING VALUES

Non-Parametric Tests (I)

SPSS Tests for Versions 9 to 13

Rank-Based Non-Parametric Tests

13: Additional ANOVA Topics. Post hoc Comparisons

Statistiek II. John Nerbonne. October 1, Dept of Information Science

Parametric and non-parametric statistical methods for the life sciences - Session I

NCSS Statistical Software

The Dummy s Guide to Data Analysis Using SPSS

Independent t- Test (Comparing Two Means)

Overview of Non-Parametric Statistics PRESENTER: ELAINE EISENBEISZ OWNER AND PRINCIPAL, OMEGA STATISTICS

Chapter 5 Analysis of variance SPSS Analysis of variance

Introduction to Analysis of Variance (ANOVA) Limitations of the t-test

Testing for differences I exercises with SPSS

UNDERSTANDING THE DEPENDENT-SAMPLES t TEST

Analysis of Data. Organizing Data Files in SPSS. Descriptive Statistics

Comparing Means in Two Populations

CHAPTER 14 NONPARAMETRIC TESTS

SPSS Explore procedure

Projects Involving Statistics (& SPSS)

Research Methodology: Tools

SPSS ADVANCED ANALYSIS WENDIANN SETHI SPRING 2011

Nonparametric tests these test hypotheses that are not statements about population parameters (e.g.,

Recall this chart that showed how most of our course would be organized:

SPSS Guide How-to, Tips, Tricks & Statistical Techniques

Two Related Samples t Test

SPSS 3: COMPARING MEANS

Nonparametric Statistics

Normality Testing in Excel

Nonparametric Statistics

Biostatistics: DESCRIPTIVE STATISTICS: 2, VARIABILITY

The Kruskal-Wallis test:

Nonparametric Two-Sample Tests. Nonparametric Tests. Sign Test

Study Guide for the Final Exam

1.5 Oneway Analysis of Variance

Analysis of Variance ANOVA

Chapter 7 Section 7.1: Inference for the Mean of a Population

Statistics. One-two sided test, Parametric and non-parametric test statistics: one group, two groups, and more than two groups samples

Additional sources Compilation of sources:

UNDERSTANDING THE TWO-WAY ANOVA

The Statistics Tutor s Quick Guide to

COMPARISONS OF CUSTOMER LOYALTY: PUBLIC & PRIVATE INSURANCE COMPANIES.

Analysing Questionnaires using Minitab (for SPSS queries contact -)

Descriptive Statistics

One-Way Analysis of Variance (ANOVA) Example Problem

Chapter 2 Probability Topics SPSS T tests

Difference tests (2): nonparametric

Tutorial 5: Hypothesis Testing

Statistics for Sports Medicine

Skewed Data and Non-parametric Methods

NCSS Statistical Software

Examining Differences (Comparing Groups) using SPSS Inferential statistics (Part I) Dwayne Devonish

1. What is the critical value for this 95% confidence interval? CV = z.025 = invnorm(0.025) = 1.96

Section 13, Part 1 ANOVA. Analysis Of Variance

II. DISTRIBUTIONS distribution normal distribution. standard scores

Two-Sample T-Tests Allowing Unequal Variance (Enter Difference)

Research Methods & Experimental Design

Come scegliere un test statistico

Outline. Topic 4 - Analysis of Variance Approach to Regression. Partitioning Sums of Squares. Total Sum of Squares. Partitioning sums of squares

NONPARAMETRIC STATISTICS 1. depend on assumptions about the underlying distribution of the data (or on the Central Limit Theorem)

How To Test For Significance On A Data Set

Friedman's Two-way Analysis of Variance by Ranks -- Analysis of k-within-group Data with a Quantitative Response Variable

Two-Sample T-Tests Assuming Equal Variance (Enter Means)

SIMPLE LINEAR CORRELATION. r can range from -1 to 1, and is independent of units of measurement. Correlation can be done on two dependent variables.

Statistical tests for SPSS

Chapter 7. One-way ANOVA

T-test & factor analysis

Introduction. Hypothesis Testing. Hypothesis Testing. Significance Testing

Outline. Definitions Descriptive vs. Inferential Statistics The t-test - One-sample t-test

Final Exam Practice Problem Answers

The Wilcoxon Rank-Sum Test

Data analysis process

Likert Scales. are the meaning of life: Dane Bertram

Reporting Statistics in Psychology

Chapter 12 Nonparametric Tests. Chapter Table of Contents

An Empirical Examination of the Relationship between Financial Trader s Decision-making and Financial Software Applications

" Y. Notation and Equations for Regression Lecture 11/4. Notation:

Intro to Parametric & Nonparametric Statistics

Introduction to Hypothesis Testing. Hypothesis Testing. Step 1: State the Hypotheses

TABLE OF CONTENTS. About Chi Squares What is a CHI SQUARE? Chi Squares Hypothesis Testing with Chi Squares... 2

individualdifferences

NCSS Statistical Software. One-Sample T-Test

QUANTITATIVE METHODS BIOLOGY FINAL HONOUR SCHOOL NON-PARAMETRIC TESTS

UNDERSTANDING ANALYSIS OF COVARIANCE (ANCOVA)

Using Excel for inferential statistics

How Far is too Far? Statistical Outlier Detection

MASTER COURSE SYLLABUS-PROTOTYPE PSYCHOLOGY 2317 STATISTICAL METHODS FOR THE BEHAVIORAL SCIENCES

HYPOTHESIS TESTING WITH SPSS:

Testing Group Differences using T-tests, ANOVA, and Nonparametric Measures

(and sex and drugs and rock 'n' roll) ANDY FIELD

Standard Deviation Estimator

One-Way Analysis of Variance

January 26, 2009 The Faculty Center for Teaching and Learning

PSYCHOLOGY 320L Problem Set #3: One-Way ANOVA and Analytical Comparisons

CALCULATIONS & STATISTICS

Transcription:

THE KRUSKAL WALLLIS TEST TEODORA H. MEHOTCHEVA Wednesday, 23 rd April 08

THE KRUSKAL-WALLIS TEST: The non-parametric alternative to ANOVA: testing for difference between several independent groups 2

NON PARAMETRIC TESTS: CHARACTERISTICS Distribution-free tests? Not exactly, they just less restrictive than parametric tests Based on ranked data By ranking the data we lose some information about the magnitude of difference between scores the non-parametric tests are less powerful than their parametric counterparts, i.e. a parametric test is more likely to detect a genuine effect in the data, if there is one, than a non-parametric test. 3

WHEN TO USE KRUSKAL-WALLIS We want to compare several independent groups but we don t meet some of the assumptions made in ANOVA: Data should come from a normal distribution Variances should be fairly similar (Levene s test) 4

EXAMPLE: EFFECT OF WEED ON CROP 4 groups: 0 weeds/meter 4 samples x group (N16) 1 weed/meter 3 weeds/meter 9 weeds/meter Weeds Corn Weeds Corn Weeds Corn Weeds Corn 0 166.7 1 166.2 3 158.6 9 162.8 0 172.2 1 157.3 3 176.4 9 142.4 0 165.0 1 166.7 3 153.1 9 162.7 0 176.9 1 161.1 3 165.0 9 162.4 Corn crop by weeds (Ex. 15.13, Moore & McCabe, 2005) 5

EFFECT OF WEED ON CROP: EXPLORING THE DATA Weeds n Mean Std.dev. 0 4 170.200 5.422 1 4 162.825 4.469 3 4 161.025 10.493 9 4 157.575 10.118 Summary statistics for Effect of Weed on Crop! For ANOVA: the largest standard deviation should NOT exceed twice the smallest. 6

EFFECT OF WEED ON CROP: EXPLORING THE DATA: Q-Q PLOTS Ex. 15.9, Moore & McCabe, 2005 7

KRUSKAL-WALLIS: HYPOTHESISING H 0 : All four populations have the same median yield. H a : Not all four median yields are equal.! ANOVA F: H 0 : µ0 = µ1 = µ3 = µ9 H a : not all four means are equal. 8

KRUSKAL-WALLIS: HYPOTHESISING Non-parametric tests hypothesize about the median instead of the mean (as parametric tests do). mean a hypothetical value not necessarily present in the data (µ= Σ x i / n) median the middle score of a set of ordered observations. In the case of even number of observations, the median is the average of the two scores on each side of what should be in the middle The mean is more sensitive to outliers than the median. Ex: 1, 5, 2, 8, 38 µ= 10,7 [(1+5+2+8+38)/5] median = 5 (1, 2, 5, 8, 38)! Median when the observations are even: Ex: 5, 2, 8, 38 = 6,5 (2, 5, 8, 38) (5+8)/2 = 6,5 9

THE KRUSKAL-WALLIS TEST: THE THEORY We take the responses from all groups and rank them; then we sum up the ranks for each group and we apply one way ANOVA to the ranks, not to the original observations. We order the scores that we have from lowest to highest, ignoring the group that the scores come from, and then we assign the lowest score a rank of 1, the next highest a rank of 2 and so on. 10

EFFECTS OF WEED ON CROP: KRUSKAL-WALLIS TEST: RANKING THE DATA Score 142,4 157,3 158,6 161,1 166,2 166,7 166,7 172,2 Rank 1 2 3 4 11 12 13 14 Act. Rank 1 2 3 4 11 12,5 12,5 14 Group 9 3 3 1 1 0 1 0! Repeated values (tied ranks) are ranked as the average of the potential ranks for those scores, i.e. (12+13)/2=12,5 11

EFFECTS OF WEED ON CROP: KRUSKAL-WALLIS TEST: RANKS When the data are ranked we collect the scores back in their groups and add up the ranks for each group = R i (i determines the particular group) Weeds Ranks Sum of ranks 0 10 12,5 14 16 52,5 1 4 6 11 12,5 33,5 3 2 3 5 15 25,0 9 1 7 8 9 25,0 Ex. 15.14, Moore & McCabe, 2005 12

THE KRUSKAL-WALLIS TEST: THE THEORY! In ANOVA, we calculate the total variation (total sum of squares, SST) by adding up the variation among the groups (sum of squares for groups, SSG) with the variation within group (sum of squares for error, SSE): SST=SSG+SSE In Kruskal-Wallis: one way ANOVA to the ranks, not the original scores. If there are N observations in all, the ranks are always the whole numbers from 1 to N. The total sum of squares for the ranks is therefore a fixed number no matter what the data are no need to look at both SSG and SSE Kruskal-Wallis = SSG for the ranks 13

KRUSKAL-WALLIS TEST: H STATISTIC The test statistic H is calculated: H = 12 Σ R i 3(N + 1) N (N+1) n i 2 The Kruskal-Wallis test rejects the H o when H is large. 14

EFFECTS OF WEED ON CROP: KRUSKAL-WALLIS TEST: H STATISTIC Our Example: I = 4, N = 16, n i =4, R = 52.5, 33.5, 25.0, 25.0 H = 12 Σ R i 3(N + 1) N (N+1) n i 2 = 12 ( 52.52 33.5 2 25 2 25 2 ) 3(17) (16)(17) 4 4 4 4 = 12 (689.0625 + 280.5625 + 156.25 + 156.25) 51 272 = 0.0441(1282,125) 51 = 56.5643 51 = 5.56 15

EFFECTS OF WEED ON CROP: KRUSKAL-WALLIS TEST: P VALUE H has approximately the chi-square distribution with k 1 degrees of freedom df = 3 (4-1) & H = 5.56 0.10 < P > 0.15 16

THE STUDY: THE EFFECT OF SOYA ON CONCENTRATION 4 groups: O Soya Meals per week 1 Soya Meal per week 4 Soya Meals per week 9 Soya Meals per week 20 participants per group N=80 Tested after one year: RT when naming words 17

THE EFFECT OF SOYA ON CONCENTRATION: EXPLORATORY STATISTICS Soya n Mean Std.dev. 0 20 4.9868 5.08437 1 20 4.6052 4.67263 4 20 4.1101 4.40991 9 20 1.6530 1.10865 Summary Statistics for Soya on Concentration! Violation of the rule of thumb for using ANOVA: the largest standard deviation should NOT exceed twice the smallest. 18

THE EFFECT OF SOYA ON CONCENTRATION: TEST OF NORMALITY Tests of Normality RT (Ms) Number of Soya Meals Per Week No Soya Meals 1 Soya Meal Per Week 4 Soyal Meals Per Week 7 Soya Meals Per Week a. Lilliefors Significance Correction Kolmogorov-Smirnov a Shapiro-Wilk Statistic df Sig. Statistic df Sig.,181 20,085,805 20,001,207 20,024,826 20,002,267 20,001,743 20,000,204 20,028,912 20,071 Significance of data the distribution is significantly different from a normal distribution, i.e. it is non-normal. 19

THE EFFECT OF SOYA ON CONCENTRATION: HOMOGENEITY OF VARIANCE Test of Homogeneity of Variance RT (Ms) Based on Mean Based on Median Based on Median and with adjusted df Based on trimmed mean Levene Statistic df1 df2 Sig. 5,117 3 76,003 2,860 3 76,042 2,860 3 58,107,045 4,070 3 76,010 Significance of data the variances in different groups are significantly different data are not homogenous 20

21 KRUSKAL-WALLIS: SPSS

22 KRUSKAL-WALLIS: SPSS

THE EFFECT OF SOYA ON CONCENTRATION: SPSS: RANKS Ranks RT (Ms) Number of Soya Meals No Soya Meals 1 Soya Meal Per Week 4 Soyal Meals Per Week 7 Soya Meals Per Week Total N Mean Rank 20 46,35 20 44,15 20 44,15 20 27,35 80 23

THE EFFECT OF SOYA ON CONCENTRATION: SPSS: TEST STATISTICS Test Statistics b,c Chi-Square df Asymp. Sig. Monte Carlo Sig. a. b. c. Sig. 99% Confidence Interval Lower Bound Upper Bound Based on 10000 sampled tables with starting seed 2000000. Kruskal Wallis Test Grouping Variable: Number of Soya Meals Per Week RT (Ms) 8,659 3,034,031 a,027,036 Test significance p <.034 Confidence Interval.028-.037 does not cross the boundary of.05 a lot of confidence that the significant effect is genuine 24

THE EFFECT OF SOYA ON CONCENTRATION: THE CONCLUSION We know that there is difference but we don t know exactly where! 25

KRUSKAL-WALLIS: FINDING THE DIFFERENCE 25,00 20 20,00 40 60 15,00 s) (M T R 10,00 59 5,00 0,00 No Soya Meals 1 Soya Meal Per Week 4 Soyal Meals Per Week Number of Soya Meals Per Week 7 Soya Meals Per Week 26

KRUSKAL-WALLIS: POST HOC TESTS: MANN-WHITNEY TEST Mann-Whitney tests = Wilcoxon Rank Sum Test a non-parametric test for comparing two independent groups based on ranking! Lots of Wilcoxon Rank Sum Tests inflation of the Type I error (the probability of falsely rejecting the H 0 ) Bonferroni correction -.05/ the number of test to be conducted The value of significance becomes too small, i.e.: 0 soya meals, 1 soya meal, 4 soya meals, 7 soya meals = 6 tests.05/6=.0083 27

KRUSKAL-WALLIS: POST HOC TESTS: MANN-WHITNEY TEST Select a number of comparisons to make, i.e.: Test 1: 1 soya meal per week compared to 0 soya meals Test 2: 4 soya meals per week compared to 0 soya meals Test 3: 7 soya meals per week compared to 0 soya meals α level =.05/3=.0167 28

KRUSKAL-WALLIS: POST HOC TESTS: MANN-WHITNEY TEST 1. 0 soya vs. 1 meal per week 2. 0 soya vs. 4 meals per week Test Statistics b Test Statistics b Mann-Whitney U Wilcoxon W Z Asymp. Sig. (2-tailed) Exact Sig. [2*(1-tailed Sig.)] a. Not corrected for ties. RT (Ms) 191,000 401,000 -,243,808,820 a b. Grouping Variable: Number of Soya Meals Per Week 3. 0 soya vs. 7 meals per week Mann-Whitney U Wilcoxon W Z Test Statistics b Asymp. Sig. (2-tailed) Exact Sig. [2*(1-tailed Sig.)] a. Not corrected for ties. RT (Ms) 104,000 314,000-2,597,009,009 a b. Grouping Variable: Number of Soya Meals Per Week Mann-Whitney U Wilcoxon W Z Asymp. Sig. (2-tailed) Exact Sig. [2*(1-tailed Sig.)] a. Not corrected for ties. RT (Ms) 188,000 398,000 -,325,745,758 a b. Grouping Variable: Number of Soya Meals Per Week! α level =.0167 not.05 29

KRUSKAL-WALLIS: TESTING FOR TRENDS: JONCKHEERE-TERPSTRA TEST If we expect that the groups we compare are ordered in a certain way. I.e. the more soya a person eats the more concentrated and faster they become (shorter RTs) 30

KRUSKAL-WALLIS: TESTING FOR TRENDS: JONCKHEERE- TERPSTRA TEST Jonckheere-Terpstra Test b Number of Levels in Number of Soya Meals Per Week N Observed J-T Statistic Mean J-T Statistic Std. Deviation of J-T Statistic Std. J-T Statistic Asymp. Sig. (2-tailed) Monte Carlo Sig. (2-tailed) Monte Carlo Sig. (1-tailed) Sig. 99% Confidence Interval Sig. 99% Confidence Interval Lower Bound Upper Bound Lower Bound Upper Bound a. Based on 10000 sampled tables with starting seed 2000000. b. Grouping Variable: Number of Soya Meals Per Week RT (Ms) 4 80 912,000 1200,000 116,333-2,476,013,012 a,009,015,006 a,004,008 normal distribution z score calculated: -2,476 If > 1.65 significant result. - descending medians scores get smaller + ascending medians scores get bigger Medians get smaller the more soya meals we eat : RTs become faster more soya better concentration and 31 more speed!

CALCULATING AN EFFECT SIZE A standardized measure of the magnitude of the observed effect the measured effect is meaningful or important Cohen s d or Pearson s correlation coefficient r: 1 > r < 0 r =.10 small effect 1% of the total variance r =.30 medium effect 9 %of total variance r =.50 large effect 25 %of total varince Converting z score into the effect size estimate r = Z N 32

CALCULATING AN EFFECT SIZE 33