Chapter 5 Analysis of variance SPSS Analysis of variance



Similar documents
SPSS Tests for Versions 9 to 13

INTERPRETING THE ONE-WAY ANALYSIS OF VARIANCE (ANOVA)

Chapter 2 Probability Topics SPSS T tests

Introduction to Analysis of Variance (ANOVA) Limitations of the t-test

Independent t- Test (Comparing Two Means)

ABSORBENCY OF PAPER TOWELS

Main Effects and Interactions

SPSS Resources. 1. See website (readings) for SPSS tutorial & Stats handout

UNDERSTANDING THE TWO-WAY ANOVA

ANOVA ANOVA. Two-Way ANOVA. One-Way ANOVA. When to use ANOVA ANOVA. Analysis of Variance. Chapter 16. A procedure for comparing more than two groups

One-Way ANOVA using SPSS SPSS ANOVA procedures found in the Compare Means analyses. Specifically, we demonstrate

SCHOOL OF HEALTH AND HUMAN SCIENCES DON T FORGET TO RECODE YOUR MISSING VALUES

COMPARISONS OF CUSTOMER LOYALTY: PUBLIC & PRIVATE INSURANCE COMPANIES.

An analysis method for a quantitative outcome and two categorical explanatory variables.

Simple Tricks for Using SPSS for Windows

January 26, 2009 The Faculty Center for Teaching and Learning

IBM SPSS Statistics 20 Part 4: Chi-Square and ANOVA

Examining Differences (Comparing Groups) using SPSS Inferential statistics (Part I) Dwayne Devonish

Multivariate Analysis of Variance. The general purpose of multivariate analysis of variance (MANOVA) is to determine

SPSS Explore procedure

SPSS Guide: Regression Analysis

Linear Models in STATA and ANOVA

Data Analysis in SPSS. February 21, If you wish to cite the contents of this document, the APA reference for them would be

8. Comparing Means Using One Way ANOVA

Profile analysis is the multivariate equivalent of repeated measures or mixed ANOVA. Profile analysis is most commonly used in two cases:

1.5 Oneway Analysis of Variance

Study Guide for the Final Exam

Descriptive Statistics

An introduction to IBM SPSS Statistics

Class 19: Two Way Tables, Conditional Distributions, Chi-Square (Text: Sections 2.5; 9.1)

Two Related Samples t Test

1. What is the critical value for this 95% confidence interval? CV = z.025 = invnorm(0.025) = 1.96

Mixed 2 x 3 ANOVA. Notes

HYPOTHESIS TESTING: CONFIDENCE INTERVALS, T-TESTS, ANOVAS, AND REGRESSION

Research Methods & Experimental Design

The Dummy s Guide to Data Analysis Using SPSS

Chapter 7 Section 7.1: Inference for the Mean of a Population

7. Comparing Means Using t-tests.

One-Way Analysis of Variance (ANOVA) Example Problem

Scatter Plots with Error Bars

A Basic Guide to Analyzing Individual Scores Data with SPSS

Section 13, Part 1 ANOVA. Analysis Of Variance

10. Comparing Means Using Repeated Measures ANOVA

Concepts of Experimental Design

Projects Involving Statistics (& SPSS)

Chapter 7. One-way ANOVA

13: Additional ANOVA Topics. Post hoc Comparisons

Unit 31 A Hypothesis Test about Correlation and Slope in a Simple Linear Regression

Statistiek II. John Nerbonne. October 1, Dept of Information Science

Ordinal Regression. Chapter

Chapter 15. Mixed Models Overview. A flexible approach to correlated data.

SPSS Guide How-to, Tips, Tricks & Statistical Techniques

Additional sources Compilation of sources:

2013 MBA Jump Start Program. Statistics Module Part 3

THE KRUSKAL WALLLIS TEST

Data Analysis Tools. Tools for Summarizing Data

IBM SPSS Statistics for Beginners for Windows

UNDERSTANDING ANALYSIS OF COVARIANCE (ANCOVA)

Regression step-by-step using Microsoft Excel

EPS 625 INTERMEDIATE STATISTICS FRIEDMAN TEST

Multivariate Analysis of Variance (MANOVA)

Analysis of Variance. MINITAB User s Guide 2 3-1

How To Run Statistical Tests in Excel

12: Analysis of Variance. Introduction

KSTAT MINI-MANUAL. Decision Sciences 434 Kellogg Graduate School of Management

Chapter 7. Comparing Means in SPSS (t-tests) Compare Means analyses. Specifically, we demonstrate procedures for running Dependent-Sample (or

Directions for using SPSS

Multiple Regression in SPSS This example shows you how to perform multiple regression. The basic command is regression : linear.

Recall this chart that showed how most of our course would be organized:

SPSS 3: COMPARING MEANS

Simple Linear Regression, Scatterplots, and Bivariate Correlation

SPSS Introduction. Yi Li

SPSS Advanced Statistics 17.0

II. DISTRIBUTIONS distribution normal distribution. standard scores

Multiple-Comparison Procedures

Chapter 9. Two-Sample Tests. Effect Sizes and Power Paired t Test Calculation

PSYCHOLOGY 320L Problem Set #3: One-Way ANOVA and Analytical Comparisons

Chapter 13 Introduction to Linear Regression and Correlation Analysis

One-Way Analysis of Variance

Working with SPSS. A Step-by-Step Guide For Prof PJ s ComS 171 students

Randomized Block Analysis of Variance

This can dilute the significance of a departure from the null hypothesis. We can focus the test on departures of a particular form.

Week 4: Standard Error and Confidence Intervals

SPSS Manual for Introductory Applied Statistics: A Variable Approach

Testing Group Differences using T-tests, ANOVA, and Nonparametric Measures

Multiple Linear Regression

DDBA 8438: The t Test for Independent Samples Video Podcast Transcript

Analysis of Variance ANOVA

Generalized Linear Models

Analysing Questionnaires using Minitab (for SPSS queries contact -)

Good luck! BUSINESS STATISTICS FINAL EXAM INSTRUCTIONS. Name:

Comparing Means in Two Populations

Binary Logistic Regression

IBM SPSS Advanced Statistics 20

SPSS TUTORIAL & EXERCISE BOOK

UNDERSTANDING THE INDEPENDENT-SAMPLES t TEST

This chapter will demonstrate how to perform multiple linear regression with IBM SPSS

Chapter 23. Inferences for Regression

Final Exam Practice Problem Answers

Transcription:

Chapter 5 Analysis of variance SPSS Analysis of variance Data file used: gss.sav How to get there: Analyze Compare Means One-way ANOVA To test the null hypothesis that several population means are equal, based on the results of several independent samples. The test variable is measured on an interval- or ratio scale (for example age), and is grouped by a variable which can be measured on a nominal or discrete ordinal scale (for example life existing of the categories, and ). An independent T test and a one-way ANOVA for two independent samples test the same hypothesis. You must select the dependent variable, and specify the factor to define the different groups. You can move more than one variable into the Dependent List to test all of them. See following figure. Button Options Here you can choose to get descriptives of the data (Descriptive), and to test for equal variances in the groups (Homogeneity-of-variance). Button Post Hoc To see if, and if yes which, groups differ among themselves, there are several possibilities. You can use the Bonferroni procedure (see following figure) when there are equal variances in the groups, which can be tested with the Homogeneity-of-variance test (Button Options).

Output of running one-way ANOVA We performed a one-way ANOVA, with age as dependent variable, and life as factor, which exists of the groups: 0 = Not applicable 1 = 2 = 3 = 8 = Don t know Oneway Descriptives Age of Respondent Total 95% Confidence Interval for Mean N Mean Std. Deviation Std. Error Lower Bound Upper Bound Minimum Maximum 65 52,62 20,059 2,488 47,64 57,59 19 89 457 47,28 18,191,851 45,61 48,95 19 89 471 44,54 16,106,742 43,08 46,00 18 87 993 46,33 17,479,555 45,24 47,42 18 89 Test of Homogeneity of Variances Age of Respondent Levene Statistic df1 df2 Sig. 8,287 2 990,000 Age of Respondent Between Groups Within Groups Total ANOVA Sum of Squares df Mean Square F Sig. 4492,439 2 2246,220 7,448,001 298568,2 990 301,584 303060,6 992 The table Descriptives speaks for itself. In the table Test of Homogeneity of Variances you find the result of Levene s Test for Equality of Variances. It tests the condition that the variances of both samples are equal, indicated by the Levene Statistic. In this statistic, a high value results normally in a significant difference, in this example that is Sig. = 0,000. Strictly speaking, the Bonferroni procedure can therefore not be used, as it assumes equal variances.

However, we are dealing with large samples, which reduces the problem, and the Bonferroni test can be used and interpreted with care. In the table ANOVA the variation (Sum Of Squares), the degrees of freedom (df), and the variance (Mean Square) are given for the within and the between groups, as well as the F value (F) and the significance of the F (Sig.). Sig. indicates whether the null hypothesis the population means are all equal has to be rejected or not. As you can see, there is much difference between the two Mean Squares (2246,220 and 301,584), resulting in a significant difference (F = 7,448; Sig. = 0,001). This means that H 0 must be rejected. Thus: the average age of people who find life,, or are not all equal. But we don t know yet which means differ from each other. Rejecting a null-hypothesis means that NOT ALL population means differ. We don t know yet whether one or more means vary from each other! Therefore, we perform the Bonferoni procedure. The output shows us: Post Hoc Tests Multiple Comparisons Dependent Variable: Age of Respondent Bonferroni (I) Is life exciting or dull (J) Is life exciting or dull *. The mean difference is significant at the.05 level. Mean Difference 95% Confidence Interval (I-J) Std. Error Sig. Lower Bound Upper Bound 5,34 2,302,062 -,18 10,86 8,08* 2,298,001 2,57 13,59-5,34 2,302,062-10,86,18 2,74* 1,140,049,01 5,48-8,08* 2,298,001-13,59-2,57-2,74* 1,140,049-5,48 -,01 The table Multiple Comparisons shows that two out of three groups vary: vs Sig. = 0,062 which is higher than the Sig. level of 0.05. These groups do not vary. vs Sig. = 0,001 which is lower than the Sig. level of 0.05. These groups vary. vs Sig. = 0,049 which is lower than the Sig. level of 0.05. These groups vary (although only just). Because the Bonferroni test assumes equal variances, which does not hold in this case, you can do a test that does not assume equal variances, for example the Tamhane s T2 test. We you perform this test, you will see that in this case it results in the same conclusions. How to get there: Analyze General Linear Model A General Linear Model is, as the name suggest, general in that it incorporates many different models, so that many different tests can be performed. Among these, are the one- and two-way ANOVA, and regression analyses. More difficult designs can be analysed as well. Univariate A univariate GLM is a test with only one dependent variable. There can be one or more independent variable or factors and/or variables.

A one-way ANOVA is a univariate GLM with exactly one independent variable (e.g. fixed factor). A two-way ANOVA is a univariate GLM with exactly two independent variables (e.g. fixed factors). You can test null hypotheses about the effects of other variables on the means of various groupings of a single dependent variable. You can investigate interactions between factors as well as the effects of individual factors. Also, the effects of covariates and covariate interactions with factors can be included. In a univariate GLM you must select in the source variable list the variable you want to test and move it into the Dependent Variable Box. You can select only one dependent variable. Then, select the variables (factors) whose values define the groups and move them into the Fixed Factor box (or Random Factor(s) box if appropriate). If you have covariates, you move them into the Covariate(s) box. To obtain the default univariate GLM that contains all main effects and interactions, click OK. See following figure. Button Plots To obtain plots so that you can examine interaction visually. In the Factors list box, you see all main factors in your model. To plot the means for the values of a singe factor, move that factor into the Horizontal Axis box and click Add. If you want to see the means for all combinations of values of two factors, move the second factor into the Separate Lines Box and click Add. The horizontal-axis factor means will be plotted separately for each value of the second factor. If you move a third factor into the Separate Plots box, separate plots for each value of this factor will be produced. See following figure.

Button Post Hoc When an overall F test has shown significance, you can use post hoc tests to evaluate differences among specific means. To pinpoint differences between all possible pairs of values of a factor variable, select the factors to be tested. Move them to the Post Hoc Tests for: list box. Additionally, select one of the multiple comparison procedures, for example Bonferroni or Tukey. Many different tests are available. For information about a test, point to its name in the dialog box and click the right mouse button. See following figure. Output of running univariate GLM Univariate Analysis of Variance Between-Subjects Factors RS Highest Degree Respondent's Sex 0 1 2 3 4 1 2 Value Label Less than HS 80 High school 480 Junior college N 67 Bachelor 181 Graduate 92 Male 450 Female 450 Tests of Between-Subjects Effects Dependent Variable: Number of Hours Worked Last Week Source Corrected Model Intercept DEGREE SEX DEGREE * SEX Error Total Corrected Total Type III Sum of Squares df Mean Square F Sig. 22211,324 a 9 2467,925 12,904,000 918896,468 1 918896,468 4804,440,000 7409,113 4 1852,278 9,685,000 9296,790 1 9296,790 48,608,000 551,792 4 137,948,721,577 170221,266 890 191,260 1762191,000 900 192432,590 899 a. R Squared =,115 (Adjusted R Squared =,106)

Profile Plots 60 Estimated Marginal Means of Number of Hours Worked Last Week 50 Estimated Marginal Means 40 30 Less than HS High school Junior college Bachelor Respondent's Sex Male Female Graduate RS Highest Degree The Profile Plot is a line plot for the average hours worked. Is there an interaction between gender and education? You see that the two lines don t cross. The shapes of the lines for males and females are quite similar. That suggests that there is no interaction. You don t expect pairs of lines drawn from real data to have exactly the same shape, even if there is no interaction between gender and education in the population. Your goal is to determine whether the interaction observed in the sample is large enough to believe that it also exists in the population. The table Between-Subjects Factors speaks for itself. The table Tests of Between-Subjects Effects is very similar to the one-way analysis of variance table ANOVA. What has changed is the number of hypotheses you are testing. In one-way analysis of variance, you tested a single hypotheses. Now you can test three hypotheses: one about the main effect of degree, one about the main effect of gender, and one about the degree-by-gender interaction. Mean Square for Degree = the variability of the sample means of the five degree groups Mean Square for Gender = the variability of the sample means of the two gender groups Error Mean Square = the variability of the observations within the 10 cell means, that is 5 (Degree) x 2 (Gender). It is a kind of Within Groups Mean Square. Remember: If the null hypotheses for an effect is true, then the corresponding F ratio is expected to be 1. You look at the observed significance level for each observed F ratio to see if you can reject the corresponding null hypotheses. Always, you first have to look at possible interaction effect, since it doesn t make sense to talk about main effects if there is significant interaction between the factors. In this example, Sig. = 0.577, so you validate the null hypotheses that there is no interaction between the two variables. The effect of the type of degree on hours worked seems to be similar for males and females. The absence of interaction tells you that it s reasonable to believe that the difference in average hours worked between males and females is the same for all degree categories. Since you didn t find an interaction between degree and gender, it makes sense to test hypotheses about the main effects of degree and gender. The F statistic for the degree main effect is 9,685. The observed significance level is 0,000. This means that H 0 must be rejected. The variable Degree has influence on the average hours worked. A posthoc test will reveal more about the differences in degree. The F statistic for the Gender main effect is 48,608. The observed significance level is 0,000. This means that H 0 must be rejected. The variable Gender has influence on the average hours worked.