Analysis of Data. Organizing Data Files in SPSS. Descriptive Statistics



Similar documents
Study Guide for the Final Exam

Descriptive Statistics

INTERPRETING THE ONE-WAY ANALYSIS OF VARIANCE (ANOVA)

Statistics Review PSY379

Additional sources Compilation of sources:

Introduction to Analysis of Variance (ANOVA) Limitations of the t-test

ANOVA ANOVA. Two-Way ANOVA. One-Way ANOVA. When to use ANOVA ANOVA. Analysis of Variance. Chapter 16. A procedure for comparing more than two groups

Reporting Statistics in Psychology

UNDERSTANDING THE TWO-WAY ANOVA

SPSS Tests for Versions 9 to 13

THE KRUSKAL WALLLIS TEST

Statistiek II. John Nerbonne. October 1, Dept of Information Science

Section 13, Part 1 ANOVA. Analysis Of Variance

individualdifferences

Research Methods & Experimental Design

13: Additional ANOVA Topics. Post hoc Comparisons

Rank-Based Non-Parametric Tests

SCHOOL OF HEALTH AND HUMAN SCIENCES DON T FORGET TO RECODE YOUR MISSING VALUES

Outline. Definitions Descriptive vs. Inferential Statistics The t-test - One-sample t-test

II. DISTRIBUTIONS distribution normal distribution. standard scores

DESCRIPTIVE STATISTICS. The purpose of statistics is to condense raw data to make it easier to answer specific questions; test hypotheses.

UNDERSTANDING ANALYSIS OF COVARIANCE (ANCOVA)

Section 3 Part 1. Relationships between two numerical variables

UNIVERSITY OF NAIROBI

Introduction to Hypothesis Testing. Hypothesis Testing. Step 1: State the Hypotheses

UNDERSTANDING THE INDEPENDENT-SAMPLES t TEST

THE UNIVERSITY OF TEXAS AT TYLER COLLEGE OF NURSING COURSE SYLLABUS NURS 5317 STATISTICS FOR HEALTH PROVIDERS. Fall 2013

The Dummy s Guide to Data Analysis Using SPSS

Statistics. Measurement. Scales of Measurement 7/18/2012

STATISTICS FOR PSYCHOLOGISTS

SPSS Explore procedure

1 Nonparametric Statistics


Outline. Topic 4 - Analysis of Variance Approach to Regression. Partitioning Sums of Squares. Total Sum of Squares. Partitioning sums of squares

Testing Group Differences using T-tests, ANOVA, and Nonparametric Measures

Post-hoc comparisons & two-way analysis of variance. Two-way ANOVA, II. Post-hoc testing for main effects. Post-hoc testing 9.

Using Excel for inferential statistics

1. What is the critical value for this 95% confidence interval? CV = z.025 = invnorm(0.025) = 1.96

Introduction to Statistics and Quantitative Research Methods

Non-Parametric Tests (I)

DATA INTERPRETATION AND STATISTICS

Two-sample hypothesis testing, II /16/2004

Correlation Coefficient The correlation coefficient is a summary statistic that describes the linear relationship between two numerical variables 2

Chapter 5 Analysis of variance SPSS Analysis of variance

1.5 Oneway Analysis of Variance

Introduction to Quantitative Methods

EPS 625 INTERMEDIATE STATISTICS FRIEDMAN TEST

Introduction to. Hypothesis Testing CHAPTER LEARNING OBJECTIVES. 1 Identify the four steps of hypothesis testing.

Analysing Questionnaires using Minitab (for SPSS queries contact -)

Foundation of Quantitative Data Analysis

DATA ANALYSIS. QEM Network HBCU-UP Fundamentals of Education Research Workshop Gerunda B. Hughes, Ph.D. Howard University

PSYCHOLOGY 320L Problem Set #3: One-Way ANOVA and Analytical Comparisons

Profile analysis is the multivariate equivalent of repeated measures or mixed ANOVA. Profile analysis is most commonly used in two cases:

Unit 31 A Hypothesis Test about Correlation and Slope in a Simple Linear Regression

Chapter 9. Two-Sample Tests. Effect Sizes and Power Paired t Test Calculation

Parametric and non-parametric statistical methods for the life sciences - Session I

Data analysis process

Error Type, Power, Assumptions. Parametric Tests. Parametric vs. Nonparametric Tests

Overview of Non-Parametric Statistics PRESENTER: ELAINE EISENBEISZ OWNER AND PRINCIPAL, OMEGA STATISTICS

research/scientific includes the following: statistical hypotheses: you have a null and alternative you accept one and reject the other

SPSS ADVANCED ANALYSIS WENDIANN SETHI SPRING 2011

SPSS 3: COMPARING MEANS

UNDERSTANDING THE DEPENDENT-SAMPLES t TEST

Examining Differences (Comparing Groups) using SPSS Inferential statistics (Part I) Dwayne Devonish

Descriptive Analysis

1/27/2013. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2

T-test & factor analysis

Analysis of Variance ANOVA

Multivariate Analysis of Variance (MANOVA)

Two-Sample T-Tests Assuming Equal Variance (Enter Means)

Chapter 7. One-way ANOVA

Part 2: Analysis of Relationship Between Two Variables

Correlational Research. Correlational Research. Stephen E. Brock, Ph.D., NCSP EDS 250. Descriptive Research 1. Correlational Research: Scatter Plots

Nonparametric Two-Sample Tests. Nonparametric Tests. Sign Test

Multiple-Comparison Procedures

STA-201-TE. 5. Measures of relationship: correlation (5%) Correlation coefficient; Pearson r; correlation and causation; proportion of common variance

Experimental Designs (revisited)

Research Methodology: Tools

The Statistics Tutor s Quick Guide to

Assumptions. Assumptions of linear models. Boxplot. Data exploration. Apply to response variable. Apply to error terms from linear model

Experimental Design for Influential Factors of Rates on Massive Open Online Courses

ELEMENTARY STATISTICS

DISCRIMINANT FUNCTION ANALYSIS (DA)

Analyzing Research Data Using Excel

Measurement & Data Analysis. On the importance of math & measurement. Steps Involved in Doing Scientific Research. Measurement

Biostatistics: Types of Data Analysis

Pearson's Correlation Tests

Week 3&4: Z tables and the Sampling Distribution of X

Two-Sample T-Tests Allowing Unequal Variance (Enter Difference)

Premaster Statistics Tutorial 4 Full solutions

CORRELATIONAL ANALYSIS: PEARSON S r Purpose of correlational analysis The purpose of performing a correlational analysis: To discover whether there

Analysis of Questionnaires and Qualitative Data Non-parametric Tests

Data Mining and Data Warehousing. Henryk Maciejewski. Data Mining Predictive modelling: regression

HYPOTHESIS TESTING: POWER OF THE TEST

15. Analysis of Variance

Standard Deviation Estimator

Handbook of Parametric and Nonparametric Statistical Procedures

Difference tests (2): nonparametric

WHAT IS A JOURNAL CLUB?

Transcription:

Analysis of Data Claudia J. Stanny PSY 67 Research Design Organizing Data Files in SPSS All data for one subject entered on the same line Identification data Between-subjects manipulations: variable to identify group level for that subject Measures for Within-subjects manipulations: each level of a within subject variable appears as a new variable in the data file ID SEX AGE HAND WD WD WD3 PIC PIC PIC3 RTW RTW RTW3 RTP RTP RTP3 3 4 6 67 7 7 3 49 6 9 98 88 798 68 773 8 3 4 6 Descriptive Statistics Biased and Unbiased Estimates of Population parameters Measures of Central Tendency Operational definitions of a typical score Mean, median, mode Measures of Variability Range, interquartile range Variance Standard deviation

Measures of Association Pearson product-moment correlation (r) Point-biserial correlation Equivalent to a between-subjects t-test Spearman rank order correlation (r s ) Phi coefficient Interpreting Correlations College GPA ATM transactions 4. 3..... 3 3. 4 HS GPA 8 6 4 7 Temperature in Pensacola Crop Yield (bushels) Days in Therapy 8 6 4 8 6 4 3 4 6 7 8 9 Patient's Belief in Doctor's Competence Inches of Rainfall per Month Effects of Outliers 3 3 Scores on a Test Scores on a Test 4 6 Hours Studied 4 6 Hours Studied

Effects of Outliers 7 7 7 7 Restricted Range 8 8 7 6 7 6 Height (in) 4 3 Height (in) 4 3 Age in years Age in years Combining Data From Different Groups Weight (lbs) 9 6 4 9 6 4 9 6 7 Height (inches) Weight (lbs) 9 6 4 9 6 4 9 Women Men 6 7 Height (inches) 3

Data From Different Groups 9 9 8 8 Interest in Remarriage 7 6 4 3 Interest in Remarriage 7 6 4 3 Widowers Widows 4 6 8 Age at Death of Spouse 4 6 8 Age at Death of Spouse Inferential Statistics Logic of Inferential Statistical Tests Assume that all participants are selected from the same population Score (DV) = µ + random error H : µ = µ Assume that the experimental conditions produce systematic changes in the scores of participants on the dependent measure Score (DV) = µ + treatment effect + random error H : µ µ Types of Statistical Errors Type I Error Probability = alpha Rejection of H when H is true Type II Error Probability = beta Failure to reject H when H is true Failure to detect a true treatment effect 4

Parametric Statistics Random selection and/or random assignment of scores from the population of interest Sampling distribution is normal Central Limit Theorem: sampling distribution will become normal as sample size increases, even when raw data are not distributed normally Homogeneity of variances within groups Problems can arise when observations in one condition are more variable than observations in other conditions Testing these assumptions statistically Analysis of Variance One-way designs One independent variable (factor) two levels of a factor a t-test will give equivalent results three or more levels of a factor requires use of ANOVA Partitioning Variance Between-groups variability Within-groups variability Partitioning Variance in ANOVA Total Variability Systematic Variability (Between Groups) Error Variability (Within Groups)

Creating an F Ratio for ANOVA SS total = SS treatment + SS error Mean Square = SS / df Expected MS BG = Var TX + Var error Expected MS WG = Var error F Ratio = MS BG / MS WG = (Var TX + Var error )/Var error Value of F ratio when H is true? Value of F ratio when H is false? Interpreting a Significant F Ratio Need to evaluate pair-wise comparisons of means following a significant F ratio Planned comparisons A priori comparisons Comparisons based on specific predictions hypothesized before data collection Unplanned comparisons Post hoc tests Need to control Experiment-wide Error Rate (familywise error rate) Probability of at least one Type I Error increases as the number of comparisons increases Planned Comparisons Contrasts should be orthogonal Outcome of each contrast should be independent Knowing the result for one comparison does not constrain the decision for other comparisons Example: Planned comparisons for a three group design: Drug A, Drug B, Control Contrast : Average of Drug versus No Drug Combines mean for drug A and drug B Contrast : Drug A versus Drug B 6

Orthogonal Contrasts Produce Independent Outcomes Assume that the comparison of the Drugs versus Control produces no difference Mean for Drugs = Mean for Control = 3 outcomes possible for the comparison of Drug A with Drug B No Difference Drug A = Drug B = Drug A > Drug B Drug A = 7 Drug B = 3 Drug A < Drug B Drug A = Drug B = 8 Knowing the decision about the first contrast does not provide information about the outcome of the second contrast Independence of Contrast Outcomes (continued) Assume that the comparison of the Drugs versus Control produces a significant difference Mean for Drugs = 4 Mean for Control = 8 3 outcomes possible for the comparison of Drug A with Drug B No Difference Drug A = 4 Drug B = 4 Drug A > Drug B Drug A = 6 Drug B = Drug A < Drug B Drug A = Drug B = 7 Knowing the decision about the first contrast does not provide information about the outcome of the second contrast Post Hoc Tests Scheffe Corrects for all possible comparisons, whether or not these are examined Bonferroni procedure Divide alpha by the number of comparisons to be made Use the critical value for the smaller alpha value for all tests Dunnett Specifically designed to compare each experimental group with a single control group 7

Post Hoc Tests Tukey HSD (default post hoc in SPSS) Controls EER over a set of pair-wise comparisons (adjusts critical value for significance) Specialized procedures available for calculation for means based on unequal sample sizes Newman-Keuls Less conservative than Tukey HSD Means in a comparison set are rank ordered Critical value for a pair-wise comparison varies with the number of steps between means in the ordered set Duncan ANOVA Two Factor Designs Additional factors require partitioning variance in different ways Nature of the partitioning will change depending on the type of design Both factors manipulated between-subjects Both factors manipulated within-subjects Mixed designs Logic of the test remains the same: Isolate the systematic effect in the numerator of the F ratio Partitioning Variance in the Between- Subjects Design 8

Partitioning Variance in the Within- Subjects Design Partitioning in a Mixed Design Fixed Effects & Random Effects Fixed Effects All levels of a factor are in the design men & women All relevant levels of a factor are in the design zero stress, moderate stress, high stress Random Effects Levels of a factor represent a sample of the possible levels in the design Individual subjects as levels of a factor Unique stimulus materials (e.g., photos, sentences, etc. used to create a manipulation) Analysis problems when or more factors are random effects 9

Interaction Effects Additive Model (no interaction) Score = µ + TXa + TXb + error Non-Additive Model (interaction) Score = µ + TXa + TXb + TXab + error Meaning of interactions: effect of a treatment varies depending on conditions created by manipulation of a second variable x 3 Factorial with No Interaction Mean Words Recalled 4 8 6 4 Men Women sad neutral happy Emotional Value of Words x Factorial with Significant Interaction 3 Mean Recall 3 sad at test happy at test sad Mood at Study happy

Other Forms of Interactions 3 4 6 7 8 9 6 4 3 3 3 Factors that Affect Statistical Power Selection of alpha level Comparing alpha =. to alpha =. Comparing one tailed tests to two-tailed tests Size of the treatment effect produced Variability of the sampling distribution for the statistical test Effects of changes in sample size Effects of changes in control of extraneous variables Power Changes with Size of Effect Created

Power Changes with Variability in the Sampling Distribution Measuring Effect Sizes (Abelson) Raw effect size Differences created measured in the same units of measurement used for the dependent measure Difference in IQ, number of words recalled, etc. Difficulty of knowing what counts as a large effect Standardized effect size Effect sizes converted to z-scores Creates equivalent scales for comparing different dependent measures Useful for meta-analysis Cohen s d, variance explained (r ), omega squared Interpreting Statistical Significance Large sample sizes can provide powerful tests in which small numeric differences are statistically reliable How should these be interpreted? Numerically small effects can be impressive Created with minimal manipulation of the IV Created when the DV is difficult to change at all Effects of any size have an important theoretical interpretation

Statistical Significance vs Practical Significance What are the practical implications of an effect? An educational program reliably produces a point increase in the IQ of -year-old children A medical intervention reduces the risk of heart attack from per, (p(heart attack) =.) by half (to p(heart attack) =.) In a population of million people, this will translate into a change from, heart attacks to heart attacks Data Transformations Linear transformations Add a constant value Shifts location of distribution without changing the variance of the distribution Can make data analysis easier (K vs ) Other transformations change both the mean and the variance of the distribution Transforming data can produce data that meet the distribution assumptions for statistical tests Non-Linear Data Transformations Square Root transformation Used when larger means correlate with larger variances (common with count data) Makes variances more homogeneous Reduces skew of distributions Arcsine transformation Used when data consist of proportions (e.g., percent correct data) 3

Non-Linear Data Transformations Log transformation Reduces skew in distributions Common with sensory data gives manageable ranges for plotting results Reciprocal transformation X = /X Reciprocal data are less variable than raw data Transforms alter the interpretation of the measure Reciprocal of a time measure is a speed measure New measure is sometimes easier to interpret GSR gets smaller as stress increases EDR (/GSR) increases with increased stress Non-Parametric Tests Do not make assumptions about the underlying distributions to compute probabilities Useful when data do not meet assumptions for normality, etc. Can be more powerful tests than parametric tests when assumptions have been violated Non-parametric analogs exists for nearly every parametric test 4