2 Sample t-test (unequal sample sizes and unequal variances)



Similar documents
Paired 2 Sample t-test

How To Test For Significance On A Data Set

Data Transforms: Natural Logarithms and Square Roots

Lesson 1: Comparison of Population Means Part c: Comparison of Two- Means

1.5 Oneway Analysis of Variance

THE FIRST SET OF EXAMPLES USE SUMMARY DATA... EXAMPLE 7.2, PAGE 227 DESCRIBES A PROBLEM AND A HYPOTHESIS TEST IS PERFORMED IN EXAMPLE 7.

Two-Sample T-Tests Assuming Equal Variance (Enter Means)

Understand the role that hypothesis testing plays in an improvement project. Know how to perform a two sample hypothesis test.

Independent t- Test (Comparing Two Means)

Odds ratio, Odds ratio test for independence, chi-squared statistic.

Skewed Data and Non-parametric Methods

Projects Involving Statistics (& SPSS)

12: Analysis of Variance. Introduction

Stats for Strategy Fall 2012 First-Discussion Handout: Stats Using Calculators and MINITAB

DDBA 8438: The t Test for Independent Samples Video Podcast Transcript

1. What is the critical value for this 95% confidence interval? CV = z.025 = invnorm(0.025) = 1.96

Two-Sample T-Tests Allowing Unequal Variance (Enter Difference)

Biostatistics: DESCRIPTIVE STATISTICS: 2, VARIABILITY

CALCULATIONS & STATISTICS

Study Guide for the Final Exam

Unit 31 A Hypothesis Test about Correlation and Slope in a Simple Linear Regression

4. Continuous Random Variables, the Pareto and Normal Distributions

LAB 4 INSTRUCTIONS CONFIDENCE INTERVALS AND HYPOTHESIS TESTING

Multiple Linear Regression

Hypothesis testing - Steps

UNDERSTANDING THE DEPENDENT-SAMPLES t TEST

Two Related Samples t Test

Comparing Means in Two Populations

Independent samples t-test. Dr. Tom Pierce Radford University

Online 12 - Sections 9.1 and 9.2-Doug Ensley

Testing Research and Statistical Hypotheses

A) B) C) D)

Analysing Questionnaires using Minitab (for SPSS queries contact -)

Descriptive Statistics

Regression Analysis: A Complete Example

One-Way Analysis of Variance (ANOVA) Example Problem

Elementary Statistics Sample Exam #3

HYPOTHESIS TESTING: CONFIDENCE INTERVALS, T-TESTS, ANOVAS, AND REGRESSION

Math 108 Exam 3 Solutions Spring 00

Week 4: Standard Error and Confidence Intervals

General Method: Difference of Means. 3. Calculate df: either Welch-Satterthwaite formula or simpler df = min(n 1, n 2 ) 1.

Analysis of Variance ANOVA

Final Exam Practice Problem Answers

Lecture Notes Module 1

Unit 26 Estimation with Confidence Intervals

3.4 Statistical inference for 2 populations based on two samples

How To Run Statistical Tests in Excel

How Far is too Far? Statistical Outlier Detection

The correlation coefficient

Stat 411/511 THE RANDOMIZATION TEST. Charlotte Wickham. stat511.cwick.co.nz. Oct

Nonparametric Two-Sample Tests. Nonparametric Tests. Sign Test

NCSS Statistical Software

Chapter 7 Section 1 Homework Set A

Exercise 1.12 (Pg )

Data Analysis Tools. Tools for Summarizing Data

Using Excel for inferential statistics

HYPOTHESIS TESTING WITH SPSS:

Case Study Call Centre Hypothesis Testing

DATA INTERPRETATION AND STATISTICS

Method To Solve Linear, Polynomial, or Absolute Value Inequalities:

Class 19: Two Way Tables, Conditional Distributions, Chi-Square (Text: Sections 2.5; 9.1)

INTERPRETING THE ONE-WAY ANALYSIS OF VARIANCE (ANOVA)

Two-sample t-tests. - Independent samples - Pooled standard devation - The equal variance assumption

One-Way Analysis of Variance

Good luck! BUSINESS STATISTICS FINAL EXAM INSTRUCTIONS. Name:

Confidence Intervals for the Difference Between Two Means

Chapter 7 Notes - Inference for Single Samples. You know already for a large sample, you can invoke the CLT so:

Part 3. Comparing Groups. Chapter 7 Comparing Paired Groups 189. Chapter 8 Comparing Two Independent Groups 217

Chapter 5 Analysis of variance SPSS Analysis of variance

Introduction. Statistics Toolbox

One-Way ANOVA using SPSS SPSS ANOVA procedures found in the Compare Means analyses. Specifically, we demonstrate

Recall this chart that showed how most of our course would be organized:

13: Additional ANOVA Topics. Post hoc Comparisons

KSTAT MINI-MANUAL. Decision Sciences 434 Kellogg Graduate School of Management

MATH 140 Lab 4: Probability and the Standard Normal Distribution

StatCrunch and Nonparametric Statistics

Comparing Two Groups. Standard Error of ȳ 1 ȳ 2. Setting. Two Independent Samples

1. The parameters to be estimated in the simple linear regression model Y=α+βx+ε ε~n(0,σ) are: a) α, β, σ b) α, β, ε c) a, b, s d) ε, 0, σ

Regression step-by-step using Microsoft Excel

How Does My TI-84 Do That

Factors affecting online sales

CHAPTER 11 CHI-SQUARE AND F DISTRIBUTIONS

= $96 = $24. (b) The degrees of freedom are. s n For the mean monthly rent, the 95% confidence interval for µ is

The Dummy s Guide to Data Analysis Using SPSS

individualdifferences

Variables Control Charts

How To Check For Differences In The One Way Anova

Chapter 9. Two-Sample Tests. Effect Sizes and Power Paired t Test Calculation

Chapter 2 Probability Topics SPSS T tests

UNDERSTANDING THE INDEPENDENT-SAMPLES t TEST

" Y. Notation and Equations for Regression Lecture 11/4. Notation:

Chapter 7 Section 7.1: Inference for the Mean of a Population

Having a coin come up heads or tails is a variable on a nominal scale. Heads is a different category from tails.

Difference of Means and ANOVA Problems

SCHOOL OF HEALTH AND HUMAN SCIENCES DON T FORGET TO RECODE YOUR MISSING VALUES

Playing with Numbers

MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS

Friedman's Two-way Analysis of Variance by Ranks -- Analysis of k-within-group Data with a Quantitative Response Variable

Transcription:

Variations of the t-test: Sample tail Sample t-test (unequal sample sizes and unequal variances) Like the last example, below we have ceramic sherd thickness measurements (in cm) of two samples representing different decorative styles from an archaeological site. However, this time we see that the sample sizes are different, but we are still interested in seeing whether the average thickness is statistically significant between the two samples or not. Let Y the sample mean of sherd thickness from sample, and Y the sample mean of sherd thickness from sample. We wish to test the hypothesis at the a 0.05 level (95%) that there is no statistical difference between the mean values of sample and. Formally, we state: H H o a : Y : Y Y Y 0 0 If the data are normally distributed (or close enough) we choose to test this hypothesis using a -tailed, sample t-test, taking into account the inequality of variances and sample sizes. Below are the data: Sample 9.746.845 6.3348 5.4338 0.830 9.356 9.66.5908 5.0997 8.00 0.8439 8.865 3.86 7.0340 3.5834 8.636.447 7.8443 5.339 6.6790 3.787 8.495 7.984.87 3.098 Sample 40.0790 8.55 35.809 6.5560 3.333 39.6987 5.476 9.6046 4.808 3.5064 39.79.468 3.078 5.369 30.58 39.803 34.696 30.9565 9.9376 3.996 7.645 37.05 33.953 3.066 37.757 9.3769 40.7894 39.6987 7.9 7.3089 36.67 8.7846 6.5954 9.7374 33.948 30.648 6.8967 8.4069 30.648 33.855 So first of all we need to look at our data, so we run the descriptive stats option in MINITAB and choose to present the samples graphically using a couple of boxplots.

Variations of the t-test: Sample tail Descriptive Statistics Variable N Mean Median Tr Mean StDev SE Mean Sample 5 3.565 3.787 3.77 3.960 0.79 Sample 40 30.8 30.09 30.5 6.49.03 Variable Sample Min 3.0 Max 9.66 Q 0.837 Q3 6.857 Sample 3. 40.79 6.57 35.53 Looking at the descriptive stats output we see that the mean of sample is smaller than sample, but also that the standard deviation of sample is smaller than sample. So straight away we know we cannot assume equal variances as we did in the last example. We notice that the sample sizes are also different; we are also going to have to deal with this issue when calculating our degrees of freedom (v or df). However, we notice that the means are very similar to the medians in both samples, and the boxplots suggest that the data is close enough to normal to go ahead with the parametric test, the t-test. Boxplots of Sample and Sample (means are indicated by solid circles) 40 30 0 0 Sample Sample So, as we know by now, as we are dealing with samples we need to take into account the measures of dispersion of both samples, though in this case we know we cannot just take the average of the two (as we did in the last example) because the variations are very unequal. There is a standard method to deal with this contingency as, understandably, this situation arises much of the time in the real world. We use what is known as the Satterthwaite Approximation: s s SE S + () n n With this equation we see that we can take into account both unequal variances and unequal sample sizes at the same time, and as such, the Satterthwaite approximation

Variations of the t-test: Sample tail 3 gives a weighted average of the standard errors. When the errors are equal, the Satterthwaite approximation gives roughly the same answer as the pooled variance procedure. If we wish to calculate a p value and compare it to our a, the t-test statistic is now calculated in the same way as before: t STAT Y Y () se p However, we have to calculate our degrees of freedom to find our t CRIT, and this is a little more complex this time as the sample sizes are unequal. In this case, the equation used to estimate v is: v s s + n n ( s / n ) ( s / n ) n (3) + n Okay, this looks ugly and it is. We do not have to be concerned with the derivation of this equation, or even why exactly it works, we just have to plug in our numbers and chug through the equation when the time comes. We can check manual calculations with the MINITAB output as MINITAB uses this algorithm if we chose the right option when running the test (more later). So let s plug and chug equation 3. To get the variances we square our standard deviations from the MINITAB output and plug the numbers in: 5.686 4.64 + 5 40 v + 4 39 ( 5.686) ( 4.64) ( 0.6739 +.0540) ( 0.06388 + 0.0849).8657 0.044878 63 You see that we round the result of this equation to the nearest integer, which is 63. To find our t CRIT we then look up v 63, a 0.05 in the table and find our t CRIT.000 (approximately). So, to calculate our standard error using the Satterthwaite approximation we plug and chug equation :

Variations of the t-test: Sample tail 4 SE S 5.686 5 + 4.64 40 0.676 +.054.967 Let us first calculate our t STAT using equation : t STAT 3.565 30.8 5.8.967 We can see straight off that 5.8 standard errors is far way from the mean, in fact we calculated our t CRIT to be + or -.000 telling us already that we are going to end up rejecting our null hypothesis in favor of the alternative. Let s now calculate our confidence limits (basic equations not shown): L L L U (3.565 30.8).000*.967 9.308 (3.565 30.8) +.000*.967 4. We see that both bounds are negative numbers indicating that they do not encompass zero, therefore the hypothesis that there is no difference between the two samples is not supported by the data; we reject the null hypothesis in favor of the alternative, at the a 0.05 level. The fact that both bounds are negative is a result of sample s mean being much smaller than sample in addition to their variances.

Variations of the t-test: Sample tail 5 Now let s run the test in MINITAB to check our results and see how our manual math skills held up against MINITAB s algorithms. To perform this test we follow these procedures: Enter your two samples in two columns >STAT EQUAL) >BASIC STATS > SAMPLE t >Choose SAMPLES IN DIFFERENT COLUMNS >Choose the alternative hypothesis (in this case NOT >Leave the confidence level at 95% >DO NOT Choose ASSUME EQUAL VARIANCES; MINITAB will use the Satterthwaite approximation as a default >OK The output from MINITAB should look like: Two Sample T-Test and Confidence Interval Two sample T for Sample vs Sample N Mean StDev SE Mean Sample 5 3.56 3.96 0.79 Sample 40 30.8 6.49.0 95% CI for mu Sample - mu Sample : ( -9.3, -4.) T-Test mu Sample mu Sample (vs not ): T -5.8 P0.0000 DF 6 First, you will see that MINITAB does not explicitly give us the Satterthwaite approximation of the standard error, but we will be able to tell if we were correct if the rest of the numbers turn out well. Looking for the degrees of freedom (df) we see that MINITAB got 6, whereas we got 63. By the time we are dealing with 60-odd degrees of freedom the critical values do not change that much so the error is acceptable; in fact the error comes from rounding error in our manual calculations as we can only input so many decimal places, whereas MINITAB can use dozens. This brings up an important point; rounding errors get magnified when we start multiplying and squaring values so always use as many decimal places as you are given in such cases. The closeness of our degrees of freedom to the

Variations of the t-test: Sample tail 6 MINITAB output lets us know that MINITAB probably used only one or two more decimal places. The t STAT in the output (T) is 5.8, the exact value we got manually indicating that our calculation of the Satterthwaite approximation was good, and as we expected, the p value is highly significant, therefore as p < a we reject the null hypothesis in favor of the alternative. Remember that as we are dealing with a Tailed t-test, the actual a value we are comparing our p value is a/ 0.05, as there are equal areas of rejection on both sides of our t distribution. Even so, we are still left with the same result. Finally, looking at the confidence limits in the output, we see that our manual calculations are exactly the same as the output. Our archaeological conclusion would be that the mean thickness of the sherds from sample are much less than the mean thickness of sherds from sample, and the statistics indicate that they (most probably) do not belong to the same statistical population of vessels, under the assumption that vessel thickness is an accurate proxy measure of functional difference. Not only are they stylistically different, they are also most probably functionally different.