Error Type, Power, Assumptions. Parametric Tests. Parametric vs. Nonparametric Tests



Similar documents
Assumptions. Assumptions of linear models. Boxplot. Data exploration. Apply to response variable. Apply to error terms from linear model

Permutation Tests for Comparing Two Populations

Data Transforms: Natural Logarithms and Square Roots

How Far is too Far? Statistical Outlier Detection

HYPOTHESIS TESTING WITH SPSS:

Analysis of Data. Organizing Data Files in SPSS. Descriptive Statistics

How To Test For Significance On A Data Set

Descriptive Statistics

Chapter 8 Hypothesis Testing Chapter 8 Hypothesis Testing 8-1 Overview 8-2 Basics of Hypothesis Testing

Nonparametric Two-Sample Tests. Nonparametric Tests. Sign Test

HYPOTHESIS TESTING: POWER OF THE TEST

II. DISTRIBUTIONS distribution normal distribution. standard scores

INTERPRETING THE ONE-WAY ANALYSIS OF VARIANCE (ANOVA)

Point Biserial Correlation Tests

Univariate Regression

Study Guide for the Final Exam

CHAPTER 14 NONPARAMETRIC TESTS

Non-Parametric Tests (I)

13: Additional ANOVA Topics. Post hoc Comparisons

THE KRUSKAL WALLLIS TEST

Introduction to Hypothesis Testing. Hypothesis Testing. Step 1: State the Hypotheses

Normality Testing in Excel

Principles of Hypothesis Testing for Public Health

Lecture Notes Module 1

Statistics in Medicine Research Lecture Series CSMC Fall 2014

Projects Involving Statistics (& SPSS)

X X X a) perfect linear correlation b) no correlation c) positive correlation (r = 1) (r = 0) (0 < r < 1)

Statistics courses often teach the two-sample t-test, linear regression, and analysis of variance

Tutorial 5: Hypothesis Testing

t-test Statistics Overview of Statistical Tests Assumptions

business statistics using Excel OXFORD UNIVERSITY PRESS Glyn Davis & Branko Pecar

Simple linear regression

Section 7.1. Introduction to Hypothesis Testing. Schrodinger s cat quantum mechanics thought experiment (1935)

Elements of statistics (MATH0487-1)

Pearson's Correlation Tests

NONPARAMETRIC STATISTICS 1. depend on assumptions about the underlying distribution of the data (or on the Central Limit Theorem)

Exact Nonparametric Tests for Comparing Means - A Personal Summary

Final Exam Practice Problem Answers

Simple Regression Theory II 2010 Samuel L. Baker

Difference tests (2): nonparametric

NCSS Statistical Software. One-Sample T-Test

Variables Control Charts

Chicago Booth BUSINESS STATISTICS Final Exam Fall 2011

This chapter discusses some of the basic concepts in inferential statistics.

Week 3&4: Z tables and the Sampling Distribution of X

HYPOTHESIS TESTING: CONFIDENCE INTERVALS, T-TESTS, ANOVAS, AND REGRESSION

StatCrunch and Nonparametric Statistics

Normal Distribution. Definition A continuous random variable has a normal distribution if its probability density. f ( y ) = 1.

Experimental Design. Power and Sample Size Determination. Proportions. Proportions. Confidence Interval for p. The Binomial Test

Descriptive Statistics

Chapter 7 Section 7.1: Inference for the Mean of a Population

Improving your data transformations: Applying the Box-Cox transformation

Recall this chart that showed how most of our course would be organized:

Hypothesis Testing --- One Mean

Skewed Data and Non-parametric Methods

STAT 350 Practice Final Exam Solution (Spring 2015)

Research Methods & Experimental Design

THE FIRST SET OF EXAMPLES USE SUMMARY DATA... EXAMPLE 7.2, PAGE 227 DESCRIBES A PROBLEM AND A HYPOTHESIS TEST IS PERFORMED IN EXAMPLE 7.

Chapter 7 Notes - Inference for Single Samples. You know already for a large sample, you can invoke the CLT so:

Math 461 Fall 2006 Test 2 Solutions

Lesson 1: Comparison of Population Means Part c: Comparison of Two- Means

1 Nonparametric Statistics

1-3 id id no. of respondents respon 1 responsible for maintenance? 1 = no, 2 = yes, 9 = blank

Dongfeng Li. Autumn 2010

QUANTITATIVE METHODS BIOLOGY FINAL HONOUR SCHOOL NON-PARAMETRIC TESTS

Module 5: Multiple Regression Analysis

Outline. Definitions Descriptive vs. Inferential Statistics The t-test - One-sample t-test

Simple Linear Regression

MODIFIED PARAMETRIC BOOTSTRAP: A ROBUST ALTERNATIVE TO CLASSICAL TEST

Likelihood: Frequentist vs Bayesian Reasoning

Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)

Stat 5102 Notes: Nonparametric Tests and. confidence interval

Simple Linear Regression Inference

Introduction to nonparametric regression: Least squares vs. Nearest neighbors

Part II Chapter 9 Chapter 10 Chapter 11 Chapter 12 Chapter 13 Chapter 14 Chapter 15 Part II

5/31/2013. Chapter 8 Hypothesis Testing. Hypothesis Testing. Hypothesis Testing. Outline. Objectives. Objectives

Hypothesis testing - Steps

Paired 2 Sample t-test

T-test & factor analysis

How To Check For Differences In The One Way Anova

1. What is the critical value for this 95% confidence interval? CV = z.025 = invnorm(0.025) = 1.96

KSTAT MINI-MANUAL. Decision Sciences 434 Kellogg Graduate School of Management

Nonparametric Statistics

Stat 411/511 THE RANDOMIZATION TEST. Charlotte Wickham. stat511.cwick.co.nz. Oct

Non-Inferiority Tests for Two Means using Differences

Descriptive Statistics and Measurement Scales

2 Sample t-test (unequal sample sizes and unequal variances)

Regression III: Advanced Methods

AP STATISTICS REVIEW (YMS Chapters 1-8)

Understand the role that hypothesis testing plays in an improvement project. Know how to perform a two sample hypothesis test.

DEALING WITH THE DATA An important assumption underlying statistical quality control is that their interpretation is based on normal distribution of t

Biostatistics: Types of Data Analysis

17. SIMPLE LINEAR REGRESSION II

Statistical Process Control (SPC) Training Guide

Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.

Paired T-Test. Chapter 208. Introduction. Technical Details. Research Questions

12: Analysis of Variance. Introduction

MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question.

Statistics. Measurement. Scales of Measurement 7/18/2012

Transcription:

Error Type, Power, Assumptions Parametric vs. Nonparametric tests Type-I & -II Error Power Revisited Meeting the Normality Assumption - Outliers, Winsorizing, Trimming - Data Transformation 1 Parametric Tests Parametric tests assume that the variable in question has a known underlying mathematical distribution that can be described (normal, binomial, poisson, etc.). This underlying distribution is the fundamental basis for all of sample-to-population inference. 2 Parametric vs. Nonparametric Tests Nonparametric tests are considered distribution-free methods because they do not rely on any underlying mathematical distribution. Q. So why even worry about what the distribution is or is not? Why not just use nonparametric tests all the time? A. Nonparametric tests usually result in loss of efficiency (the ability to detect a false hypothesis). Efficiency is tied to error type. 3

Error Type - Truth Table - Ho Accepted Rejected True Correct Type - I False Type - II Correct 4 Type-I Error Before you apply a statistical test, you must specify an acceptable level of Type-I error. Usually, one accepts that there will always be some deviant observations by chance alone and that 5% error is acceptable. Recall, Type-I error is expressed as a probability and is symbolized by α. Thus, a Type-I error of α = 0.05 corresponds to a 5% error level and specifies the rejection region or critical region of a statistical test. 5 Error Type Q. So why not specify a very small error rate such as 0.01 or 0.001? A. Because as your Type-I error rate diminishes, Type-II error increases! Unfortunately, while Type-II error is important, it is difficult to evaluate in many biological applications. 6

Type-II Error Type-II error is the probability of accepting a false H o. Type-II error is also referred to as a probability and symbolized as β. Β is harder to specify because it requires knowledge of the alternate hypothesis (which is unknown in most circumstances). β is not fixed, but may increase to a maximum of 1- α. 7 Power Important Concept: Power = 1 - β Power and β are complements. Thus, for any given test, we would like power to be as high as possible and β to be as low as possible. 8 Power Since we can not generally provide an alternative hypothesis, we must describe β or 1 - β as a continuum of alternative values. This is known as a Power Curve. To improve the power of a test (i.e., decrease β) while keeping α fixed, we vary N. 9

Power Curve 1 Power (1 - β ) 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 α = 0.05 N = 5 N = 35 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 μ Power Curves for testing: Ho: μ = 45 Ha: μ 45 at α = 0.05 and N = 5, 35 10 Nonparametric Tests Q. Well then, doesn t this mean nonparametric tests are undesirable or inferior? A. No! They just have less efficiency. They are the appropriate test to use when the conditions warrant. 11 Parametric vs. Nonparametric Tests In general, Parametric tests are more conservative (i.e., less likely to make a Type-I Error). Nonparametric tests are more liberal (i.e., more likely to make a Type-I Error). Thus, in most biological applications, one should always attempt to use a parametric test first. 12

Meeting the Normality Assumption Q. What if you are unable to meet the assumption of normality? You can not continue to do parametric statistics if this has not been met, correct? A. The best strategy is to first try a simple manipulation or re-arrangement of the data. This may allow you to meet the normality assumption and continue with parametric statistics. 13 Data Manipulations Options for Data Manipulation: Delete outliers Winsorize data Trim data These procedures are legal as long as: (1) they are exercised judiciously (2) never used to adjust a P-value 14 Outliers Handling outliers is tricky business. Do these values represent natural biological variability, or are they fluke values, or are they a mistake in data collection or recording? During EDA, use box-plots to help identify outliers. Carefully examine outliers. Mild outliers are usually biologically possible. Severe outliers are often mistakes. 20.0 13.3 6.7 0.0 15

Outliers Data can be normalized if there are mild outliers usually by winsorization, trimming, or transformation. Generally, severe outliers must be deleted from the data to achieve normality. CAUTION: Do not ever delete more than 5% of your data. Severe outliers can legitimately fall within this range. However, if there are more than 5% severe outliers, usually something else is going on. 16 Winsorizing Data Usually, but not necessarily, performed in a symmetrical fashion. Rank data, then give extremes the same value as adjacent rank. Recompute stats & test of normality. Example: 1, 2, 3, 4, 5, 7, 18 Mean = 5.7 M-I Normality: reject 2, 2, 3, 4, 5, 7, 7 Mean = 4.3 M-I Normality: accept 17 Trimming Data Alternatively, data can be trimmed from the tails. Usually, drop X min & X max This reduces N and may affect Power. Example: 1, 2, 3, 4, 5, 7, 18 Mean = 5.7 M-I Normality: reject 2, 3, 4, 5, 7 Mean = 4.2 M-I Normality: accept 18

Winsorization vs. Trimming Note that, in our example, there was very little difference in the effect of trimming vs. winsorization. There are no hard and fast rules as to when to apply one and not the other. Winsorization is probably more appropriate when sample sizes are small and you need to protect your power. 19 Data Transformations The necessity to transform data may arise under the conditions of non-independence or non-normality. Data transformation seems like a lot of manipulation at first glance, but it just involves placing your data on another scale. Data from a linear scale can be transformed on to a log 10 scale (or any other). This often corrects a variety of problems. Different transformations are available to correct different problems. 20 Effect of Transformation (ln) 21

Data Transformations Typical Transformations: Logarithmic Square Root Angular Box-Cox Reciprocal Power 22 Logarithmic Transformations Logarithmic transformations are useful starting points when: (1) mean is correlated with variance (2) data are skewed to the right Can take a variety of forms: Y = log 10 (Y) Y = log 10 (Y+1) Y = ln (Y) Y = ln (Y+1) 23 Transformation Example Y 1 2 2 3 4 5 6 6 12 18 K 2 Omnibus: Reject Normality C1 20.0 15.0 10.0 5.0 Normal Probability Plot of C1 0.0-2.0-1.0 0.0 1.0 2.0 Expected Normals Y = log 10 (Y) 0 0.301 0.301 K 2 Omnibus: Accept Normality 0.477 0.602 Normal Probability Plot of C2 0.699 1.4 0.778 1.0 0.778 0.6 1.079 0.2 1.255-0.2-2.0-1.0 0.0 1.0 2.0 C2 Expected Normals 24

> Y<-c(1,2,2,3,4,5,6,6,12,18) > shapiro.test(y) Shapiro-Wilk normality test data: Y W = 0.816, p-value = 0.02268 > Ylog<-log(Y) > Ylog [1] 0.0000000 0.6931472 0.6931472 1.0986123 1.3862944 [6] 1.6094379 1.7917595 1.7917595 2.4849066 2.8903718 > shapiro.test(ylog) Shapiro-Wilk normality test data: Ylog W = 0.9797, p-value = 0.9633 25 Square Root Transformations Most appropriate when data are counts (e.g., number of leaves, number of flowers, etc.). Count data tend to more closely follow a Poisson distribution. A square root transformation brings closer to normal. Variety of forms: Y'= Y 0.05 Y'= Y Y 1 Y'= Y 3 8 Y ' = 3 Y 26 Angular Transformations Whenever the data are proportions or percentages, you should consider an angular transformation. Percentages tend to usually follow a binomial distribution. Typical transforms: =arcsin p where p ranges 0 1 =arcsin Y 3 8 N 3 4 27

Box-Cox Transformation When there is no a priori reason for choosing one transformation over another, the Box-Cox transformation might be an appropriate place to start. Iterate through a series of power functions until normality is maximized: 28 Reciprocal Transformation Reciprocal transforms often prove useful when the standard deviations of the groups of data are proportional to the square of the means of the groups. Iterate through a series of until normality is maximized: 29 Power Transformation A power transformation is often effective in dealing with two situations: (1) if S decreases with increasing y (2) if the distribution is skewed to the left Iterate through a series of λ until normality is maximized: Y' = Y λ 30

Example 13.1 (p. 323) > BR [1] 1.34 1.96 2.49 1.27 1.19 1.15 1.29 1.05 1.10 1.21 1.31 [12] 1.26 1.38 1.49 1.84 1.84 3.06 2.65 4.25 3.35 2.55 1.72 [23] 1.52 1.49 1.67 1.78 1.71 1.88 0.83 1.16 1.31 1.40 > qqnorm(br,col="red") > qqline(br,col="red") > hist(br,col= red ) 31 > BRlog [1] 0.29266961 0.67294447 0.91228271 0.23901690 [5] 0.17395331 0.13976194 0.25464222 0.04879016 [9] 0.09531018 0.19062036 0.27002714 0.23111172 [13] 0.32208350 0.39877612 0.60976557 0.60976557 [17] 1.11841492 0.97455964 1.44691898 1.20896035 [21] 0.93609336 0.54232429 0.41871033 0.39877612 [25] 0.51282363 0.57661336 0.53649337 0.63127178 [29] -0.18632958 0.14842001 0.27002714 0.33647224 > qqnorm(brlog,col="red") > qqline(brlog,col="red") > hist(brlog,col="red") 32 > shapiro.test(br) Shapiro-Wilk normality test data: BR W = 0.8175, p-value = 8.851e-05 > shapiro.test(brlog) Shapiro-Wilk normality test data: BRlog W = 0.938, p-value = 0.06551 33