Estimation of σ 2, the variance of ɛ



Similar documents
Coefficient of Determination

Hypothesis testing - Steps

Interaction between quantitative predictors

3.4 Statistical inference for 2 populations based on two samples

Two Related Samples t Test

1. What is the critical value for this 95% confidence interval? CV = z.025 = invnorm(0.025) = 1.96

HYPOTHESIS TESTING: POWER OF THE TEST

LAB 4 INSTRUCTIONS CONFIDENCE INTERVALS AND HYPOTHESIS TESTING

How To Test For Significance On A Data Set

Outline. Topic 4 - Analysis of Variance Approach to Regression. Partitioning Sums of Squares. Total Sum of Squares. Partitioning sums of squares

HYPOTHESIS TESTING: CONFIDENCE INTERVALS, T-TESTS, ANOVAS, AND REGRESSION

Final Exam Practice Problem Answers

Simple Regression Theory II 2010 Samuel L. Baker

HYPOTHESIS TESTING (ONE SAMPLE) - CHAPTER 7 1. used confidence intervals to answer questions such as...

Hypothesis Testing. Hypothesis Testing

1 Simple Linear Regression I Least Squares Estimation

Correlational Research

Regression Analysis: A Complete Example

Notes on Applied Linear Regression

Chapter 7 Notes - Inference for Single Samples. You know already for a large sample, you can invoke the CLT so:

Difference of Means and ANOVA Problems

Chapter 2. Hypothesis testing in one population

Simple Linear Regression Inference

Introduction to Hypothesis Testing. Hypothesis Testing. Step 1: State the Hypotheses

Section 7.1. Introduction to Hypothesis Testing. Schrodinger s cat quantum mechanics thought experiment (1935)

Week TSX Index

HYPOTHESIS TESTING (ONE SAMPLE) - CHAPTER 7 1. used confidence intervals to answer questions such as...

Hypothesis Testing --- One Mean

1. The parameters to be estimated in the simple linear regression model Y=α+βx+ε ε~n(0,σ) are: a) α, β, σ b) α, β, ε c) a, b, s d) ε, 0, σ

Multiple Linear Regression

Part 2: Analysis of Relationship Between Two Variables

Chapter 13 Introduction to Nonlinear Regression( 非 線 性 迴 歸 )

Chapter 8 Hypothesis Testing Chapter 8 Hypothesis Testing 8-1 Overview 8-2 Basics of Hypothesis Testing

Pearson's Correlation Tests

Section 1: Simple Linear Regression

An Introduction to Statistics Course (ECOE 1302) Spring Semester 2011 Chapter 10- TWO-SAMPLE TESTS

Additional sources Compilation of sources:

Univariate Regression

STAT 350 Practice Final Exam Solution (Spring 2015)

MULTIPLE REGRESSION EXAMPLE

2. What is the general linear model to be used to model linear trend? (Write out the model) = or

Econometrics Simple Linear Regression

Regression step-by-step using Microsoft Excel

The Normal distribution

Least Squares Estimation

DDBA 8438: Introduction to Hypothesis Testing Video Podcast Transcript

BA 275 Review Problems - Week 6 (10/30/06-11/3/06) CD Lessons: 53, 54, 55, 56 Textbook: pp , ,

SIMPLE LINEAR CORRELATION. r can range from -1 to 1, and is independent of units of measurement. Correlation can be done on two dependent variables.

" Y. Notation and Equations for Regression Lecture 11/4. Notation:

5. Linear Regression

Factors affecting online sales

2013 MBA Jump Start Program. Statistics Module Part 3

17. SIMPLE LINEAR REGRESSION II

Week 5: Multiple Linear Regression

Using Stata for One Sample Tests

BA 275 Review Problems - Week 5 (10/23/06-10/27/06) CD Lessons: 48, 49, 50, 51, 52 Textbook: pp

Bowerman, O'Connell, Aitken Schermer, & Adcock, Business Statistics in Practice, Canadian edition

12: Analysis of Variance. Introduction

THE FIRST SET OF EXAMPLES USE SUMMARY DATA... EXAMPLE 7.2, PAGE 227 DESCRIBES A PROBLEM AND A HYPOTHESIS TEST IS PERFORMED IN EXAMPLE 7.

5. Multiple regression

2. Simple Linear Regression

MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS

22. HYPOTHESIS TESTING

Hypothesis Testing. Reminder of Inferential Statistics. Hypothesis Testing: Introduction

Principles of Hypothesis Testing for Public Health

General Regression Formulae ) (N-2) (1 - r 2 YX

Lesson 9 Hypothesis Testing

Point Biserial Correlation Tests

2. Linear regression with multiple regressors

Data Analysis Tools. Tools for Summarizing Data

Chapter 4 Statistical Inference in Quality Control and Improvement. Statistical Quality Control (D. C. Montgomery)

Math 251, Review Questions for Test 3 Rough Answers

Nominal and Real U.S. GDP

Section Format Day Begin End Building Rm# Instructor. 001 Lecture Tue 6:45 PM 8:40 PM Silver 401 Ballerini

How Far is too Far? Statistical Outlier Detection

Chapter 13 Introduction to Linear Regression and Correlation Analysis

Two-Sample T-Tests Assuming Equal Variance (Enter Means)

Statistics Review PSY379

Unit 31 A Hypothesis Test about Correlation and Slope in a Simple Linear Regression

Nonlinear Regression Functions. SW Ch 8 1/54/

Lesson 1: Comparison of Population Means Part c: Comparison of Two- Means

Two-Sample T-Tests Allowing Unequal Variance (Enter Difference)

Machine Learning and Data Mining. Regression Problem. (adapted from) Prof. Alexander Ihler

Good luck! BUSINESS STATISTICS FINAL EXAM INSTRUCTIONS. Name:

Confidence intervals

Illustration (and the use of HLM)

An Introduction to Statistical Tests for the SAS Programmer Sara Beck, Fred Hutchinson Cancer Research Center, Seattle, WA

SUMAN DUVVURU STAT 567 PROJECT REPORT

Descriptive Statistics

Introduction to Hypothesis Testing OPRE 6301

Basic Statistics and Data Analysis for Health Researchers from Foreign Countries

Analysis of Variance ANOVA

Multinomial and Ordinal Logistic Regression

NCSS Statistical Software. One-Sample T-Test

Module 5: Multiple Regression Analysis

CHAPTER 13. Experimental Design and Analysis of Variance

5/31/2013. Chapter 8 Hypothesis Testing. Hypothesis Testing. Hypothesis Testing. Outline. Objectives. Objectives

Chapter 3 Quantitative Demand Analysis

Independent t- Test (Comparing Two Means)

Transcription:

Estimation of σ 2, the variance of ɛ The variance of the errors σ 2 indicates how much observations deviate from the fitted surface. If σ 2 is small, parameters β 0, β 1,..., β k will be reliably estimated and predictions ŷ will also be reliable. Since σ 2 is typically unknown, we estimate it from the sample as: ˆσ 2 = S 2 = SSE n number of parameters in model = SSE n (k + 1). Stat 328 - Fall 2004 1

Estimation of σ 2 (cont d) As in the case of simple linear regression: SSE = i (y i ŷ i ) 2 Also as before, the predicted value of y for a given set of x s is: ŷ i = b 0 + b 1 x 1i + b 2 x 2i +... + b k x ki. Stat 328 - Fall 2004 2

Estimation of σ 2 (cont d) As in simple regression, SAS and JMP call S 2 the MSE and S the RMSE. See example on page 169 of text. We have: Three predictors of home sale price, so k = 3. Sample size n = 20. From SAS output, MSE = 62, 718, 204 and RMSE = 7, 919. If we had wished to compute MSE by hand, we would have done so as: MSE = SSE n (k + 1) = 1003491259 20 (3 + 1). Stat 328 - Fall 2004 3

Inferences about β parameters A (1 α)% confidence interval for β j is given by: b j ± tα 2,n (k+1) standard error of b j We use ˆσ bj or S bj to denote the standard error of b j, and obtain its value from SAS or JMP output. The standard errors of the regression estimates are given in the column labeled Standard Error, both in SAS and in JMP. We can obtain confidence intervals for any of the regression coefficients in the model (and also for the intercept). Stat 328 - Fall 2004 4

Inferences about β parameters Example: see example on page 169. We wish to obtain a 90% CI for β 2 : or (0.45, 1.19). b 2 ± t 0.05,16ˆσ b2 0.82 ± 1.746(0.21), As before, we say that the interval has a 90% probability of covering the true value of β 2. Stat 328 - Fall 2004 5

Inferences about β parameters (cont d) We can also test hypotheses about the β s following the usual steps: 1. Set up the hypotheses to be tested, either one or two-tailed. 2. Choose level α, determine critical value and set up rejection region. 3. Compute test statistic. 4. Compare test statistic to critical value and reach conclusion. 5. Interpret results. Hypotheses: The null is always H 0 : β j = 0. Alternative for a two-tailed test: H a : β j 0. Alternative for a one-tailed test: H a : β j < 0 or H a : β j > 0. Stat 328 - Fall 2004 6

Inferences about β parameters (cont d) Critical value: For a two-tailed test, the critical values are ±t α/2,n k 1. For a one-tailed test, the critical value is t α,n k 1 (if H a : β j < 0) or +t α,n k 1 (if H a : β j > 0). Critical or rejection region: For a two-tailed test: Reject H 0 if test statistic t < t α/2,n k 1 or t > t α/2,n k 1. For a one-tailed test: Reject H 0 if t < t α,n k 1 [or t > +t α,n k 1 for a bigger-than alternative.] Test statistic: As in simple linear regression: t = b j ˆσ bj Stat 328 - Fall 2004 7

Inferences about β parameters (cont d) How do we interpret results of hypotheses tests? Suppose we reject H 0 : β j = 0 while conducting a two-tailed test. Conclusion: Data suggest that the response y and the jth predictor x j are linearly associated when other predictors are held constant. If we fail to reject H 0, then reasons might be: 1. There is no association between y and x j. 2. A linear association exists (when other x s are held constant) but a Type II error occurred (defined as concluding H 0 when H a is true). 3. The association between y and x j exists, but it is more complex than linear. Stat 328 - Fall 2004 8

Multiple coefficient of determination R 2 The multiple coefficient of determination R 2 is a measure of how well the linear model fits the data. As in simple linear regression, R 2 is defined as: R 2 = SS yy SSE SS yy = 1 SSE SS yy, and 0 R 2 1. The closer R 2 is to one, the better the model fits the data. If R 2 is equal to 0.65 (for example) we say that about 65% of the sample variability observed in the response can be attributed to (or explained by) the predictors in the model. Stat 328 - Fall 2004 9

The adjusted R 2 As it happens, we can artifially increase R 2 simply by adding predictors to the model. For example, if we have n = 2 observations, a simple linear regression of y on x will result in R 2 = 1 even if x and y are not associated. To get R 2 to be equal to 1 all we need to do is fit a model with n parameters to a dataset of size n. Then, R 2 makes sense as a measure of goodness of fit only if n is a lot larger than k. We can penalize the R 2 every time we add a new predictor to the model. The penalized R 2 is called the adjusted R 2 and it is sometimes more useful than the plain R 2. Stat 328 - Fall 2004 10

The adjusted R 2 (cont d) The adjusted R 2 is denoted R 2 a and is computed as: R 2 a = 1 = 1 [ ] ( ) (n 1) SSE n (k + 1) SS yy [ ] (n 1) (1 R 2 ). n (k + 1) Stat 328 - Fall 2004 11

The adjusted R 2 (cont d) Note that As k increases, n (k + 1) increases and SSE decreases. If new predictor contributes information about y, SSE will decrease faster than the increase in n (k + 1), so Ra 2 will increase. Else, Ra 2 will decrease when we add a new predictor to the model. [The ordinary R 2 always increases with added predictors even if new predictors contribute no information about y.] R 2 a adjusts for sample size and number of predictors k and cannot be forced to be equal to 1. R 2 a < R 2 always. Stat 328 - Fall 2004 12