The F distribution

Similar documents
Regression Analysis: A Complete Example

CHAPTER 13 SIMPLE LINEAR REGRESSION. Opening Example. Simple Regression. Linear Regression

1. What is the critical value for this 95% confidence interval? CV = z.025 = invnorm(0.025) = 1.96

Chapter 7: Simple linear regression Learning Objectives

Multiple Linear Regression

Part 2: Analysis of Relationship Between Two Variables

Chapter 13 Introduction to Linear Regression and Correlation Analysis

Factors affecting online sales

We extended the additive model in two variables to the interaction model by adding a third term to the equation.

DEPARTMENT OF PSYCHOLOGY UNIVERSITY OF LANCASTER MSC IN PSYCHOLOGICAL RESEARCH METHODS ANALYSING AND INTERPRETING DATA 2 PART 1 WEEK 9

Hypothesis testing - Steps

2. Simple Linear Regression

Final Exam Practice Problem Answers

Unit 31 A Hypothesis Test about Correlation and Slope in a Simple Linear Regression

Using R for Linear Regression

Coefficient of Determination

Simple Linear Regression Inference

" Y. Notation and Equations for Regression Lecture 11/4. Notation:

Recall this chart that showed how most of our course would be organized:

Data Mining and Data Warehousing. Henryk Maciejewski. Data Mining Predictive modelling: regression

NCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( )

1.5 Oneway Analysis of Variance

Statistical Models in R

Statistical Models in R

Introduction to Regression and Data Analysis

X X X a) perfect linear correlation b) no correlation c) positive correlation (r = 1) (r = 0) (0 < r < 1)

Simple linear regression

Chapter 2 Probability Topics SPSS T tests

Chapter 5 Analysis of variance SPSS Analysis of variance

Module 5: Multiple Regression Analysis

2013 MBA Jump Start Program. Statistics Module Part 3

Simple Regression Theory II 2010 Samuel L. Baker

IAPRI Quantitative Analysis Capacity Building Series. Multiple regression analysis & interpreting results

Correlation and Simple Linear Regression

POLYNOMIAL AND MULTIPLE REGRESSION. Polynomial regression used to fit nonlinear (e.g. curvilinear) data into a least squares linear regression model.

SPSS Guide: Regression Analysis

August 2012 EXAMINATIONS Solution Part I

17. SIMPLE LINEAR REGRESSION II

International Statistical Institute, 56th Session, 2007: Phil Everson

Premaster Statistics Tutorial 4 Full solutions

Univariate Regression

Estimation of σ 2, the variance of ɛ

Testing for Lack of Fit

Example: Boats and Manatees

Please follow the directions once you locate the Stata software in your computer. Room 114 (Business Lab) has computers with Stata software

Regression III: Advanced Methods

SIMPLE LINEAR CORRELATION. r can range from -1 to 1, and is independent of units of measurement. Correlation can be done on two dependent variables.

Interaction between quantitative predictors

Predictor Coef StDev T P Constant X S = R-Sq = 0.0% R-Sq(adj) = 0.

12: Analysis of Variance. Introduction

Week TSX Index

Outline. Topic 4 - Analysis of Variance Approach to Regression. Partitioning Sums of Squares. Total Sum of Squares. Partitioning sums of squares

GLM I An Introduction to Generalized Linear Models

Elementary Statistics Sample Exam #3

Regression and Correlation

WEB APPENDIX. Calculating Beta Coefficients. b Beta Rise Run Y X

One-Way Analysis of Variance

The correlation coefficient

4. Simple regression. QBUS6840 Predictive Analytics.

MULTIPLE LINEAR REGRESSION ANALYSIS USING MICROSOFT EXCEL. by Michael L. Orlov Chemistry Department, Oregon State University (1996)

APPLICATION OF LINEAR REGRESSION MODEL FOR POISSON DISTRIBUTION IN FORECASTING

Chapter 4 and 5 solutions

MULTIPLE REGRESSION EXAMPLE

1. The parameters to be estimated in the simple linear regression model Y=α+βx+ε ε~n(0,σ) are: a) α, β, σ b) α, β, ε c) a, b, s d) ε, 0, σ

STAT 350 Practice Final Exam Solution (Spring 2015)

5. Linear Regression

Lets suppose we rolled a six-sided die 150 times and recorded the number of times each outcome (1-6) occured. The data is

Section 14 Simple Linear Regression: Introduction to Least Squares Regression

Answer: C. The strength of a correlation does not change if units change by a linear transformation such as: Fahrenheit = 32 + (5/9) * Centigrade

2. What is the general linear model to be used to model linear trend? (Write out the model) = or

Generalized Linear Models

Point Biserial Correlation Tests

SAS Software to Fit the Generalized Linear Model

Chicago Booth BUSINESS STATISTICS Final Exam Fall 2011

Chapter 23. Inferences for Regression

KSTAT MINI-MANUAL. Decision Sciences 434 Kellogg Graduate School of Management

Introduction to Analysis of Variance (ANOVA) Limitations of the t-test

A Primer on Forecasting Business Performance

Notes on Applied Linear Regression

Independent t- Test (Comparing Two Means)

Example G Cost of construction of nuclear power plants

One-Way Analysis of Variance: A Guide to Testing Differences Between Multiple Groups

Simple Methods and Procedures Used in Forecasting

N-Way Analysis of Variance

MULTIPLE REGRESSIONS ON SOME SELECTED MACROECONOMIC VARIABLES ON STOCK MARKET RETURNS FROM

Dimensionality Reduction: Principal Components Analysis

Econometrics Simple Linear Regression

EM375 STATISTICS AND MEASUREMENT UNCERTAINTY LEAST SQUARES LINEAR REGRESSION ANALYSIS

DATA INTERPRETATION AND STATISTICS

ECON 142 SKETCH OF SOLUTIONS FOR APPLIED EXERCISE #2

INTERPRETING THE ONE-WAY ANALYSIS OF VARIANCE (ANOVA)

Multiple Regression in SPSS This example shows you how to perform multiple regression. The basic command is regression : linear.

Chapter 2 Simple Comparative Experiments Solutions

The importance of graphing the data: Anscombe s regression examples

2. Linear regression with multiple regressors

Pearson s Correlation

1 Simple Linear Regression I Least Squares Estimation

Lesson 1: Comparison of Population Means Part c: Comparison of Two- Means

Business Statistics. Successful completion of Introductory and/or Intermediate Algebra courses is recommended before taking Business Statistics.

Regression step-by-step using Microsoft Excel

Transcription:

10-5.1 The F distribution

11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis is a statistical technique that is very useful for these types of problems. For example, in a chemical process, suppose that the yield of the product is related to the process-operating temperature. Regression analysis can be used to build a model to predict yield at a given temperature level.

11-1 Empirical Models

11-1 Empirical Models Based on the scatter diagram, it is probably reasonable to assume that the mean of the random variable Y is related to x by the following straight-line relationship: where the slope and intercept of the line are called regression coefficients. The simple linear regression model is given by where ε is the random error term.

11-1 Empirical Models We think of the regression model as an empirical model. Suppose that the mean and variance of ε are 0 and σ 2, respectively, then The variance of Y given x is

11-1 Empirical Models The true regression model is a line of mean values: where β 1 can be interpreted as the change in the mean of Y for a unit change in x. Also, the variability of Y at a particular value of x is determined by the error variance, σ 2. This implies there is a distribution of Y-values at each x and that the variance of this distribution is the same at each x.

11-1 Empirical Models Figure 11-2 The distribution of Y for a given value of x for the oxygen purity-hydrocarbon data.

11-2 Simple Linear Regression The case of simple linear regression considers a single regressor or predictor x and a dependent or response variable Y. The expected value of Y at each level of x is a random variable: We assume that each observation, Y, can be described by the model

11-2 Simple Linear Regression Suppose that we have n pairs of observations (x 1, y 1 ), (x 2, y 2 ),, (x n, y n ). Figure 11-3 Deviations of the data from the estimated regression model.

11-2 Simple Linear Regression The method of least squares is used to estimate the parameters, β 0 and β 1 by minimizing the sum of the squares of the vertical deviations in Figure 11-3. Figure 11-3 Deviations of the data from the estimated regression model.

11-2 Simple Linear Regression Using Equation 11-2, the n observations in the sample can be expressed as The sum of the squares of the deviations of the observations from the true regression line is

11-2 Simple Linear Regression

11-2 Simple Linear Regression

11-2 Simple Linear Regression

11-2 Simple Linear Regression

11-2 Simple Linear Regression Notation

11-2 Simple Linear Regression Example 11-1

11-2 Simple Linear Regression Example 11-1

11-2 Simple Linear Regression Example 11-1 Figure 11-4 Scatter plot of oxygen purity y versus hydrocarbon level x and regression model ŷ = 74.20 + 14.97x.

11-2 Simple Linear Regression Estimating σ 2 The error sum of squares is It can be shown that the expected value of the error sum of squares is E(SS E ) = (n 2)σ 2.

11-2 Simple Linear Regression Estimating σ 2 An unbiased estimator of σ 2 is where SS E can be easily computed using where SS T = n (y i y ) 2 = y 2 i ny 2 = S yy i=1 n i=1

11-3 Properties of the Least Squares Estimators Slope Properties Intercept Properties

11-4 Hypothesis Tests in Simple Linear Regression 11-4.1 Use of t-tests Suppose we wish to test An appropriate test statistic would be

11-4 Hypothesis Tests in Simple Linear Regression Assumptions: To test hypotheses about the slope and intercept of the regression model, we must make the additional assumption that the error component in the model, ε, is normally distributed. Thus, the complete assumptions are that the errors are normally and independently distributed with mean zero and variance σ 2, abbreviated NID(0, σ 2 ).

11-4 Hypothesis Tests in Simple Linear Regression 11-4.1 Use of t-tests The test statistic could also be written as: We would reject the null hypothesis if

11-4.1 Use of t-tests Suppose we wish to test An appropriate test statistic would be We would reject the null hypothesis if

11-4 Hypothesis Tests in Simple Linear Regression 11-4.1 Use of t-tests An important special case of the hypotheses of Equation 11-18 is These hypotheses relate to the significance of regression. Failure to reject H 0 is equivalent to concluding that there is no linear relationship between x and Y.

11-4 Hypothesis Tests in Simple Linear Regression Figure 11-5 The hypothesis H 0 : β 1 = 0 is not rejected.

11-4 Hypothesis Tests in Simple Linear Regression Figure 11-6 The hypothesis H 0 : β 1 = 0 is rejected.

11-4 Hypothesis Tests in Simple Linear Regression Example 11-2

R commands and outputs > dat=read.table("table11-1.txt", h=t)! > g=lm(y~x, dat)! > summary(g)! Coefficients:! Estimate Std. Error t value Pr(> t )! (Intercept) 74.283 1.593 46.62 < 2e-16 ***! x 14.947 1.317 11.35 1.23e-09 ***! Residual standard error: 1.087 on 18 degrees of freedom! Multiple R-Squared: 0.8774,!Adjusted R-squared: 0.8706! F-statistic: 128.9 on 1 and 18 DF, p-value: 1.227e-09! > anova(g)! Analysis of Variance Table! Response: y! Df Sum Sq Mean Sq F value Pr(>F)! x 1 152.127 152.127 128.86 1.227e-09 ***! Residuals 18 21.250 1.181!

10-5.1 The F Distribution

10-5.1 The F Distribution

10-5.1 The F Distribution The lower-tail percentage points f 1-α,u,ν can be found as follows.

11-4 Hypothesis Tests in Simple Linear Regression 11-4.2 Analysis of Variance Approach to Test Significance of Regression The analysis of variance identity is Symbolically,

11-4 Hypothesis Tests in Simple Linear Regression 11-4.2 Analysis of Variance Approach to Test Significance of Regression If the null hypothesis, H 0 : β 1 = 0 is true, the statistic follows the F 1,n-2 distribution and we would reject if f 0 > f α,1,n-2.

11-4 Hypothesis Tests in Simple Linear Regression 11-4.2 Analysis of Variance Approach to Test Significance of Regression The quantities, MS R and MS E are called mean squares. Analysis of variance table:

11-4 Hypothesis Tests in Simple Linear Regression Example 11-3

11-4 Hypothesis Tests in Simple Linear Regression

11-5 Confidence Intervals 11-5.1 Confidence Intervals on the Slope and Intercept Definition

11-6 Confidence Intervals Example 11-4

11-5 Confidence Intervals 11-5.2 Confidence Interval on the Mean Response Definition

11-5 Confidence Intervals Example 11-5

11-5 Confidence Intervals

11-5 Confidence Intervals Example 11-5 Figure 11-7 Scatter diagram of oxygen purity data from Example 11-1 with fitted regression line and 95 percent confidence limits on µ Y x0.

11-6 Prediction of New Observations If x 0 is the value of the regressor variable of interest, is the point estimator of the new or future value of the response, Y 0.

11-6 Prediction of New Observations Definition

11-6 Prediction of New Observations Example 11-6

11-6 Prediction of New Observations Example 11-6

11-6 Prediction of New Observations Example 11-6 Figure 11-8 Scatter diagram of oxygen purity data from Example 11-1 with fitted regression line, 95% prediction limits (outer lines), and 95% confidence limits on µ Y x0.

11-7 Adequacy of the Regression Model Fitting a regression model requires several assumptions. 1. Errors are uncorrelated random variables with mean zero; 2. Errors have constant variance; and, 3. Errors be normally distributed. The analyst should always consider the validity of these assumptions to be doubtful and conduct analyses to examine the adequacy of the model

11-7 Adequacy of the Regression Model 11-7.1 Residual Analysis The residuals from a regression model are e i = y i - ŷ i, where y i is an actual observation and ŷ i is the corresponding fitted value from the regression model. Analysis of the residuals is frequently helpful in checking the assumption that the errors are approximately normally distributed with constant variance, and in determining whether additional terms in the model would be useful.

11-7 Adequacy of the Regression Model 11-7.1 Residual Analysis Figure 11-9 Patterns for residual plots. (a) satisfactory, (b) funnel, (c) double bow, (d) nonlinear. [Adapted from Montgomery, Peck, and Vining (2001).]

11-7 Adequacy of the Regression Model Example 11-7

11-7 Adequacy of the Regression Model Example 11-7 Figure 11-10 Normal probability plot of residuals, Example 11-7.

11-7 Adequacy of the Regression Model Example 11-7 Figure 11-11 Plot of residuals versus predicted oxygen purity, ŷ, Example 11-7.

11-7 Adequacy of the Regression Model 11-7.2 Coefficient of Determination (R 2 ) The quantity is called the coefficient of determination and is often used to judge the adequacy of a regression model. 0 R 2 1; We often refer (loosely) to R 2 as the amount of variability in the data explained or accounted for by the regression model.

11-7 Adequacy of the Regression Model 11-7.2 Coefficient of Determination (R 2 ) For the oxygen purity regression model, R 2 = SS R /SS T = 152.13/173.38 = 0.877 Thus, the model accounts for 87.7% of the variability in the data.

11-9 Transformation and Logistic Regression

11-9 Transformation and Logistic Regression Example 11-9 Table 11-5 Observed Values and Regressor Variable x i for Example 11-9. y i

11-9 Transformation and Logistic Regression

11-9 Transformation and Logistic Regression

11-9 Transformation and Logistic Regression