Simple Linear Regression Inference



Similar documents
Univariate Regression

Chapter 7: Simple linear regression Learning Objectives

Multiple Linear Regression

Elements of statistics (MATH0487-1)

Part 2: Analysis of Relationship Between Two Variables

1. What is the critical value for this 95% confidence interval? CV = z.025 = invnorm(0.025) = 1.96

" Y. Notation and Equations for Regression Lecture 11/4. Notation:

Chicago Booth BUSINESS STATISTICS Final Exam Fall 2011

Regression Analysis: A Complete Example

17. SIMPLE LINEAR REGRESSION II

Outline. Topic 4 - Analysis of Variance Approach to Regression. Partitioning Sums of Squares. Total Sum of Squares. Partitioning sums of squares

Factors affecting online sales

Unit 31 A Hypothesis Test about Correlation and Slope in a Simple Linear Regression

Chapter 7 Notes - Inference for Single Samples. You know already for a large sample, you can invoke the CLT so:

CHAPTER 13 SIMPLE LINEAR REGRESSION. Opening Example. Simple Regression. Linear Regression

2. What is the general linear model to be used to model linear trend? (Write out the model) = or

POLYNOMIAL AND MULTIPLE REGRESSION. Polynomial regression used to fit nonlinear (e.g. curvilinear) data into a least squares linear regression model.

COMPARISONS OF CUSTOMER LOYALTY: PUBLIC & PRIVATE INSURANCE COMPANIES.

DATA INTERPRETATION AND STATISTICS

Final Exam Practice Problem Answers

2013 MBA Jump Start Program. Statistics Module Part 3

Least Squares Estimation

Chapter 13 Introduction to Linear Regression and Correlation Analysis

Introduction to Regression and Data Analysis

Additional sources Compilation of sources:

11. Analysis of Case-control Studies Logistic Regression

Bivariate Statistics Session 2: Measuring Associations Chi-Square Test

NCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( )

Week TSX Index

HYPOTHESIS TESTING: POWER OF THE TEST

Please follow the directions once you locate the Stata software in your computer. Room 114 (Business Lab) has computers with Stata software

Class 19: Two Way Tables, Conditional Distributions, Chi-Square (Text: Sections 2.5; 9.1)

Introduction to General and Generalized Linear Models

Difference of Means and ANOVA Problems

Section 13, Part 1 ANOVA. Analysis Of Variance

3.4 Statistical inference for 2 populations based on two samples

Testing for Lack of Fit

II. DISTRIBUTIONS distribution normal distribution. standard scores

How To Check For Differences In The One Way Anova

MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS

Generalized Linear Models

A Primer on Forecasting Business Performance

2. Simple Linear Regression

12.5: CHI-SQUARE GOODNESS OF FIT TESTS

Consider a study in which. How many subjects? The importance of sample size calculations. An insignificant effect: two possibilities.

1 Simple Linear Regression I Least Squares Estimation

Chapter 8 Hypothesis Testing Chapter 8 Hypothesis Testing 8-1 Overview 8-2 Basics of Hypothesis Testing

Using Excel for inferential statistics

5. Linear Regression

Time series Forecasting using Holt-Winters Exponential Smoothing

Econometrics Simple Linear Regression

Two-Sample T-Tests Assuming Equal Variance (Enter Means)

2. Linear regression with multiple regressors

LAB 4 INSTRUCTIONS CONFIDENCE INTERVALS AND HYPOTHESIS TESTING

Statistical Models in R

Descriptive Statistics

Two-Sample T-Tests Allowing Unequal Variance (Enter Difference)

Simple linear regression

Correlation and Simple Linear Regression

12: Analysis of Variance. Introduction

Principles of Hypothesis Testing for Public Health

Calculating P-Values. Parkland College. Isela Guerra Parkland College. Recommended Citation

Chapter 3 Quantitative Demand Analysis

Causal Forecasting Models

Hypothesis testing - Steps

Institute of Actuaries of India Subject CT3 Probability and Mathematical Statistics

Business Statistics. Successful completion of Introductory and/or Intermediate Algebra courses is recommended before taking Business Statistics.

Estimation of σ 2, the variance of ɛ

Overview Classes Logistic regression (5) 19-3 Building and applying logistic regression (6) 26-3 Generalizations of logistic regression (7)

Point Biserial Correlation Tests

Example: Boats and Manatees

WHAT IS A JOURNAL CLUB?

Regression Analysis (Spring, 2000)

1. The parameters to be estimated in the simple linear regression model Y=α+βx+ε ε~n(0,σ) are: a) α, β, σ b) α, β, ε c) a, b, s d) ε, 0, σ

Introduction to Quantitative Methods

SAS Software to Fit the Generalized Linear Model

Section 14 Simple Linear Regression: Introduction to Least Squares Regression

SPSS TUTORIAL & EXERCISE BOOK

Linear Models for Continuous Data

Introduction. Hypothesis Testing. Hypothesis Testing. Significance Testing

Interaction between quantitative predictors

August 2012 EXAMINATIONS Solution Part I

KSTAT MINI-MANUAL. Decision Sciences 434 Kellogg Graduate School of Management

Analysis of Variance. MINITAB User s Guide 2 3-1

Dongfeng Li. Autumn 2010

Multiple Regression in SPSS This example shows you how to perform multiple regression. The basic command is regression : linear.

Course Text. Required Computing Software. Course Description. Course Objectives. StraighterLine. Business Statistics

Basic Statistics and Data Analysis for Health Researchers from Foreign Countries

Premaster Statistics Tutorial 4 Full solutions

START Selected Topics in Assurance

Description. Textbook. Grading. Objective

The Big Picture. Correlation. Scatter Plots. Data

Bowerman, O'Connell, Aitken Schermer, & Adcock, Business Statistics in Practice, Canadian edition

Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)

Simple Regression Theory II 2010 Samuel L. Baker

Lecture 8. Confidence intervals and the central limit theorem

Confidence Intervals for Cp

Once saved, if the file was zipped you will need to unzip it. For the files that I will be posting you need to change the preferences.

Lets suppose we rolled a six-sided die 150 times and recorded the number of times each outcome (1-6) occured. The data is

Transcription:

Simple Linear Regression Inference 1

Inference requirements The Normality assumption of the stochastic term e is needed for inference even if it is not a OLS requirement. Therefore we have: Interpretation The interpretation of the Confidence Interval (CI) is fixed to a level α. If we sample more than a set of observations, each set would probably give a di_erent OLS estimate of β and therefore different CI. (1- α)% of these intervals would include β, and only (α)% of the sets would deviate from β by more than a specified Δ. 2

Standardize a Gaussian variable As we have previously seen: We can standardize such statement such that: 3

The t-distribution However σ 2 in unknown, and it must be estimated by s 2. Where t ν is a t distribution with ν= n 2 degrees of freedom. 4

CI limits Hence the CI for β at a (1 α)% can be calculated as follows: The CI specifies a range of values within which the true parameter is credible to lie. The level of likelihood is specified by the choice of α. 5

Hypothesis Testing Step 1. Fix a significance level α Step 2. Specify the null hypothesis Step 3. Specify the alternative hypothesis Step 4. Calculate the test statistic Step 5. Determine the acceptance/rejection region 6

Null and Alternative Hypothesis Given a starting conjecture (null hypothesis), considered true as working hypothesis, we may verify the Hypothesis evaluating the discrepancy between the sample observations and what one would expect from the null hypothesis. If, given a specific level of significance α, the discrepancy is significant, the null hypothesis is rejected. Otherwise it cannot be statistically rejected. It is important to notice that you may never conclusively accept the null hypothesis (Falsification Theory). 7

Hypothesis Testing: β1 = 0 We want to test whether there is a linear relationship between X and Y at a generic α = 0.05 level of significance. The hypothesis will then be: H 0 : β1 = 0 H A : β1 0 8

Hypothesis Testing: β1 = 0 The test statistic as before defined: Hence we reject the null hypothesis if: 9

Hypothesis Testing: β1 = 0 The Golden Rule When n is large, the t-distribution tends to a normal distribution, hence if we choose a 5% significance level as above we can use an approximate rule of thumb and reject the hypothesis whenever the test statistics is greater the 2. 10

Hypothesis Testing: b1 = 0 P-value and Significant levels The p-value is a probability. The rule is: we reject the null hypothesis if p-value< α. 11

P-value example Suppose you fix α = 0.5, therefore α/2 = 0.025, and the p-value associated to the coefficient β 1 is 0.001. It means that effectively there is a 1% chance that the relationship emerged randomly and a 99% chance that the relationship is real. Graphically we have: 12

Hypothesis Testing: β1 = c Similarly if we are testing for a specific constant value named c: H 0 : β1 = c H A : β1 c Hence the rejection criterion is: Keeping in mind the possibility of a Normal approximation for large samples. 13

The meaning of β 1 The parameter β 1 expresses how much change in Y follows a unitary change in X on average. if β 1 > 0 an increase in X is proportional to the increase in Y (direct relationship) if β 1 < 0 an increase in X is inversely proportional to the increase in Y (indirect relationship) 14

Regression and Correlation 15

Linear relationship The Σx i y i function measures the intensity of the linear relationship between X and Y. When this amount is expressed as mean observation contribution it is called covariance. 16

Linear relationship The covariance is a measure that depends on the measurement scalar. An alternative that is a relative index is the Pearson correlation coefficient. 17

Relationship and Dependence One must not however mistake the Pearson coefficient with the regression coefficient β 1 of a simple linear regression, as they have different formulas and different meanings. They can be linked by the following formula: Clearly the meaning also differs. The Pearson coefficient (r) measures the linear bidirectional relationship between X and Y (i.e. Y and X). 18

Linear relationship The regression coefficient β 1 measures the linear dependence of Y from X. Given a linear relationship, the dependence may be from X Y or Y X. In regression analysis one must ex ante decide which dependence to investigate. Correlation analysis in not concerned with causal links. Regression analysis is based of causal links. 19

Linear relationship Correlation does not imply causality. Example: If there are many hospital recoveries (X) in zones with many doctors (Y), it does not imply that doctors cause recoveries, but it identifies a strong linear link. 20

Residuals squared 21

Residual properties 22

Sum of squares TSS = Total Sum of Squares RSS = Residual Sum of Squares ESS = Explained Sum of Squares 23

R 2 : coefficient of determination The coefficient of determination R 2 expresses the proportion of total deviance explained by the regression model of Y on X. As one cannot explain more then all the existing deviance we can write as follows: 24

R 2 properties R 2 = 0 implies that the explicative contribution of the model is null, hence the deviance is completely expressed by the random component. R 2 = 1 implies that all N observations lie on the regression line, hence the whole variability is explained by the model. 25

R 2 properties All intermediate cases imply that the higher/lower the value of R 2 the more/less of variability is expressed by the choice of the model. A model with an R 2 = 0.80 implies that 80% of the variability is explained by the chosen model. The R 2 is said to be a fitting index as it measures the adaptiveness of the chosen model to the specified dataset. 26

R 2 and r Hence the coefficient of determination is equal to the coefficient of correlation squared. Another equally simple and efficient relationship is can be obtained by: 27

ANalysis Of VAriance The total variability decomposition shows the error component and the model component. We also know that: 28

ANalysis Of VAriance As the square of a Normal distribution is a Chi-squared distribution: Omitting here the proof for: The ratio of χ 2 divided by the d.f. is known to be an F distribution, as follows: 29

ANalysis Of VAriance One may then verify the null hypothesis H 0 : β 1 = 0 with respect to H A : β 1 0 with the following test statistic: A strong linear relationship between X and Y will determine a high valued test statistic, which supports the model, and reject the null hypothesis. Therefore: * F F 1; n 2 H0 : 1 0 is rejected empirical value theoretical value 30

Anova table 31

Forecast Often the estimated regression model is used to predict the expected Y value (i.e ˆ Y ) which corresponds to a specific value of the X variable in this case indicated with Xκ. The standard error of such value is equal to: 32

Forecast The 1 α confidence intervals for such value are then specified as: We may notice that the value of the s.e. increases proportionally to the distance between Xκ and X, therefore taking distant forecasts will worsen the quality of the estimate. It can also happen that the linear relation breaks down for very high or low Xκ values, as it is limited to the observe scatter plot. In this case taking a forecast outside the range of the observed values can be misleading. 33

Residual analysis The residual analysis is a graphical approach used to evaluate if the assumptions of the model are valid and thus to determine whether the regression model selected is appropriate. The assumption that can be graphically evaluated are: Linearity; Independence of errors; Normality of errors; Equal variance. 34

Linearity To evaluate linearity, plot the residuals on the vertical axis against the corresponding Xi values on the horizontal axis. If the linear model is appropriate fro the data, there will be no apparent pattern in this plot. 35

Independence For data collected over time, residuals can be plotted versus the time dimension to search for a relationship between consecutive residuals. Normality It can be evaluated with a normal probability plot that compare the actual versus the theoretical values of the residuals. 36

Equal variance Plot residuals versus X i. 37

Business Exercises Managerial decisions are often based on the relationship between two or more variables. For example, after considering the relationship between advertising expenditures and sales, a marketing manager might attempt to predict sales for a given level of advertising expenditure. A public utility might use the relationship between daily temperature and the demand for electricity to predict electricity consumption on next months temperature forecasts. 38

Business Exercises A candy bar manufacturer is interested in trying to estimate how sales are influenced by the price of their product. To do this the company randomly chooses six small cities and offers the candy bar at different prices. Using the candy bar sales as the dependent variable, the company will conduct a simple linear regression on the data below: 39

Business Exercises: ˆ ˆ 0, 1 40

Business Exercises: TSS 41

Business Exercises: TSS 42

Business Exercises: ESS 43

Numerical Example 44

45

46

Therefore one can conclude that 95 times out of 100 the true value β 1 lies between 0.25 and 0.37. 47

The linear relationship is very strong, 97%. As an alternative to the F-test one can present the t-test. Which implies that the variable of circulating vehicles is significant in the regression. 48