2SLS HATCO SPSS and SHAZAM Example. by Eddie Oczkowski. August X9: Usage Level (how much of the firm s total product is purchased from HATCO).

Save this PDF as:
 WORD  PNG  TXT  JPG

Size: px
Start display at page:

Download "2SLS HATCO SPSS and SHAZAM Example. by Eddie Oczkowski. August X9: Usage Level (how much of the firm s total product is purchased from HATCO)."

Transcription

1 2SLS HATCO SPSS and SHAZAM Example by Eddie Oczkowski August 200 This example illustrates how to use SPSS to estimate and evaluate a 2SLS latent variable model. The bulk of the example relates to SPSS, the SHAZAM code is provided on the final page. We employ data from Hair et al (Multivariate Data Analysis, 998). The data pertain to a company called HATCO and relate to purchase outcomes from and perceptions of the company. The models presented may not necessarily be good models, we simply use them for presentation purposes. Consider a model which has a single dependent variable (usage) and two latent independent variables (strategy and image). Dependent variable X9: Usage Level (how much of the firm s total product is purchased from HATCO). Latent Independent Variables Strategy X: Delivery Speed (assume this is the scaling variable) X2: Price Level X3: Price Flexibility X7: Product Quality Image X4: Manufacturer s Image (assume this is the scaling variable) X6: Salesforce image

2 2SLS Estimation The 2SLS option is gained via: Analyze Regression 2-Stage Least Squares For our basic model (usage against strategy and image) the variable boxes are filled by: Dependent Variable: Explanatory Variables: Instrumental Variables: X9 X and X4 (these are our scaling variables) X2, X3, X7 and X6 (these are our non-scaling variables) 2

3 For the diagnostic testing of the model it is useful to save the residuals and predictions from this model using Options. Part of the output from this 2SLS model is: Two-stage Least Squares Equation number: Dependent variable.. X9 Multiple R R Square Adjusted R Square Standard Error Analysis of Variance: DF Sum of Squares Mean Square Regression Residuals F = Signif F =

4 Variables in the Equation Variable B SE B Beta T Sig T X X (Constant) The following new variables are being created: Name Label FIT_ Fit for X9 from 2SLS, MOD_2 Equation ERR_ Error for X9 from 2SLS, MOD_2 Equation Comments: The R-Square is 0.34 and F-statistic being significant indicates reasonable overall fit. The two independent variables are both statistically significant with expected positive signs. Two variables have been created: FIT_ is the IV fitted value variable while ERR_ is the IV residual. 2SLS as two OLS Regressions Consider now the 2 step method for calculating estimates. This should be employed to get the 2SLS forecasts and residuals for later diagnostic testing. The first step is to run a regression for each scaling variable against all instruments and save predictions. OLS Regression: X against X2, X3, X6, X7, save predictions. OLS Regression: X4 against X2, X3, X6, X7, save predictions. Recall the R-square values from these runs can be examined to ascertain the possible usefulness of the instruments. 4

5 The standard OLS option is gained via: Analyze Regression Linear The st regression is: OLS Regression: X against X2, X3, X6, X7, save predictions. 5

6 Save the predictions in the Save box. Part of the output from the regression is: Regression Model Model Summary b Adjusted Std. Error of R R Square R Square the Estimate.604 a a. Predictors: (Constant), Product Quality, Salesforce Image, Price Flexibility, Price Level b. Dependent Variable: Delivery Speed 6

7 Model (Constant) Price Level Price Flexibility Salesforce Image Product Quality Coefficients a Unstandardized Coefficients a. Dependent Variable: Delivery Speed Standardi zed Coefficien ts B Std. Error Beta t Sig E Comments: The R-square exceeds 0.0 and some variables are significant, this indicates some instrument acceptability. Note, however, that Price Level appears not to be a good instrument. A new variable with the predictions has been saved here: pre_. The same approach is used for the other scaling variable. OLS Regression: X4 against X2, X3, X6, X7, save predictions. Part of the output from this regression is: Regression Model Model Summary b Adjusted Std. Error of R R Square R Square the Estimate.799 a a. Predictors: (Constant), Product Quality, Salesforce Image, Price Flexibility, Price Level b. Dependent Variable: Manufacturer Image 7

8 Model (Constant) Price Level Price Flexibility Salesforce Image Product Quality Coefficients a Unstandardized Coefficients a. Dependent Variable: Manufacturer Image Standardi zed Coefficien ts B Std. Error Beta t Sig E E Comments: The R-square is much better here, and so the instruments appear to be better for image rather than strategy. Here clearly Salesforce Image is the key instrument for the image scaling variable. A new variable with the predictions has been saved here: pre_2. The final step in the process is to OLS regress the dependent variable (X9) on the two new prediction variables (pre_ and pre_2). 8

9 To produce the 2SLS forecasts and residuals we need to use the Save option: Part of the output from the 2 nd stage regression is: Regression Model Model Summary b Adjusted Std. Error of R R Square R Square the Estimate.530 a a. Predictors: (Constant), Unstandardized Predicted Value, Unstandardized Predicted Value b. Dependent Variable: Usage Level 9

10 Model Regression Residual Total ANOVA b Sum of Squares df Mean Square F Sig a a. Predictors: (Constant), Unstandardized Predicted Value, Unstandardized Predicted Value b. Dependent Variable: Usage Level Model (Constant) Unstandardized Predicted Value Unstandardized Predicted Value a. Dependent Variable: Usage Level Coefficients a Unstandardized Coefficients Standardi zed Coefficien ts B Std. Error Beta t Sig Comments: Note how the parameter estimates are the same between this regression and the initial 2SLS model. Also note how the standard errors (and hence t and significance 2 levels) are different. The reported R-square is the ( GR ) generalized R-square referred to in the notes and this indicates how 28.% of the variation in the data is explained. This is different to the initially presented R-square in the 2SLS model of 34.6%. Two new variables have been saved: pre_3 which are the 2SLS forecasts and res_ which are the 2SLS residuals. Over-identifying Restrictions Test To perform this test we perform a regression of the IV residuals (err_) against all the instruments: X2, X3, X6, X7. Note the R-square from this regression and multiply it by the sample size (N = 00) to get the test statistic. In this case the degrees of freedom (no. of instruments less no. of RHS variables) is (4 2 = 2). At the 5% level of significance the critical value for a chi-square with d.f. = 2 is:

11 The relevant regression window is: Part of the output from this regression is: Regression Model Model Summary Adjusted Std. Error of R R Square R Square the Estimate.680 a a. Predictors: (Constant), Product Quality, Salesforce Image, Price Flexibility, Price Level

12 Model (Constant) Price Level Price Flexibility Salesforce Image Product Quality Coefficients a Unstandardized Coefficients Standardi zed Coefficien ts B Std. Error Beta t Sig a. Dependent Variable: Error for X9 from 2SLS, MOD_2 Equation Comments: The R-square is and so the test statistic is: N * R-Square = 00 (0.462) = 46.2, this far exceeds the critical value of 5.99 and therefore we conclude that there is a model specification problem or the instruments are invalid. There is a major problem here. Note, all the instruments are significant in this equation illustrating how the instruments can explain significant amounts of the variation in the residuals. RESET (Specification Error Test) To perform this test we first need to compute the square of the 2SLS forecasts. That is we need to compute: pre_3 *pre_3. We can call the new variable whatever we want, say, pre_32. 2

13 To do this we use the option: Transform Compute The new variable pre_32 is now added to the original 2SLS model. That is, we employ the original dependent, independent and instrumental variables, but we add to the independent variables and instrumental variables pre_32. Part of the output from this 2SLS regression is: Two-stage Least Squares Dependent variable.. X9 Multiple R R Square Adjusted R Square.2483 Standard Error

14 Variables in the Equation Variable B SE B Beta T Sig T X X PRE_ (Constant) Comments: The test statistic is the t-ratio for pre_32. In this case the t-ratio is with a p-value of This is highly insignificant. This implies that there are no omitted variables and the functional form can be trusted. Taken together with the previous test, this may imply that the problems with the model relate to inadequate instruments. Heteroscedasticity Test To perform this test we initially have to square the IV residuals using the compute option: err_2 = err_ * err_ 4

15 This new variable (err_2) is then regressed against the 2SLS forecasts (pre_32) and the t-ratio on the forecast variable represents the test statistic. The output from this regression is: Regression Model Model Summary Adjusted Std. Error of R R Square R Square the Estimate.069 a a. Predictors: (Constant), PRE_32 5

16 Model (Constant) PRE_32 Unstandardized Coefficients a. Dependent Variable: ERR_2 Coefficients a Standardi zed Coefficien ts B Std. Error Beta t Sig E The t-ratio on pre_32 is with a p-value of 0.496, this is highly insignificant indicating the absence of heteroscedastcity. Interaction Effects To illustrate interaction effects, assume that strategy and image interact to create a new interaction latent independent variable. This variable is in addition to the original two independent variables. To create the new variables we employ the transform compute option. For the new independent variable we multiply the scaling variables by each other: say XX4 = X*X4 6

17 The instruments for this new variable are the products of all the remaining non-scaling variables across the two constructs. Since there is only one non-scaling variable for image we simply multiply it with the non-scaling variables for strategy to get our instruments: X2X6 =X2*X6 X3X6 = X3*X6 X7X6 = X7*X6 Thus the original 2SLS model is run again with one new explanatory variable XX4 and three new instrumental variables X2X6, X3X6, X7X6. Part of the output from this 2SLS regression is: Two-stage Least Squares Dependent variable.. X9 Multiple R R Square.3486 Adjusted R Square Standard Error

18 Variables in the Equation Variable B SE B Beta T Sig T X X XX (Constant) Comments: Note, this model appears to be inferior to the original specification. All the variables are now insignificant, including the new interaction term XX4. Non-nested Testing To illustrate these tests consider two models: Model A: Model B: Usage Strategy Usage Image Assume we wish to ascertain which variable better explains usage. We will conduct a paired test alternating the role of Models A and B. Case H0: Null model: Usage Strategy H: Alternative model: Usage Image In terms of our notation, our x s are the strategy indicators while the w s are the image indicators. The three steps are:. Regression: X4 on X6 and save the predictions (pre_4). 2. 2SLS regression X9 on X and pre_4 (instruments: X2, X3, X7 and pre_4). 3. The t-ratio on the pre_4 variable is the test statistic. The output from this 2SLS regression is: Two-stage Least Squares Dependent variable.. X9 Multiple R R Square.3445 Adjusted R Square Standard Error

19 Variables in the Equation Variable B SE B Beta T Sig T X PRE_ (Constant) Comments: The t-ratio for Pre_4 is 2.77 with a p-value of , this is highly significant. This implies that the alternative model H image rejects the null model H0 strategy. Case 2 H0: Null model: Usage Image H: Alternative model: Usage Strategy In terms of our notation our, x s are the image indicators while the w s are the strategy indicators. The three steps are:. Regression: X on X2,X3,X7 and save the predictions (pre_5). 4. 2SLS regression X9 on X4 and pre_5 (instruments: X6 and pre_5). 5. The t-ratio on the pre_5 variable is the test statistic. The output from this 2SLS regression is: Two-stage Least Squares Dependent variable.. X9 Multiple R R Square Adjusted R Square Standard Error Variables in the Equation Variable B SE B Beta T Sig T X PRE_ (Constant) Comments: The t-ratio for Pre_5 is with a p-value of , this is highly significant. This implies that the alternative model H strategy rejects the null model H0 image. In summary these results combined imply that both models reject each other and therefore it is erroneous to use either in isolation. 9

20 2SLS HATCO SHAZAM EXAMPLE This section presents the SHAZAM code corresponding to the SPSS example. * Original 2SLS model 2SLS X9 X X4 (X2 X3 X7 X6) / PREDICT=FIT_ RESID=ERR_ * 2 step OLS version to get 2SLS predictions, residuals and GR^2 OLS X X2 X3 X6 X7 / PREDICT=PRE_ OLS X4 X2 X3 X6 X7 / PREDICT=PRE_2 OLS X9 PRE_ PRE_2 / PREDICT=PRE_3 RESID=RES_ * Over-identifying restrictions test OLS ERR_ X2 X3 X6 X7 *RESET test GENR PRE_32=PRE_3*PRE_3 2SLS X9 X X4 PRE_32 (X2 X3 X7 X6 PRE_32) * Heteroscedasticity Test GENR ERR_2=ERR_*ERR_ OLS ERR_2 PRE_32 * Interactions Model Specification GENR XX4=X*X4 GENR X2X6=X2*X6 GENR X3X6=X3*X6 GENR X7X6=X7*X6 2SLS X9 X X4 XX4 (X2 X3 X7 X6 X2X6 X3X6 X7X6) * Non-nested Test Case OLS X4 X6 / PREDICT=PRE_4 2SLS X9 X PRE_4 (X2 X3 X7 PRE_4) * Non-nested Test Case 2 OLS X X2 X3 X7 / PREDICT=PRE_5 2SLS X9 X4 PRE_5 (X6 PRE_5) 20

Linear Regression in SPSS

Linear Regression in SPSS Linear Regression in SPSS Data: mangunkill.sav Goals: Examine relation between number of handguns registered (nhandgun) and number of man killed (mankill) checking Predict number of man killed using number

More information

SPSS Guide: Regression Analysis

SPSS Guide: Regression Analysis SPSS Guide: Regression Analysis I put this together to give you a step-by-step guide for replicating what we did in the computer lab. It should help you run the tests we covered. The best way to get familiar

More information

Multiple Regression. Page 24

Multiple Regression. Page 24 Multiple Regression Multiple regression is an extension of simple (bi-variate) regression. The goal of multiple regression is to enable a researcher to assess the relationship between a dependent (predicted)

More information

Multiple Regression in SPSS This example shows you how to perform multiple regression. The basic command is regression : linear.

Multiple Regression in SPSS This example shows you how to perform multiple regression. The basic command is regression : linear. Multiple Regression in SPSS This example shows you how to perform multiple regression. The basic command is regression : linear. In the main dialog box, input the dependent variable and several predictors.

More information

Chapter Seven. Multiple regression An introduction to multiple regression Performing a multiple regression on SPSS

Chapter Seven. Multiple regression An introduction to multiple regression Performing a multiple regression on SPSS Chapter Seven Multiple regression An introduction to multiple regression Performing a multiple regression on SPSS Section : An introduction to multiple regression WHAT IS MULTIPLE REGRESSION? Multiple

More information

Multiple Linear Regression

Multiple Linear Regression Multiple Linear Regression A regression with two or more explanatory variables is called a multiple regression. Rather than modeling the mean response as a straight line, as in simple regression, it is

More information

Regression step-by-step using Microsoft Excel

Regression step-by-step using Microsoft Excel Step 1: Regression step-by-step using Microsoft Excel Notes prepared by Pamela Peterson Drake, James Madison University Type the data into the spreadsheet The example used throughout this How to is a regression

More information

Regression Analysis. Data Calculations Output

Regression Analysis. Data Calculations Output Regression Analysis In an attempt to find answers to questions such as those posed above, empirical labour economists use a useful tool called regression analysis. Regression analysis is essentially a

More information

DEPARTMENT OF PSYCHOLOGY UNIVERSITY OF LANCASTER MSC IN PSYCHOLOGICAL RESEARCH METHODS ANALYSING AND INTERPRETING DATA 2 PART 1 WEEK 9

DEPARTMENT OF PSYCHOLOGY UNIVERSITY OF LANCASTER MSC IN PSYCHOLOGICAL RESEARCH METHODS ANALYSING AND INTERPRETING DATA 2 PART 1 WEEK 9 DEPARTMENT OF PSYCHOLOGY UNIVERSITY OF LANCASTER MSC IN PSYCHOLOGICAL RESEARCH METHODS ANALYSING AND INTERPRETING DATA 2 PART 1 WEEK 9 Analysis of covariance and multiple regression So far in this course,

More information

Exploring Relationships using SPSS inferential statistics (Part II) Dwayne Devonish

Exploring Relationships using SPSS inferential statistics (Part II) Dwayne Devonish Exploring Relationships using SPSS inferential statistics (Part II) Dwayne Devonish Reminder: Types of Variables Categorical Variables Based on qualitative type variables. Gender, Ethnicity, religious

More information

Chapter 13 Introduction to Linear Regression and Correlation Analysis

Chapter 13 Introduction to Linear Regression and Correlation Analysis Chapter 3 Student Lecture Notes 3- Chapter 3 Introduction to Linear Regression and Correlation Analsis Fall 2006 Fundamentals of Business Statistics Chapter Goals To understand the methods for displaing

More information

We extended the additive model in two variables to the interaction model by adding a third term to the equation.

We extended the additive model in two variables to the interaction model by adding a third term to the equation. Quadratic Models We extended the additive model in two variables to the interaction model by adding a third term to the equation. Similarly, we can extend the linear model in one variable to the quadratic

More information

Bivariate Analysis. Correlation. Correlation. Pearson's Correlation Coefficient. Variable 1. Variable 2

Bivariate Analysis. Correlation. Correlation. Pearson's Correlation Coefficient. Variable 1. Variable 2 Bivariate Analysis Variable 2 LEVELS >2 LEVELS COTIUOUS Correlation Used when you measure two continuous variables. Variable 2 2 LEVELS X 2 >2 LEVELS X 2 COTIUOUS t-test X 2 X 2 AOVA (F-test) t-test AOVA

More information

Statistical Models in R

Statistical Models in R Statistical Models in R Some Examples Steven Buechler Department of Mathematics 276B Hurley Hall; 1-6233 Fall, 2007 Outline Statistical Models Linear Models in R Regression Regression analysis is the appropriate

More information

Factor Analysis. Principal components factor analysis. Use of extracted factors in multivariate dependency models

Factor Analysis. Principal components factor analysis. Use of extracted factors in multivariate dependency models Factor Analysis Principal components factor analysis Use of extracted factors in multivariate dependency models 2 KEY CONCEPTS ***** Factor Analysis Interdependency technique Assumptions of factor analysis

More information

Instrumental Variables Regression. Instrumental Variables (IV) estimation is used when the model has endogenous s.

Instrumental Variables Regression. Instrumental Variables (IV) estimation is used when the model has endogenous s. Instrumental Variables Regression Instrumental Variables (IV) estimation is used when the model has endogenous s. IV can thus be used to address the following important threats to internal validity: Omitted

More information

Part 2: Analysis of Relationship Between Two Variables

Part 2: Analysis of Relationship Between Two Variables Part 2: Analysis of Relationship Between Two Variables Linear Regression Linear correlation Significance Tests Multiple regression Linear Regression Y = a X + b Dependent Variable Independent Variable

More information

Example: Boats and Manatees

Example: Boats and Manatees Figure 9-6 Example: Boats and Manatees Slide 1 Given the sample data in Table 9-1, find the value of the linear correlation coefficient r, then refer to Table A-6 to determine whether there is a significant

More information

MULTIPLE REGRESSION ANALYSIS OF MAIN ECONOMIC INDICATORS IN TOURISM. R, analysis of variance, Student test, multivariate analysis

MULTIPLE REGRESSION ANALYSIS OF MAIN ECONOMIC INDICATORS IN TOURISM. R, analysis of variance, Student test, multivariate analysis Journal of tourism [No. 8] MULTIPLE REGRESSION ANALYSIS OF MAIN ECONOMIC INDICATORS IN TOURISM Assistant Ph.D. Erika KULCSÁR Babeş Bolyai University of Cluj Napoca, Romania Abstract This paper analysis

More information

Lecture 5 Hypothesis Testing in Multiple Linear Regression

Lecture 5 Hypothesis Testing in Multiple Linear Regression Lecture 5 Hypothesis Testing in Multiple Linear Regression BIOST 515 January 20, 2004 Types of tests 1 Overall test Test for addition of a single variable Test for addition of a group of variables Overall

More information

Simple linear regression

Simple linear regression Simple linear regression Introduction Simple linear regression is a statistical method for obtaining a formula to predict values of one variable from another where there is a causal relationship between

More information

Regression Analysis (Spring, 2000)

Regression Analysis (Spring, 2000) Regression Analysis (Spring, 2000) By Wonjae Purposes: a. Explaining the relationship between Y and X variables with a model (Explain a variable Y in terms of Xs) b. Estimating and testing the intensity

More information

NCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( )

NCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( ) Chapter 340 Principal Components Regression Introduction is a technique for analyzing multiple regression data that suffer from multicollinearity. When multicollinearity occurs, least squares estimates

More information

1. What is the critical value for this 95% confidence interval? CV = z.025 = invnorm(0.025) = 1.96

1. What is the critical value for this 95% confidence interval? CV = z.025 = invnorm(0.025) = 1.96 1 Final Review 2 Review 2.1 CI 1-propZint Scenario 1 A TV manufacturer claims in its warranty brochure that in the past not more than 10 percent of its TV sets needed any repair during the first two years

More information

Bivariate Regression Analysis. The beginning of many types of regression

Bivariate Regression Analysis. The beginning of many types of regression Bivariate Regression Analysis The beginning of many types of regression TOPICS Beyond Correlation Forecasting Two points to estimate the slope Meeting the BLUE criterion The OLS method Purpose of Regression

More information

Class 19: Two Way Tables, Conditional Distributions, Chi-Square (Text: Sections 2.5; 9.1)

Class 19: Two Way Tables, Conditional Distributions, Chi-Square (Text: Sections 2.5; 9.1) Spring 204 Class 9: Two Way Tables, Conditional Distributions, Chi-Square (Text: Sections 2.5; 9.) Big Picture: More than Two Samples In Chapter 7: We looked at quantitative variables and compared the

More information

Outline. Topic 4 - Analysis of Variance Approach to Regression. Partitioning Sums of Squares. Total Sum of Squares. Partitioning sums of squares

Outline. Topic 4 - Analysis of Variance Approach to Regression. Partitioning Sums of Squares. Total Sum of Squares. Partitioning sums of squares Topic 4 - Analysis of Variance Approach to Regression Outline Partitioning sums of squares Degrees of freedom Expected mean squares General linear test - Fall 2013 R 2 and the coefficient of correlation

More information

Simple Linear Regression Inference

Simple Linear Regression Inference Simple Linear Regression Inference 1 Inference requirements The Normality assumption of the stochastic term e is needed for inference even if it is not a OLS requirement. Therefore we have: Interpretation

More information

Instrumental Variables & 2SLS

Instrumental Variables & 2SLS Instrumental Variables & 2SLS y 1 = β 0 + β 1 y 2 + β 2 z 1 +... β k z k + u y 2 = π 0 + π 1 z k+1 + π 2 z 1 +... π k z k + v Economics 20 - Prof. Schuetze 1 Why Use Instrumental Variables? Instrumental

More information

Doing Multiple Regression with SPSS. In this case, we are interested in the Analyze options so we choose that menu. If gives us a number of choices:

Doing Multiple Regression with SPSS. In this case, we are interested in the Analyze options so we choose that menu. If gives us a number of choices: Doing Multiple Regression with SPSS Multiple Regression for Data Already in Data Editor Next we want to specify a multiple regression analysis for these data. The menu bar for SPSS offers several options:

More information

The Multiple Regression Model: Hypothesis Tests and the Use of Nonsample Information

The Multiple Regression Model: Hypothesis Tests and the Use of Nonsample Information Chapter 8 The Multiple Regression Model: Hypothesis Tests and the Use of Nonsample Information An important new development that we encounter in this chapter is using the F- distribution to simultaneously

More information

Simple Linear Regression in SPSS STAT 314

Simple Linear Regression in SPSS STAT 314 Simple Linear Regression in SPSS STAT 314 1. Ten Corvettes between 1 and 6 years old were randomly selected from last year s sales records in Virginia Beach, Virginia. The following data were obtained,

More information

1.1. Simple Regression in Excel (Excel 2010).

1.1. Simple Regression in Excel (Excel 2010). .. Simple Regression in Excel (Excel 200). To get the Data Analysis tool, first click on File > Options > Add-Ins > Go > Select Data Analysis Toolpack & Toolpack VBA. Data Analysis is now available under

More information

Regression Analysis: A Complete Example

Regression Analysis: A Complete Example Regression Analysis: A Complete Example This section works out an example that includes all the topics we have discussed so far in this chapter. A complete example of regression analysis. PhotoDisc, Inc./Getty

More information

Chapter 7: Simple linear regression Learning Objectives

Chapter 7: Simple linear regression Learning Objectives Chapter 7: Simple linear regression Learning Objectives Reading: Section 7.1 of OpenIntro Statistics Video: Correlation vs. causation, YouTube (2:19) Video: Intro to Linear Regression, YouTube (5:18) -

More information

This chapter will demonstrate how to perform multiple linear regression with IBM SPSS

This chapter will demonstrate how to perform multiple linear regression with IBM SPSS CHAPTER 7B Multiple Regression: Statistical Methods Using IBM SPSS This chapter will demonstrate how to perform multiple linear regression with IBM SPSS first using the standard method and then using the

More information

Module 5: Multiple Regression Analysis

Module 5: Multiple Regression Analysis Using Statistical Data Using to Make Statistical Decisions: Data Multiple to Make Regression Decisions Analysis Page 1 Module 5: Multiple Regression Analysis Tom Ilvento, University of Delaware, College

More information

Chapter 5 Analysis of variance SPSS Analysis of variance

Chapter 5 Analysis of variance SPSS Analysis of variance Chapter 5 Analysis of variance SPSS Analysis of variance Data file used: gss.sav How to get there: Analyze Compare Means One-way ANOVA To test the null hypothesis that several population means are equal,

More information

5. Linear Regression

5. Linear Regression 5. Linear Regression Outline.................................................................... 2 Simple linear regression 3 Linear model............................................................. 4

More information

Chapter 23. Inferences for Regression

Chapter 23. Inferences for Regression Chapter 23. Inferences for Regression Topics covered in this chapter: Simple Linear Regression Simple Linear Regression Example 23.1: Crying and IQ The Problem: Infants who cry easily may be more easily

More information

Comparing Nested Models

Comparing Nested Models Comparing Nested Models ST 430/514 Two models are nested if one model contains all the terms of the other, and at least one additional term. The larger model is the complete (or full) model, and the smaller

More information

Rockefeller College University at Albany

Rockefeller College University at Albany Rockefeller College University at Albany PAD 705 Handout: Hypothesis Testing on Multiple Parameters In many cases we may wish to know whether two or more variables are jointly significant in a regression.

More information

HYPOTHESIS TESTING: CONFIDENCE INTERVALS, T-TESTS, ANOVAS, AND REGRESSION

HYPOTHESIS TESTING: CONFIDENCE INTERVALS, T-TESTS, ANOVAS, AND REGRESSION HYPOTHESIS TESTING: CONFIDENCE INTERVALS, T-TESTS, ANOVAS, AND REGRESSION HOD 2990 10 November 2010 Lecture Background This is a lightning speed summary of introductory statistical methods for senior undergraduate

More information

Regression, least squares

Regression, least squares Regression, least squares Joe Felsenstein Department of Genome Sciences and Department of Biology Regression, least squares p.1/24 Fitting a straight line X Two distinct cases: The X values are chosen

More information

Multicollinearity Richard Williams, University of Notre Dame, http://www3.nd.edu/~rwilliam/ Last revised January 13, 2015

Multicollinearity Richard Williams, University of Notre Dame, http://www3.nd.edu/~rwilliam/ Last revised January 13, 2015 Multicollinearity Richard Williams, University of Notre Dame, http://www3.nd.edu/~rwilliam/ Last revised January 13, 2015 Stata Example (See appendices for full example).. use http://www.nd.edu/~rwilliam/stats2/statafiles/multicoll.dta,

More information

MULTIPLE REGRESSION WITH CATEGORICAL DATA

MULTIPLE REGRESSION WITH CATEGORICAL DATA DEPARTMENT OF POLITICAL SCIENCE AND INTERNATIONAL RELATIONS Posc/Uapp 86 MULTIPLE REGRESSION WITH CATEGORICAL DATA I. AGENDA: A. Multiple regression with categorical variables. Coding schemes. Interpreting

More information

Instrumental Variables & 2SLS

Instrumental Variables & 2SLS Instrumental Variables & 2SLS y 1 = β 0 + β 1 y 2 + β 2 z 1 +... β k z k + u y 2 = π 0 + π 1 z k+1 + π 2 z 1 +... π k z k + v Economics 20 - Prof. Schuetze 1 Why Use Instrumental Variables? Instrumental

More information

Regression. Name: Class: Date: Multiple Choice Identify the choice that best completes the statement or answers the question.

Regression. Name: Class: Date: Multiple Choice Identify the choice that best completes the statement or answers the question. Class: Date: Regression Multiple Choice Identify the choice that best completes the statement or answers the question. 1. Given the least squares regression line y8 = 5 2x: a. the relationship between

More information

SPSS BASICS. (Data used in this tutorial: General Social Survey 2000 and 2002) Ex: Mother s Education to eliminate responses 97,98, 99;

SPSS BASICS. (Data used in this tutorial: General Social Survey 2000 and 2002) Ex: Mother s Education to eliminate responses 97,98, 99; SPSS BASICS (Data used in this tutorial: General Social Survey 2000 and 2002) How to do Recoding Eliminating Response Categories Ex: Mother s Education to eliminate responses 97,98, 99; When we run a frequency

More information

Correlation and Simple Linear Regression

Correlation and Simple Linear Regression Correlation and Simple Linear Regression We are often interested in studying the relationship among variables to determine whether they are associated with one another. When we think that changes in a

More information

13. Poisson Regression Analysis

13. Poisson Regression Analysis 136 Poisson Regression Analysis 13. Poisson Regression Analysis We have so far considered situations where the outcome variable is numeric and Normally distributed, or binary. In clinical work one often

More information

Data Mining and Data Warehousing. Henryk Maciejewski. Data Mining Predictive modelling: regression

Data Mining and Data Warehousing. Henryk Maciejewski. Data Mining Predictive modelling: regression Data Mining and Data Warehousing Henryk Maciejewski Data Mining Predictive modelling: regression Algorithms for Predictive Modelling Contents Regression Classification Auxiliary topics: Estimation of prediction

More information

Chapter 5: Basic Statistics and Hypothesis Testing

Chapter 5: Basic Statistics and Hypothesis Testing Chapter 5: Basic Statistics and Hypothesis Testing In this chapter: 1. Viewing the t-value from an OLS regression (UE 5.2.1) 2. Calculating critical t-values and applying the decision rule (UE 5.2.2) 3.

More information

Sydney Roberts Predicting Age Group Swimmers 50 Freestyle Time 1. 1. Introduction p. 2. 2. Statistical Methods Used p. 5. 3. 10 and under Males p.

Sydney Roberts Predicting Age Group Swimmers 50 Freestyle Time 1. 1. Introduction p. 2. 2. Statistical Methods Used p. 5. 3. 10 and under Males p. Sydney Roberts Predicting Age Group Swimmers 50 Freestyle Time 1 Table of Contents 1. Introduction p. 2 2. Statistical Methods Used p. 5 3. 10 and under Males p. 8 4. 11 and up Males p. 10 5. 10 and under

More information

MULTIPLE REGRESSIONS ON SOME SELECTED MACROECONOMIC VARIABLES ON STOCK MARKET RETURNS FROM 1986-2010

MULTIPLE REGRESSIONS ON SOME SELECTED MACROECONOMIC VARIABLES ON STOCK MARKET RETURNS FROM 1986-2010 Advances in Economics and International Finance AEIF Vol. 1(1), pp. 1-11, December 2014 Available online at http://www.academiaresearch.org Copyright 2014 Academia Research Full Length Research Paper MULTIPLE

More information

Generalized Linear Models

Generalized Linear Models Generalized Linear Models We have previously worked with regression models where the response variable is quantitative and normally distributed. Now we turn our attention to two types of models where the

More information

Introduction to Regression and Data Analysis

Introduction to Regression and Data Analysis Statlab Workshop Introduction to Regression and Data Analysis with Dan Campbell and Sherlock Campbell October 28, 2008 I. The basics A. Types of variables Your variables may take several forms, and it

More information

IAPRI Quantitative Analysis Capacity Building Series. Multiple regression analysis & interpreting results

IAPRI Quantitative Analysis Capacity Building Series. Multiple regression analysis & interpreting results IAPRI Quantitative Analysis Capacity Building Series Multiple regression analysis & interpreting results How important is R-squared? R-squared Published in Agricultural Economics 0.45 Best article of the

More information

Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares

Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares Many economic models involve endogeneity: that is, a theoretical relationship does not fit

More information

Univariate Regression

Univariate Regression Univariate Regression Correlation and Regression The regression line summarizes the linear relationship between 2 variables Correlation coefficient, r, measures strength of relationship: the closer r is

More information

Statistics 112 Regression Cheatsheet Section 1B - Ryan Rosario

Statistics 112 Regression Cheatsheet Section 1B - Ryan Rosario Statistics 112 Regression Cheatsheet Section 1B - Ryan Rosario I have found that the best way to practice regression is by brute force That is, given nothing but a dataset and your mind, compute everything

More information

Final Exam Practice Problem Answers

Final Exam Practice Problem Answers Final Exam Practice Problem Answers The following data set consists of data gathered from 77 popular breakfast cereals. The variables in the data set are as follows: Brand: The brand name of the cereal

More information

1.5 Oneway Analysis of Variance

1.5 Oneway Analysis of Variance Statistics: Rosie Cornish. 200. 1.5 Oneway Analysis of Variance 1 Introduction Oneway analysis of variance (ANOVA) is used to compare several means. This method is often used in scientific or medical experiments

More information

Additional sources Compilation of sources: http://lrs.ed.uiuc.edu/tseportal/datacollectionmethodologies/jin-tselink/tselink.htm

Additional sources Compilation of sources: http://lrs.ed.uiuc.edu/tseportal/datacollectionmethodologies/jin-tselink/tselink.htm Mgt 540 Research Methods Data Analysis 1 Additional sources Compilation of sources: http://lrs.ed.uiuc.edu/tseportal/datacollectionmethodologies/jin-tselink/tselink.htm http://web.utk.edu/~dap/random/order/start.htm

More information

SPSS: Descriptive and Inferential Statistics. For Windows

SPSS: Descriptive and Inferential Statistics. For Windows For Windows August 2012 Table of Contents Section 1: Summarizing Data...3 1.1 Descriptive Statistics...3 Section 2: Inferential Statistics... 10 2.1 Chi-Square Test... 10 2.2 T tests... 11 2.3 Correlation...

More information

DEPARTMENT OF ECONOMICS. Unit ECON 12122 Introduction to Econometrics. Notes 4 2. R and F tests

DEPARTMENT OF ECONOMICS. Unit ECON 12122 Introduction to Econometrics. Notes 4 2. R and F tests DEPARTMENT OF ECONOMICS Unit ECON 11 Introduction to Econometrics Notes 4 R and F tests These notes provide a summary of the lectures. They are not a complete account of the unit material. You should also

More information

Introduction to Stata

Introduction to Stata Introduction to Stata September 23, 2014 Stata is one of a few statistical analysis programs that social scientists use. Stata is in the mid-range of how easy it is to use. Other options include SPSS,

More information

ECON 142 SKETCH OF SOLUTIONS FOR APPLIED EXERCISE #2

ECON 142 SKETCH OF SOLUTIONS FOR APPLIED EXERCISE #2 University of California, Berkeley Prof. Ken Chay Department of Economics Fall Semester, 005 ECON 14 SKETCH OF SOLUTIONS FOR APPLIED EXERCISE # Question 1: a. Below are the scatter plots of hourly wages

More information

Data Analysis for Marketing Research - Using SPSS

Data Analysis for Marketing Research - Using SPSS North South University, School of Business MKT 63 Marketing Research Instructor: Mahmood Hussain, PhD Data Analysis for Marketing Research - Using SPSS Introduction In this part of the class, we will learn

More information

Week TSX Index 1 8480 2 8470 3 8475 4 8510 5 8500 6 8480

Week TSX Index 1 8480 2 8470 3 8475 4 8510 5 8500 6 8480 1) The S & P/TSX Composite Index is based on common stock prices of a group of Canadian stocks. The weekly close level of the TSX for 6 weeks are shown: Week TSX Index 1 8480 2 8470 3 8475 4 8510 5 8500

More information

ANOVA - Analysis of Variance

ANOVA - Analysis of Variance ANOVA - Analysis of Variance ANOVA - Analysis of Variance Extends independent-samples t test Compares the means of groups of independent observations Don t be fooled by the name. ANOVA does not compare

More information

MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS

MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS MSR = Mean Regression Sum of Squares MSE = Mean Squared Error RSS = Regression Sum of Squares SSE = Sum of Squared Errors/Residuals α = Level of Significance

More information

Study on the Working Capital Management Efficiency in Indian Leather Industry- An Empirical Analysis

Study on the Working Capital Management Efficiency in Indian Leather Industry- An Empirical Analysis Study on the Working Capital Management Efficiency in Indian Leather Industry- An Empirical Analysis Mr. N.Suresh Babu 1 Prof. G.V.Chalam 2 Research scholar Professor in Finance Dept. of Commerce and Business

More information

KSTAT MINI-MANUAL. Decision Sciences 434 Kellogg Graduate School of Management

KSTAT MINI-MANUAL. Decision Sciences 434 Kellogg Graduate School of Management KSTAT MINI-MANUAL Decision Sciences 434 Kellogg Graduate School of Management Kstat is a set of macros added to Excel and it will enable you to do the statistics required for this course very easily. To

More information

MULTIPLE REGRESSION EXAMPLE

MULTIPLE REGRESSION EXAMPLE MULTIPLE REGRESSION EXAMPLE For a sample of n = 166 college students, the following variables were measured: Y = height X 1 = mother s height ( momheight ) X 2 = father s height ( dadheight ) X 3 = 1 if

More information

Testing for Lack of Fit

Testing for Lack of Fit Chapter 6 Testing for Lack of Fit How can we tell if a model fits the data? If the model is correct then ˆσ 2 should be an unbiased estimate of σ 2. If we have a model which is not complex enough to fit

More information

Canonical Correlation

Canonical Correlation Chapter 400 Introduction Canonical correlation analysis is the study of the linear relations between two sets of variables. It is the multivariate extension of correlation analysis. Although we will present

More information

One-Way Analysis of Variance: A Guide to Testing Differences Between Multiple Groups

One-Way Analysis of Variance: A Guide to Testing Differences Between Multiple Groups One-Way Analysis of Variance: A Guide to Testing Differences Between Multiple Groups In analysis of variance, the main research question is whether the sample means are from different populations. The

More information

Testing for Granger causality between stock prices and economic growth

Testing for Granger causality between stock prices and economic growth MPRA Munich Personal RePEc Archive Testing for Granger causality between stock prices and economic growth Pasquale Foresti 2006 Online at http://mpra.ub.uni-muenchen.de/2962/ MPRA Paper No. 2962, posted

More information

Psychology 205: Research Methods in Psychology

Psychology 205: Research Methods in Psychology Psychology 205: Research Methods in Psychology Using R to analyze the data for study 2 Department of Psychology Northwestern University Evanston, Illinois USA November, 2012 1 / 38 Outline 1 Getting ready

More information

Predictability Study of ISIP Reading and STAAR Reading: Prediction Bands. March 2014

Predictability Study of ISIP Reading and STAAR Reading: Prediction Bands. March 2014 Predictability Study of ISIP Reading and STAAR Reading: Prediction Bands March 2014 Chalie Patarapichayatham 1, Ph.D. William Fahle 2, Ph.D. Tracey R. Roden 3, M.Ed. 1 Research Assistant Professor in the

More information

Lectures 8, 9 & 10. Multiple Regression Analysis

Lectures 8, 9 & 10. Multiple Regression Analysis Lectures 8, 9 & 0. Multiple Regression Analysis In which you learn how to apply the principles and tests outlined in earlier lectures to more realistic models involving more than explanatory variable and

More information

Collinearity of independent variables. Collinearity is a condition in which some of the independent variables are highly correlated.

Collinearity of independent variables. Collinearity is a condition in which some of the independent variables are highly correlated. Collinearity of independent variables Collinearity is a condition in which some of the independent variables are highly correlated. Why is this a problem? Collinearity tends to inflate the variance of

More information

Using R for Linear Regression

Using R for Linear Regression Using R for Linear Regression In the following handout words and symbols in bold are R functions and words and symbols in italics are entries supplied by the user; underlined words and symbols are optional

More information

One-Way Analysis of Variance

One-Way Analysis of Variance One-Way Analysis of Variance Note: Much of the math here is tedious but straightforward. We ll skim over it in class but you should be sure to ask questions if you don t understand it. I. Overview A. We

More information

2. Simple Linear Regression

2. Simple Linear Regression Research methods - II 3 2. Simple Linear Regression Simple linear regression is a technique in parametric statistics that is commonly used for analyzing mean response of a variable Y which changes according

More information

Canonical Correlation Analysis

Canonical Correlation Analysis Canonical Correlation Analysis LEARNING OBJECTIVES Upon completing this chapter, you should be able to do the following: State the similarities and differences between multiple regression, factor analysis,

More information

August 2012 EXAMINATIONS Solution Part I

August 2012 EXAMINATIONS Solution Part I August 01 EXAMINATIONS Solution Part I (1) In a random sample of 600 eligible voters, the probability that less than 38% will be in favour of this policy is closest to (B) () In a large random sample,

More information

Week 5: Multiple Linear Regression

Week 5: Multiple Linear Regression BUS41100 Applied Regression Analysis Week 5: Multiple Linear Regression Parameter estimation and inference, forecasting, diagnostics, dummy variables Robert B. Gramacy The University of Chicago Booth School

More information

ANALYSIS OF USER ACCEPTANCE OF A NETWORK MONITORING SYSTEM WITH A FOCUS ON ICT TEACHERS

ANALYSIS OF USER ACCEPTANCE OF A NETWORK MONITORING SYSTEM WITH A FOCUS ON ICT TEACHERS ANALYSIS OF USER ACCEPTANCE OF A NETWORK MONITORING SYSTEM WITH A FOCUS ON ICT TEACHERS Siti Rahayu Abdul Aziz 1, Mohamad Ibrahim 2, and Suhaimi Sauti 3 1 Universiti Teknologi MARA, Malaysia, rahayu@fskm.uitm.edu.my

More information

Pearson s Correlation

Pearson s Correlation Pearson s Correlation Correlation the degree to which two variables are associated (co-vary). Covariance may be either positive or negative. Its magnitude depends on the units of measurement. Assumes the

More information

Regression analysis in practice with GRETL

Regression analysis in practice with GRETL Regression analysis in practice with GRETL Prerequisites You will need the GNU econometrics software GRETL installed on your computer (http://gretl.sourceforge.net/), together with the sample files that

More information

INTERPRETING THE ONE-WAY ANALYSIS OF VARIANCE (ANOVA)

INTERPRETING THE ONE-WAY ANALYSIS OF VARIANCE (ANOVA) INTERPRETING THE ONE-WAY ANALYSIS OF VARIANCE (ANOVA) As with other parametric statistics, we begin the one-way ANOVA with a test of the underlying assumptions. Our first assumption is the assumption of

More information

Premaster Statistics Tutorial 4 Full solutions

Premaster Statistics Tutorial 4 Full solutions Premaster Statistics Tutorial 4 Full solutions Regression analysis Q1 (based on Doane & Seward, 4/E, 12.7) a. Interpret the slope of the fitted regression = 125,000 + 150. b. What is the prediction for

More information

Ordinal Regression. Chapter

Ordinal Regression. Chapter Ordinal Regression Chapter 4 Many variables of interest are ordinal. That is, you can rank the values, but the real distance between categories is unknown. Diseases are graded on scales from least severe

More information

Chapter 10. Analysis of Covariance. 10.1 Multiple regression

Chapter 10. Analysis of Covariance. 10.1 Multiple regression Chapter 10 Analysis of Covariance An analysis procedure for looking at group effects on a continuous outcome when some other continuous explanatory variable also has an effect on the outcome. This chapter

More information

COMPARISONS OF CUSTOMER LOYALTY: PUBLIC & PRIVATE INSURANCE COMPANIES.

COMPARISONS OF CUSTOMER LOYALTY: PUBLIC & PRIVATE INSURANCE COMPANIES. 277 CHAPTER VI COMPARISONS OF CUSTOMER LOYALTY: PUBLIC & PRIVATE INSURANCE COMPANIES. This chapter contains a full discussion of customer loyalty comparisons between private and public insurance companies

More information

Running Head: HUMAN RESOURCE PRACTICES AND ENTERPRISE PERFORMANCE. Pakistan. Muzaffar Asad. Syed Hussain Haider. Muhammad Bilal Akhtar

Running Head: HUMAN RESOURCE PRACTICES AND ENTERPRISE PERFORMANCE. Pakistan. Muzaffar Asad. Syed Hussain Haider. Muhammad Bilal Akhtar Running Head: HUMAN RESOURCE PRACTICES AND ENTERPRISE PERFORMANCE Human Resource Practices and Enterprise Performance in Small and Medium Enterprises of Pakistan Muzaffar Asad Syed Hussain Haider Muhammad

More information

REGRESSION LINES IN STATA

REGRESSION LINES IN STATA REGRESSION LINES IN STATA THOMAS ELLIOTT 1. Introduction to Regression Regression analysis is about eploring linear relationships between a dependent variable and one or more independent variables. Regression

More information

Data Analysis. Lecture Empirical Model Building and Methods (Empirische Modellbildung und Methoden) SS Analysis of Experiments - Introduction

Data Analysis. Lecture Empirical Model Building and Methods (Empirische Modellbildung und Methoden) SS Analysis of Experiments - Introduction Data Analysis Lecture Empirical Model Building and Methods (Empirische Modellbildung und Methoden) Prof. Dr. Dr. h.c. Dieter Rombach Dr. Andreas Jedlitschka SS 2014 Analysis of Experiments - Introduction

More information