2SLS HATCO SPSS and SHAZAM Example. by Eddie Oczkowski. August X9: Usage Level (how much of the firm s total product is purchased from HATCO).


 Margery Sanders
 1 years ago
 Views:
Transcription
1 2SLS HATCO SPSS and SHAZAM Example by Eddie Oczkowski August 200 This example illustrates how to use SPSS to estimate and evaluate a 2SLS latent variable model. The bulk of the example relates to SPSS, the SHAZAM code is provided on the final page. We employ data from Hair et al (Multivariate Data Analysis, 998). The data pertain to a company called HATCO and relate to purchase outcomes from and perceptions of the company. The models presented may not necessarily be good models, we simply use them for presentation purposes. Consider a model which has a single dependent variable (usage) and two latent independent variables (strategy and image). Dependent variable X9: Usage Level (how much of the firm s total product is purchased from HATCO). Latent Independent Variables Strategy X: Delivery Speed (assume this is the scaling variable) X2: Price Level X3: Price Flexibility X7: Product Quality Image X4: Manufacturer s Image (assume this is the scaling variable) X6: Salesforce image
2 2SLS Estimation The 2SLS option is gained via: Analyze Regression 2Stage Least Squares For our basic model (usage against strategy and image) the variable boxes are filled by: Dependent Variable: Explanatory Variables: Instrumental Variables: X9 X and X4 (these are our scaling variables) X2, X3, X7 and X6 (these are our nonscaling variables) 2
3 For the diagnostic testing of the model it is useful to save the residuals and predictions from this model using Options. Part of the output from this 2SLS model is: Twostage Least Squares Equation number: Dependent variable.. X9 Multiple R R Square Adjusted R Square Standard Error Analysis of Variance: DF Sum of Squares Mean Square Regression Residuals F = Signif F =
4 Variables in the Equation Variable B SE B Beta T Sig T X X (Constant) The following new variables are being created: Name Label FIT_ Fit for X9 from 2SLS, MOD_2 Equation ERR_ Error for X9 from 2SLS, MOD_2 Equation Comments: The RSquare is 0.34 and Fstatistic being significant indicates reasonable overall fit. The two independent variables are both statistically significant with expected positive signs. Two variables have been created: FIT_ is the IV fitted value variable while ERR_ is the IV residual. 2SLS as two OLS Regressions Consider now the 2 step method for calculating estimates. This should be employed to get the 2SLS forecasts and residuals for later diagnostic testing. The first step is to run a regression for each scaling variable against all instruments and save predictions. OLS Regression: X against X2, X3, X6, X7, save predictions. OLS Regression: X4 against X2, X3, X6, X7, save predictions. Recall the Rsquare values from these runs can be examined to ascertain the possible usefulness of the instruments. 4
5 The standard OLS option is gained via: Analyze Regression Linear The st regression is: OLS Regression: X against X2, X3, X6, X7, save predictions. 5
6 Save the predictions in the Save box. Part of the output from the regression is: Regression Model Model Summary b Adjusted Std. Error of R R Square R Square the Estimate.604 a a. Predictors: (Constant), Product Quality, Salesforce Image, Price Flexibility, Price Level b. Dependent Variable: Delivery Speed 6
7 Model (Constant) Price Level Price Flexibility Salesforce Image Product Quality Coefficients a Unstandardized Coefficients a. Dependent Variable: Delivery Speed Standardi zed Coefficien ts B Std. Error Beta t Sig E Comments: The Rsquare exceeds 0.0 and some variables are significant, this indicates some instrument acceptability. Note, however, that Price Level appears not to be a good instrument. A new variable with the predictions has been saved here: pre_. The same approach is used for the other scaling variable. OLS Regression: X4 against X2, X3, X6, X7, save predictions. Part of the output from this regression is: Regression Model Model Summary b Adjusted Std. Error of R R Square R Square the Estimate.799 a a. Predictors: (Constant), Product Quality, Salesforce Image, Price Flexibility, Price Level b. Dependent Variable: Manufacturer Image 7
8 Model (Constant) Price Level Price Flexibility Salesforce Image Product Quality Coefficients a Unstandardized Coefficients a. Dependent Variable: Manufacturer Image Standardi zed Coefficien ts B Std. Error Beta t Sig E E Comments: The Rsquare is much better here, and so the instruments appear to be better for image rather than strategy. Here clearly Salesforce Image is the key instrument for the image scaling variable. A new variable with the predictions has been saved here: pre_2. The final step in the process is to OLS regress the dependent variable (X9) on the two new prediction variables (pre_ and pre_2). 8
9 To produce the 2SLS forecasts and residuals we need to use the Save option: Part of the output from the 2 nd stage regression is: Regression Model Model Summary b Adjusted Std. Error of R R Square R Square the Estimate.530 a a. Predictors: (Constant), Unstandardized Predicted Value, Unstandardized Predicted Value b. Dependent Variable: Usage Level 9
10 Model Regression Residual Total ANOVA b Sum of Squares df Mean Square F Sig a a. Predictors: (Constant), Unstandardized Predicted Value, Unstandardized Predicted Value b. Dependent Variable: Usage Level Model (Constant) Unstandardized Predicted Value Unstandardized Predicted Value a. Dependent Variable: Usage Level Coefficients a Unstandardized Coefficients Standardi zed Coefficien ts B Std. Error Beta t Sig Comments: Note how the parameter estimates are the same between this regression and the initial 2SLS model. Also note how the standard errors (and hence t and significance 2 levels) are different. The reported Rsquare is the ( GR ) generalized Rsquare referred to in the notes and this indicates how 28.% of the variation in the data is explained. This is different to the initially presented Rsquare in the 2SLS model of 34.6%. Two new variables have been saved: pre_3 which are the 2SLS forecasts and res_ which are the 2SLS residuals. Overidentifying Restrictions Test To perform this test we perform a regression of the IV residuals (err_) against all the instruments: X2, X3, X6, X7. Note the Rsquare from this regression and multiply it by the sample size (N = 00) to get the test statistic. In this case the degrees of freedom (no. of instruments less no. of RHS variables) is (4 2 = 2). At the 5% level of significance the critical value for a chisquare with d.f. = 2 is:
11 The relevant regression window is: Part of the output from this regression is: Regression Model Model Summary Adjusted Std. Error of R R Square R Square the Estimate.680 a a. Predictors: (Constant), Product Quality, Salesforce Image, Price Flexibility, Price Level
12 Model (Constant) Price Level Price Flexibility Salesforce Image Product Quality Coefficients a Unstandardized Coefficients Standardi zed Coefficien ts B Std. Error Beta t Sig a. Dependent Variable: Error for X9 from 2SLS, MOD_2 Equation Comments: The Rsquare is and so the test statistic is: N * RSquare = 00 (0.462) = 46.2, this far exceeds the critical value of 5.99 and therefore we conclude that there is a model specification problem or the instruments are invalid. There is a major problem here. Note, all the instruments are significant in this equation illustrating how the instruments can explain significant amounts of the variation in the residuals. RESET (Specification Error Test) To perform this test we first need to compute the square of the 2SLS forecasts. That is we need to compute: pre_3 *pre_3. We can call the new variable whatever we want, say, pre_32. 2
13 To do this we use the option: Transform Compute The new variable pre_32 is now added to the original 2SLS model. That is, we employ the original dependent, independent and instrumental variables, but we add to the independent variables and instrumental variables pre_32. Part of the output from this 2SLS regression is: Twostage Least Squares Dependent variable.. X9 Multiple R R Square Adjusted R Square.2483 Standard Error
14 Variables in the Equation Variable B SE B Beta T Sig T X X PRE_ (Constant) Comments: The test statistic is the tratio for pre_32. In this case the tratio is with a pvalue of This is highly insignificant. This implies that there are no omitted variables and the functional form can be trusted. Taken together with the previous test, this may imply that the problems with the model relate to inadequate instruments. Heteroscedasticity Test To perform this test we initially have to square the IV residuals using the compute option: err_2 = err_ * err_ 4
15 This new variable (err_2) is then regressed against the 2SLS forecasts (pre_32) and the tratio on the forecast variable represents the test statistic. The output from this regression is: Regression Model Model Summary Adjusted Std. Error of R R Square R Square the Estimate.069 a a. Predictors: (Constant), PRE_32 5
16 Model (Constant) PRE_32 Unstandardized Coefficients a. Dependent Variable: ERR_2 Coefficients a Standardi zed Coefficien ts B Std. Error Beta t Sig E The tratio on pre_32 is with a pvalue of 0.496, this is highly insignificant indicating the absence of heteroscedastcity. Interaction Effects To illustrate interaction effects, assume that strategy and image interact to create a new interaction latent independent variable. This variable is in addition to the original two independent variables. To create the new variables we employ the transform compute option. For the new independent variable we multiply the scaling variables by each other: say XX4 = X*X4 6
17 The instruments for this new variable are the products of all the remaining nonscaling variables across the two constructs. Since there is only one nonscaling variable for image we simply multiply it with the nonscaling variables for strategy to get our instruments: X2X6 =X2*X6 X3X6 = X3*X6 X7X6 = X7*X6 Thus the original 2SLS model is run again with one new explanatory variable XX4 and three new instrumental variables X2X6, X3X6, X7X6. Part of the output from this 2SLS regression is: Twostage Least Squares Dependent variable.. X9 Multiple R R Square.3486 Adjusted R Square Standard Error
18 Variables in the Equation Variable B SE B Beta T Sig T X X XX (Constant) Comments: Note, this model appears to be inferior to the original specification. All the variables are now insignificant, including the new interaction term XX4. Nonnested Testing To illustrate these tests consider two models: Model A: Model B: Usage Strategy Usage Image Assume we wish to ascertain which variable better explains usage. We will conduct a paired test alternating the role of Models A and B. Case H0: Null model: Usage Strategy H: Alternative model: Usage Image In terms of our notation, our x s are the strategy indicators while the w s are the image indicators. The three steps are:. Regression: X4 on X6 and save the predictions (pre_4). 2. 2SLS regression X9 on X and pre_4 (instruments: X2, X3, X7 and pre_4). 3. The tratio on the pre_4 variable is the test statistic. The output from this 2SLS regression is: Twostage Least Squares Dependent variable.. X9 Multiple R R Square.3445 Adjusted R Square Standard Error
19 Variables in the Equation Variable B SE B Beta T Sig T X PRE_ (Constant) Comments: The tratio for Pre_4 is 2.77 with a pvalue of , this is highly significant. This implies that the alternative model H image rejects the null model H0 strategy. Case 2 H0: Null model: Usage Image H: Alternative model: Usage Strategy In terms of our notation our, x s are the image indicators while the w s are the strategy indicators. The three steps are:. Regression: X on X2,X3,X7 and save the predictions (pre_5). 4. 2SLS regression X9 on X4 and pre_5 (instruments: X6 and pre_5). 5. The tratio on the pre_5 variable is the test statistic. The output from this 2SLS regression is: Twostage Least Squares Dependent variable.. X9 Multiple R R Square Adjusted R Square Standard Error Variables in the Equation Variable B SE B Beta T Sig T X PRE_ (Constant) Comments: The tratio for Pre_5 is with a pvalue of , this is highly significant. This implies that the alternative model H strategy rejects the null model H0 image. In summary these results combined imply that both models reject each other and therefore it is erroneous to use either in isolation. 9
20 2SLS HATCO SHAZAM EXAMPLE This section presents the SHAZAM code corresponding to the SPSS example. * Original 2SLS model 2SLS X9 X X4 (X2 X3 X7 X6) / PREDICT=FIT_ RESID=ERR_ * 2 step OLS version to get 2SLS predictions, residuals and GR^2 OLS X X2 X3 X6 X7 / PREDICT=PRE_ OLS X4 X2 X3 X6 X7 / PREDICT=PRE_2 OLS X9 PRE_ PRE_2 / PREDICT=PRE_3 RESID=RES_ * Overidentifying restrictions test OLS ERR_ X2 X3 X6 X7 *RESET test GENR PRE_32=PRE_3*PRE_3 2SLS X9 X X4 PRE_32 (X2 X3 X7 X6 PRE_32) * Heteroscedasticity Test GENR ERR_2=ERR_*ERR_ OLS ERR_2 PRE_32 * Interactions Model Specification GENR XX4=X*X4 GENR X2X6=X2*X6 GENR X3X6=X3*X6 GENR X7X6=X7*X6 2SLS X9 X X4 XX4 (X2 X3 X7 X6 X2X6 X3X6 X7X6) * Nonnested Test Case OLS X4 X6 / PREDICT=PRE_4 2SLS X9 X PRE_4 (X2 X3 X7 PRE_4) * Nonnested Test Case 2 OLS X X2 X3 X7 / PREDICT=PRE_5 2SLS X9 X4 PRE_5 (X6 PRE_5) 20
Linear Regression in SPSS
Linear Regression in SPSS Data: mangunkill.sav Goals: Examine relation between number of handguns registered (nhandgun) and number of man killed (mankill) checking Predict number of man killed using number
More informationSPSS Guide: Regression Analysis
SPSS Guide: Regression Analysis I put this together to give you a stepbystep guide for replicating what we did in the computer lab. It should help you run the tests we covered. The best way to get familiar
More informationMultiple Regression. Page 24
Multiple Regression Multiple regression is an extension of simple (bivariate) regression. The goal of multiple regression is to enable a researcher to assess the relationship between a dependent (predicted)
More informationMultiple Regression in SPSS This example shows you how to perform multiple regression. The basic command is regression : linear.
Multiple Regression in SPSS This example shows you how to perform multiple regression. The basic command is regression : linear. In the main dialog box, input the dependent variable and several predictors.
More informationChapter Seven. Multiple regression An introduction to multiple regression Performing a multiple regression on SPSS
Chapter Seven Multiple regression An introduction to multiple regression Performing a multiple regression on SPSS Section : An introduction to multiple regression WHAT IS MULTIPLE REGRESSION? Multiple
More informationMultiple Linear Regression
Multiple Linear Regression A regression with two or more explanatory variables is called a multiple regression. Rather than modeling the mean response as a straight line, as in simple regression, it is
More informationRegression stepbystep using Microsoft Excel
Step 1: Regression stepbystep using Microsoft Excel Notes prepared by Pamela Peterson Drake, James Madison University Type the data into the spreadsheet The example used throughout this How to is a regression
More informationRegression Analysis. Data Calculations Output
Regression Analysis In an attempt to find answers to questions such as those posed above, empirical labour economists use a useful tool called regression analysis. Regression analysis is essentially a
More informationDEPARTMENT OF PSYCHOLOGY UNIVERSITY OF LANCASTER MSC IN PSYCHOLOGICAL RESEARCH METHODS ANALYSING AND INTERPRETING DATA 2 PART 1 WEEK 9
DEPARTMENT OF PSYCHOLOGY UNIVERSITY OF LANCASTER MSC IN PSYCHOLOGICAL RESEARCH METHODS ANALYSING AND INTERPRETING DATA 2 PART 1 WEEK 9 Analysis of covariance and multiple regression So far in this course,
More informationExploring Relationships using SPSS inferential statistics (Part II) Dwayne Devonish
Exploring Relationships using SPSS inferential statistics (Part II) Dwayne Devonish Reminder: Types of Variables Categorical Variables Based on qualitative type variables. Gender, Ethnicity, religious
More informationChapter 13 Introduction to Linear Regression and Correlation Analysis
Chapter 3 Student Lecture Notes 3 Chapter 3 Introduction to Linear Regression and Correlation Analsis Fall 2006 Fundamentals of Business Statistics Chapter Goals To understand the methods for displaing
More informationWe extended the additive model in two variables to the interaction model by adding a third term to the equation.
Quadratic Models We extended the additive model in two variables to the interaction model by adding a third term to the equation. Similarly, we can extend the linear model in one variable to the quadratic
More informationBivariate Analysis. Correlation. Correlation. Pearson's Correlation Coefficient. Variable 1. Variable 2
Bivariate Analysis Variable 2 LEVELS >2 LEVELS COTIUOUS Correlation Used when you measure two continuous variables. Variable 2 2 LEVELS X 2 >2 LEVELS X 2 COTIUOUS ttest X 2 X 2 AOVA (Ftest) ttest AOVA
More informationStatistical Models in R
Statistical Models in R Some Examples Steven Buechler Department of Mathematics 276B Hurley Hall; 16233 Fall, 2007 Outline Statistical Models Linear Models in R Regression Regression analysis is the appropriate
More informationFactor Analysis. Principal components factor analysis. Use of extracted factors in multivariate dependency models
Factor Analysis Principal components factor analysis Use of extracted factors in multivariate dependency models 2 KEY CONCEPTS ***** Factor Analysis Interdependency technique Assumptions of factor analysis
More informationInstrumental Variables Regression. Instrumental Variables (IV) estimation is used when the model has endogenous s.
Instrumental Variables Regression Instrumental Variables (IV) estimation is used when the model has endogenous s. IV can thus be used to address the following important threats to internal validity: Omitted
More informationPart 2: Analysis of Relationship Between Two Variables
Part 2: Analysis of Relationship Between Two Variables Linear Regression Linear correlation Significance Tests Multiple regression Linear Regression Y = a X + b Dependent Variable Independent Variable
More informationExample: Boats and Manatees
Figure 96 Example: Boats and Manatees Slide 1 Given the sample data in Table 91, find the value of the linear correlation coefficient r, then refer to Table A6 to determine whether there is a significant
More informationMULTIPLE REGRESSION ANALYSIS OF MAIN ECONOMIC INDICATORS IN TOURISM. R, analysis of variance, Student test, multivariate analysis
Journal of tourism [No. 8] MULTIPLE REGRESSION ANALYSIS OF MAIN ECONOMIC INDICATORS IN TOURISM Assistant Ph.D. Erika KULCSÁR Babeş Bolyai University of Cluj Napoca, Romania Abstract This paper analysis
More informationLecture 5 Hypothesis Testing in Multiple Linear Regression
Lecture 5 Hypothesis Testing in Multiple Linear Regression BIOST 515 January 20, 2004 Types of tests 1 Overall test Test for addition of a single variable Test for addition of a group of variables Overall
More informationSimple linear regression
Simple linear regression Introduction Simple linear regression is a statistical method for obtaining a formula to predict values of one variable from another where there is a causal relationship between
More informationRegression Analysis (Spring, 2000)
Regression Analysis (Spring, 2000) By Wonjae Purposes: a. Explaining the relationship between Y and X variables with a model (Explain a variable Y in terms of Xs) b. Estimating and testing the intensity
More informationNCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( )
Chapter 340 Principal Components Regression Introduction is a technique for analyzing multiple regression data that suffer from multicollinearity. When multicollinearity occurs, least squares estimates
More information1. What is the critical value for this 95% confidence interval? CV = z.025 = invnorm(0.025) = 1.96
1 Final Review 2 Review 2.1 CI 1propZint Scenario 1 A TV manufacturer claims in its warranty brochure that in the past not more than 10 percent of its TV sets needed any repair during the first two years
More informationBivariate Regression Analysis. The beginning of many types of regression
Bivariate Regression Analysis The beginning of many types of regression TOPICS Beyond Correlation Forecasting Two points to estimate the slope Meeting the BLUE criterion The OLS method Purpose of Regression
More informationClass 19: Two Way Tables, Conditional Distributions, ChiSquare (Text: Sections 2.5; 9.1)
Spring 204 Class 9: Two Way Tables, Conditional Distributions, ChiSquare (Text: Sections 2.5; 9.) Big Picture: More than Two Samples In Chapter 7: We looked at quantitative variables and compared the
More informationOutline. Topic 4  Analysis of Variance Approach to Regression. Partitioning Sums of Squares. Total Sum of Squares. Partitioning sums of squares
Topic 4  Analysis of Variance Approach to Regression Outline Partitioning sums of squares Degrees of freedom Expected mean squares General linear test  Fall 2013 R 2 and the coefficient of correlation
More informationSimple Linear Regression Inference
Simple Linear Regression Inference 1 Inference requirements The Normality assumption of the stochastic term e is needed for inference even if it is not a OLS requirement. Therefore we have: Interpretation
More informationInstrumental Variables & 2SLS
Instrumental Variables & 2SLS y 1 = β 0 + β 1 y 2 + β 2 z 1 +... β k z k + u y 2 = π 0 + π 1 z k+1 + π 2 z 1 +... π k z k + v Economics 20  Prof. Schuetze 1 Why Use Instrumental Variables? Instrumental
More informationDoing Multiple Regression with SPSS. In this case, we are interested in the Analyze options so we choose that menu. If gives us a number of choices:
Doing Multiple Regression with SPSS Multiple Regression for Data Already in Data Editor Next we want to specify a multiple regression analysis for these data. The menu bar for SPSS offers several options:
More informationThe Multiple Regression Model: Hypothesis Tests and the Use of Nonsample Information
Chapter 8 The Multiple Regression Model: Hypothesis Tests and the Use of Nonsample Information An important new development that we encounter in this chapter is using the F distribution to simultaneously
More informationSimple Linear Regression in SPSS STAT 314
Simple Linear Regression in SPSS STAT 314 1. Ten Corvettes between 1 and 6 years old were randomly selected from last year s sales records in Virginia Beach, Virginia. The following data were obtained,
More information1.1. Simple Regression in Excel (Excel 2010).
.. Simple Regression in Excel (Excel 200). To get the Data Analysis tool, first click on File > Options > AddIns > Go > Select Data Analysis Toolpack & Toolpack VBA. Data Analysis is now available under
More informationRegression Analysis: A Complete Example
Regression Analysis: A Complete Example This section works out an example that includes all the topics we have discussed so far in this chapter. A complete example of regression analysis. PhotoDisc, Inc./Getty
More informationChapter 7: Simple linear regression Learning Objectives
Chapter 7: Simple linear regression Learning Objectives Reading: Section 7.1 of OpenIntro Statistics Video: Correlation vs. causation, YouTube (2:19) Video: Intro to Linear Regression, YouTube (5:18) 
More informationThis chapter will demonstrate how to perform multiple linear regression with IBM SPSS
CHAPTER 7B Multiple Regression: Statistical Methods Using IBM SPSS This chapter will demonstrate how to perform multiple linear regression with IBM SPSS first using the standard method and then using the
More informationModule 5: Multiple Regression Analysis
Using Statistical Data Using to Make Statistical Decisions: Data Multiple to Make Regression Decisions Analysis Page 1 Module 5: Multiple Regression Analysis Tom Ilvento, University of Delaware, College
More informationChapter 5 Analysis of variance SPSS Analysis of variance
Chapter 5 Analysis of variance SPSS Analysis of variance Data file used: gss.sav How to get there: Analyze Compare Means Oneway ANOVA To test the null hypothesis that several population means are equal,
More information5. Linear Regression
5. Linear Regression Outline.................................................................... 2 Simple linear regression 3 Linear model............................................................. 4
More informationChapter 23. Inferences for Regression
Chapter 23. Inferences for Regression Topics covered in this chapter: Simple Linear Regression Simple Linear Regression Example 23.1: Crying and IQ The Problem: Infants who cry easily may be more easily
More informationComparing Nested Models
Comparing Nested Models ST 430/514 Two models are nested if one model contains all the terms of the other, and at least one additional term. The larger model is the complete (or full) model, and the smaller
More informationRockefeller College University at Albany
Rockefeller College University at Albany PAD 705 Handout: Hypothesis Testing on Multiple Parameters In many cases we may wish to know whether two or more variables are jointly significant in a regression.
More informationHYPOTHESIS TESTING: CONFIDENCE INTERVALS, TTESTS, ANOVAS, AND REGRESSION
HYPOTHESIS TESTING: CONFIDENCE INTERVALS, TTESTS, ANOVAS, AND REGRESSION HOD 2990 10 November 2010 Lecture Background This is a lightning speed summary of introductory statistical methods for senior undergraduate
More informationRegression, least squares
Regression, least squares Joe Felsenstein Department of Genome Sciences and Department of Biology Regression, least squares p.1/24 Fitting a straight line X Two distinct cases: The X values are chosen
More informationMulticollinearity Richard Williams, University of Notre Dame, http://www3.nd.edu/~rwilliam/ Last revised January 13, 2015
Multicollinearity Richard Williams, University of Notre Dame, http://www3.nd.edu/~rwilliam/ Last revised January 13, 2015 Stata Example (See appendices for full example).. use http://www.nd.edu/~rwilliam/stats2/statafiles/multicoll.dta,
More informationMULTIPLE REGRESSION WITH CATEGORICAL DATA
DEPARTMENT OF POLITICAL SCIENCE AND INTERNATIONAL RELATIONS Posc/Uapp 86 MULTIPLE REGRESSION WITH CATEGORICAL DATA I. AGENDA: A. Multiple regression with categorical variables. Coding schemes. Interpreting
More informationInstrumental Variables & 2SLS
Instrumental Variables & 2SLS y 1 = β 0 + β 1 y 2 + β 2 z 1 +... β k z k + u y 2 = π 0 + π 1 z k+1 + π 2 z 1 +... π k z k + v Economics 20  Prof. Schuetze 1 Why Use Instrumental Variables? Instrumental
More informationRegression. Name: Class: Date: Multiple Choice Identify the choice that best completes the statement or answers the question.
Class: Date: Regression Multiple Choice Identify the choice that best completes the statement or answers the question. 1. Given the least squares regression line y8 = 5 2x: a. the relationship between
More informationSPSS BASICS. (Data used in this tutorial: General Social Survey 2000 and 2002) Ex: Mother s Education to eliminate responses 97,98, 99;
SPSS BASICS (Data used in this tutorial: General Social Survey 2000 and 2002) How to do Recoding Eliminating Response Categories Ex: Mother s Education to eliminate responses 97,98, 99; When we run a frequency
More informationCorrelation and Simple Linear Regression
Correlation and Simple Linear Regression We are often interested in studying the relationship among variables to determine whether they are associated with one another. When we think that changes in a
More information13. Poisson Regression Analysis
136 Poisson Regression Analysis 13. Poisson Regression Analysis We have so far considered situations where the outcome variable is numeric and Normally distributed, or binary. In clinical work one often
More informationData Mining and Data Warehousing. Henryk Maciejewski. Data Mining Predictive modelling: regression
Data Mining and Data Warehousing Henryk Maciejewski Data Mining Predictive modelling: regression Algorithms for Predictive Modelling Contents Regression Classification Auxiliary topics: Estimation of prediction
More informationChapter 5: Basic Statistics and Hypothesis Testing
Chapter 5: Basic Statistics and Hypothesis Testing In this chapter: 1. Viewing the tvalue from an OLS regression (UE 5.2.1) 2. Calculating critical tvalues and applying the decision rule (UE 5.2.2) 3.
More informationSydney Roberts Predicting Age Group Swimmers 50 Freestyle Time 1. 1. Introduction p. 2. 2. Statistical Methods Used p. 5. 3. 10 and under Males p.
Sydney Roberts Predicting Age Group Swimmers 50 Freestyle Time 1 Table of Contents 1. Introduction p. 2 2. Statistical Methods Used p. 5 3. 10 and under Males p. 8 4. 11 and up Males p. 10 5. 10 and under
More informationMULTIPLE REGRESSIONS ON SOME SELECTED MACROECONOMIC VARIABLES ON STOCK MARKET RETURNS FROM 19862010
Advances in Economics and International Finance AEIF Vol. 1(1), pp. 111, December 2014 Available online at http://www.academiaresearch.org Copyright 2014 Academia Research Full Length Research Paper MULTIPLE
More informationGeneralized Linear Models
Generalized Linear Models We have previously worked with regression models where the response variable is quantitative and normally distributed. Now we turn our attention to two types of models where the
More informationIntroduction to Regression and Data Analysis
Statlab Workshop Introduction to Regression and Data Analysis with Dan Campbell and Sherlock Campbell October 28, 2008 I. The basics A. Types of variables Your variables may take several forms, and it
More informationIAPRI Quantitative Analysis Capacity Building Series. Multiple regression analysis & interpreting results
IAPRI Quantitative Analysis Capacity Building Series Multiple regression analysis & interpreting results How important is Rsquared? Rsquared Published in Agricultural Economics 0.45 Best article of the
More informationWooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares
Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares Many economic models involve endogeneity: that is, a theoretical relationship does not fit
More informationUnivariate Regression
Univariate Regression Correlation and Regression The regression line summarizes the linear relationship between 2 variables Correlation coefficient, r, measures strength of relationship: the closer r is
More informationStatistics 112 Regression Cheatsheet Section 1B  Ryan Rosario
Statistics 112 Regression Cheatsheet Section 1B  Ryan Rosario I have found that the best way to practice regression is by brute force That is, given nothing but a dataset and your mind, compute everything
More informationFinal Exam Practice Problem Answers
Final Exam Practice Problem Answers The following data set consists of data gathered from 77 popular breakfast cereals. The variables in the data set are as follows: Brand: The brand name of the cereal
More information1.5 Oneway Analysis of Variance
Statistics: Rosie Cornish. 200. 1.5 Oneway Analysis of Variance 1 Introduction Oneway analysis of variance (ANOVA) is used to compare several means. This method is often used in scientific or medical experiments
More informationAdditional sources Compilation of sources: http://lrs.ed.uiuc.edu/tseportal/datacollectionmethodologies/jintselink/tselink.htm
Mgt 540 Research Methods Data Analysis 1 Additional sources Compilation of sources: http://lrs.ed.uiuc.edu/tseportal/datacollectionmethodologies/jintselink/tselink.htm http://web.utk.edu/~dap/random/order/start.htm
More informationSPSS: Descriptive and Inferential Statistics. For Windows
For Windows August 2012 Table of Contents Section 1: Summarizing Data...3 1.1 Descriptive Statistics...3 Section 2: Inferential Statistics... 10 2.1 ChiSquare Test... 10 2.2 T tests... 11 2.3 Correlation...
More informationDEPARTMENT OF ECONOMICS. Unit ECON 12122 Introduction to Econometrics. Notes 4 2. R and F tests
DEPARTMENT OF ECONOMICS Unit ECON 11 Introduction to Econometrics Notes 4 R and F tests These notes provide a summary of the lectures. They are not a complete account of the unit material. You should also
More informationIntroduction to Stata
Introduction to Stata September 23, 2014 Stata is one of a few statistical analysis programs that social scientists use. Stata is in the midrange of how easy it is to use. Other options include SPSS,
More informationECON 142 SKETCH OF SOLUTIONS FOR APPLIED EXERCISE #2
University of California, Berkeley Prof. Ken Chay Department of Economics Fall Semester, 005 ECON 14 SKETCH OF SOLUTIONS FOR APPLIED EXERCISE # Question 1: a. Below are the scatter plots of hourly wages
More informationData Analysis for Marketing Research  Using SPSS
North South University, School of Business MKT 63 Marketing Research Instructor: Mahmood Hussain, PhD Data Analysis for Marketing Research  Using SPSS Introduction In this part of the class, we will learn
More informationWeek TSX Index 1 8480 2 8470 3 8475 4 8510 5 8500 6 8480
1) The S & P/TSX Composite Index is based on common stock prices of a group of Canadian stocks. The weekly close level of the TSX for 6 weeks are shown: Week TSX Index 1 8480 2 8470 3 8475 4 8510 5 8500
More informationANOVA  Analysis of Variance
ANOVA  Analysis of Variance ANOVA  Analysis of Variance Extends independentsamples t test Compares the means of groups of independent observations Don t be fooled by the name. ANOVA does not compare
More informationMULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS
MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS MSR = Mean Regression Sum of Squares MSE = Mean Squared Error RSS = Regression Sum of Squares SSE = Sum of Squared Errors/Residuals α = Level of Significance
More informationStudy on the Working Capital Management Efficiency in Indian Leather Industry An Empirical Analysis
Study on the Working Capital Management Efficiency in Indian Leather Industry An Empirical Analysis Mr. N.Suresh Babu 1 Prof. G.V.Chalam 2 Research scholar Professor in Finance Dept. of Commerce and Business
More informationKSTAT MINIMANUAL. Decision Sciences 434 Kellogg Graduate School of Management
KSTAT MINIMANUAL Decision Sciences 434 Kellogg Graduate School of Management Kstat is a set of macros added to Excel and it will enable you to do the statistics required for this course very easily. To
More informationMULTIPLE REGRESSION EXAMPLE
MULTIPLE REGRESSION EXAMPLE For a sample of n = 166 college students, the following variables were measured: Y = height X 1 = mother s height ( momheight ) X 2 = father s height ( dadheight ) X 3 = 1 if
More informationTesting for Lack of Fit
Chapter 6 Testing for Lack of Fit How can we tell if a model fits the data? If the model is correct then ˆσ 2 should be an unbiased estimate of σ 2. If we have a model which is not complex enough to fit
More informationCanonical Correlation
Chapter 400 Introduction Canonical correlation analysis is the study of the linear relations between two sets of variables. It is the multivariate extension of correlation analysis. Although we will present
More informationOneWay Analysis of Variance: A Guide to Testing Differences Between Multiple Groups
OneWay Analysis of Variance: A Guide to Testing Differences Between Multiple Groups In analysis of variance, the main research question is whether the sample means are from different populations. The
More informationTesting for Granger causality between stock prices and economic growth
MPRA Munich Personal RePEc Archive Testing for Granger causality between stock prices and economic growth Pasquale Foresti 2006 Online at http://mpra.ub.unimuenchen.de/2962/ MPRA Paper No. 2962, posted
More informationPsychology 205: Research Methods in Psychology
Psychology 205: Research Methods in Psychology Using R to analyze the data for study 2 Department of Psychology Northwestern University Evanston, Illinois USA November, 2012 1 / 38 Outline 1 Getting ready
More informationPredictability Study of ISIP Reading and STAAR Reading: Prediction Bands. March 2014
Predictability Study of ISIP Reading and STAAR Reading: Prediction Bands March 2014 Chalie Patarapichayatham 1, Ph.D. William Fahle 2, Ph.D. Tracey R. Roden 3, M.Ed. 1 Research Assistant Professor in the
More informationLectures 8, 9 & 10. Multiple Regression Analysis
Lectures 8, 9 & 0. Multiple Regression Analysis In which you learn how to apply the principles and tests outlined in earlier lectures to more realistic models involving more than explanatory variable and
More informationCollinearity of independent variables. Collinearity is a condition in which some of the independent variables are highly correlated.
Collinearity of independent variables Collinearity is a condition in which some of the independent variables are highly correlated. Why is this a problem? Collinearity tends to inflate the variance of
More informationUsing R for Linear Regression
Using R for Linear Regression In the following handout words and symbols in bold are R functions and words and symbols in italics are entries supplied by the user; underlined words and symbols are optional
More informationOneWay Analysis of Variance
OneWay Analysis of Variance Note: Much of the math here is tedious but straightforward. We ll skim over it in class but you should be sure to ask questions if you don t understand it. I. Overview A. We
More information2. Simple Linear Regression
Research methods  II 3 2. Simple Linear Regression Simple linear regression is a technique in parametric statistics that is commonly used for analyzing mean response of a variable Y which changes according
More informationCanonical Correlation Analysis
Canonical Correlation Analysis LEARNING OBJECTIVES Upon completing this chapter, you should be able to do the following: State the similarities and differences between multiple regression, factor analysis,
More informationAugust 2012 EXAMINATIONS Solution Part I
August 01 EXAMINATIONS Solution Part I (1) In a random sample of 600 eligible voters, the probability that less than 38% will be in favour of this policy is closest to (B) () In a large random sample,
More informationWeek 5: Multiple Linear Regression
BUS41100 Applied Regression Analysis Week 5: Multiple Linear Regression Parameter estimation and inference, forecasting, diagnostics, dummy variables Robert B. Gramacy The University of Chicago Booth School
More informationANALYSIS OF USER ACCEPTANCE OF A NETWORK MONITORING SYSTEM WITH A FOCUS ON ICT TEACHERS
ANALYSIS OF USER ACCEPTANCE OF A NETWORK MONITORING SYSTEM WITH A FOCUS ON ICT TEACHERS Siti Rahayu Abdul Aziz 1, Mohamad Ibrahim 2, and Suhaimi Sauti 3 1 Universiti Teknologi MARA, Malaysia, rahayu@fskm.uitm.edu.my
More informationPearson s Correlation
Pearson s Correlation Correlation the degree to which two variables are associated (covary). Covariance may be either positive or negative. Its magnitude depends on the units of measurement. Assumes the
More informationRegression analysis in practice with GRETL
Regression analysis in practice with GRETL Prerequisites You will need the GNU econometrics software GRETL installed on your computer (http://gretl.sourceforge.net/), together with the sample files that
More informationINTERPRETING THE ONEWAY ANALYSIS OF VARIANCE (ANOVA)
INTERPRETING THE ONEWAY ANALYSIS OF VARIANCE (ANOVA) As with other parametric statistics, we begin the oneway ANOVA with a test of the underlying assumptions. Our first assumption is the assumption of
More informationPremaster Statistics Tutorial 4 Full solutions
Premaster Statistics Tutorial 4 Full solutions Regression analysis Q1 (based on Doane & Seward, 4/E, 12.7) a. Interpret the slope of the fitted regression = 125,000 + 150. b. What is the prediction for
More informationOrdinal Regression. Chapter
Ordinal Regression Chapter 4 Many variables of interest are ordinal. That is, you can rank the values, but the real distance between categories is unknown. Diseases are graded on scales from least severe
More informationChapter 10. Analysis of Covariance. 10.1 Multiple regression
Chapter 10 Analysis of Covariance An analysis procedure for looking at group effects on a continuous outcome when some other continuous explanatory variable also has an effect on the outcome. This chapter
More informationCOMPARISONS OF CUSTOMER LOYALTY: PUBLIC & PRIVATE INSURANCE COMPANIES.
277 CHAPTER VI COMPARISONS OF CUSTOMER LOYALTY: PUBLIC & PRIVATE INSURANCE COMPANIES. This chapter contains a full discussion of customer loyalty comparisons between private and public insurance companies
More informationRunning Head: HUMAN RESOURCE PRACTICES AND ENTERPRISE PERFORMANCE. Pakistan. Muzaffar Asad. Syed Hussain Haider. Muhammad Bilal Akhtar
Running Head: HUMAN RESOURCE PRACTICES AND ENTERPRISE PERFORMANCE Human Resource Practices and Enterprise Performance in Small and Medium Enterprises of Pakistan Muzaffar Asad Syed Hussain Haider Muhammad
More informationREGRESSION LINES IN STATA
REGRESSION LINES IN STATA THOMAS ELLIOTT 1. Introduction to Regression Regression analysis is about eploring linear relationships between a dependent variable and one or more independent variables. Regression
More informationData Analysis. Lecture Empirical Model Building and Methods (Empirische Modellbildung und Methoden) SS Analysis of Experiments  Introduction
Data Analysis Lecture Empirical Model Building and Methods (Empirische Modellbildung und Methoden) Prof. Dr. Dr. h.c. Dieter Rombach Dr. Andreas Jedlitschka SS 2014 Analysis of Experiments  Introduction
More information