SPSS Guide: Regression Analysis



Similar documents
Multiple Regression in SPSS This example shows you how to perform multiple regression. The basic command is regression : linear.

Multiple Regression. Page 24

1.1. Simple Regression in Excel (Excel 2010).

Chapter 13 Introduction to Linear Regression and Correlation Analysis

Simple Linear Regression, Scatterplots, and Bivariate Correlation

This chapter will demonstrate how to perform multiple linear regression with IBM SPSS

HYPOTHESIS TESTING: CONFIDENCE INTERVALS, T-TESTS, ANOVAS, AND REGRESSION

Unit 31 A Hypothesis Test about Correlation and Slope in a Simple Linear Regression

Doing Multiple Regression with SPSS. In this case, we are interested in the Analyze options so we choose that menu. If gives us a number of choices:

DEPARTMENT OF PSYCHOLOGY UNIVERSITY OF LANCASTER MSC IN PSYCHOLOGICAL RESEARCH METHODS ANALYSING AND INTERPRETING DATA 2 PART 1 WEEK 9

Regression step-by-step using Microsoft Excel

Chapter Seven. Multiple regression An introduction to multiple regression Performing a multiple regression on SPSS

Chapter 23. Inferences for Regression

Directions for using SPSS

Data Analysis for Marketing Research - Using SPSS

Chapter 5 Analysis of variance SPSS Analysis of variance

KSTAT MINI-MANUAL. Decision Sciences 434 Kellogg Graduate School of Management

Course Objective This course is designed to give you a basic understanding of how to run regressions in SPSS.

Formula for linear models. Prediction, extrapolation, significance test against zero slope.

Two Related Samples t Test

" Y. Notation and Equations for Regression Lecture 11/4. Notation:

Part 2: Analysis of Relationship Between Two Variables

A STUDY ON ONBOARDING PROCESS IN SIFY TECHNOLOGIES, CHENNAI

Multicollinearity Richard Williams, University of Notre Dame, Last revised January 13, 2015

Chapter 7: Simple linear regression Learning Objectives

Week TSX Index

Rockefeller College University at Albany

NCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( )

Introduction to Regression and Data Analysis

Multiple Linear Regression

MULTIPLE REGRESSION ANALYSIS OF MAIN ECONOMIC INDICATORS IN TOURISM. R, analysis of variance, Student test, multivariate analysis

Comparing Nested Models

Pearson's Correlation Tests

Pearson s Correlation

Chapter 2 Probability Topics SPSS T tests

Module 5: Multiple Regression Analysis

Section 14 Simple Linear Regression: Introduction to Least Squares Regression

Stepwise Regression. Chapter 311. Introduction. Variable Selection Procedures. Forward (Step-Up) Selection

The Dummy s Guide to Data Analysis Using SPSS

Testing for Lack of Fit

Binary Logistic Regression

How To Run Statistical Tests in Excel

Independent t- Test (Comparing Two Means)

Illustration (and the use of HLM)

Outline. Topic 4 - Analysis of Variance Approach to Regression. Partitioning Sums of Squares. Total Sum of Squares. Partitioning sums of squares

CRITICAL FACTORS AFFECTING THE UTILIZATION OF CLOUD COMPUTING Alberto Daniel Salinas Montemayor 1, Jesús Fabián Lopez 2, Jesús Cruz Álvarez 3

Comparing a Multiple Regression Model Across Groups

Moderator and Mediator Analysis

1. What is the critical value for this 95% confidence interval? CV = z.025 = invnorm(0.025) = 1.96

INTERPRETING THE ONE-WAY ANALYSIS OF VARIANCE (ANOVA)

ABSORBENCY OF PAPER TOWELS

Example: Boats and Manatees

Univariate Regression

Predictability Study of ISIP Reading and STAAR Reading: Prediction Bands. March 2014

Correlation and Regression Analysis: SPSS

ANALYSIS OF USER ACCEPTANCE OF A NETWORK MONITORING SYSTEM WITH A FOCUS ON ICT TEACHERS

Using R for Linear Regression

Factor Analysis. Principal components factor analysis. Use of extracted factors in multivariate dependency models

Didacticiel - Études de cas

Ordinal Regression. Chapter

11. Analysis of Case-control Studies Logistic Regression

An SPSS companion book. Basic Practice of Statistics

We extended the additive model in two variables to the interaction model by adding a third term to the equation.

Section Format Day Begin End Building Rm# Instructor. 001 Lecture Tue 6:45 PM 8:40 PM Silver 401 Ballerini

5. Linear Regression

Elementary Statistics Sample Exam #3

ANALYSIS OF TREND CHAPTER 5

Profile analysis is the multivariate equivalent of repeated measures or mixed ANOVA. Profile analysis is most commonly used in two cases:

Simple linear regression

ECON 142 SKETCH OF SOLUTIONS FOR APPLIED EXERCISE #2

Linear Models in STATA and ANOVA

Final Exam Practice Problem Answers

Regression and Correlation

Association Between Variables

Bill Burton Albert Einstein College of Medicine April 28, 2014 EERS: Managing the Tension Between Rigor and Resources 1

Chapter 15. Mixed Models Overview. A flexible approach to correlated data.

Regression Clustering

APPLICATION OF LINEAR REGRESSION MODEL FOR POISSON DISTRIBUTION IN FORECASTING

Lin s Concordance Correlation Coefficient

Multiple Regression: What Is It?

Projects Involving Statistics (& SPSS)

THE INFLUENCE OF MARKETING INTELLIGENCE ON PERFORMANCES OF ROMANIAN RETAILERS. Adrian MICU 1 Angela-Eliza MICU 2 Nicoleta CRISTACHE 3 Edit LUKACS 4

2. What is the general linear model to be used to model linear trend? (Write out the model) = or

An Introduction to Path Analysis. nach 3

Module 3: Correlation and Covariance

August 2012 EXAMINATIONS Solution Part I

Regression Analysis (Spring, 2000)

USING SPSS FOR DESCRIPTIVE ANALYSIS

Lets suppose we rolled a six-sided die 150 times and recorded the number of times each outcome (1-6) occured. The data is

January 26, 2009 The Faculty Center for Teaching and Learning

Indices of Model Fit STRUCTURAL EQUATION MODELING 2013

1.5 Oneway Analysis of Variance

CHAPTER 13 SIMPLE LINEAR REGRESSION. Opening Example. Simple Regression. Linear Regression

An analysis appropriate for a quantitative outcome and a single quantitative explanatory. 9.1 The model behind linear regression

A Basic Guide to Analyzing Individual Scores Data with SPSS

Multiple Regression Using SPSS

SPSS Resources. 1. See website (readings) for SPSS tutorial & Stats handout

1. The parameters to be estimated in the simple linear regression model Y=α+βx+ε ε~n(0,σ) are: a) α, β, σ b) α, β, ε c) a, b, s d) ε, 0, σ

A SUCCESFULL WAY TO SELL HANDICRAFTS IN ALBANIA

Transcription:

SPSS Guide: Regression Analysis I put this together to give you a step-by-step guide for replicating what we did in the computer lab. It should help you run the tests we covered. The best way to get familiar with these techniques is just to play around with the data and run tests. As you do it, though, think of the research questions from your project and how these tests can answer them. For the final exam, be sure you are able to: Peform a regression analysis (both bivariate and multivariate) Interpret the output o Which variables are significant? o How is an unstandardized coefficient used? Standardized? o Which is the b0 coefficient? o Which are the b coefficients? Create a regression equation from the output Plug in coefficients to predict a Y variable from the X variable(s)

In SPSS, go to Analyze>Regression>Linear Regression Enter your dependent variable (Y) in the dependent box, and your independent variable(s) in the Independent(s) box. Under the Statistics button, make sure that Estimates is checked under Regression Coefficient, and fit is checked. Click continue, and then OK on the main menu.

Your output should look like this: Entered/Removed b Entered Removed Method 1 CON3 Perceived Invasiveness of SNS Ads a. Enter a. All requested variables entered. Summary R R Square Adjusted R Square Std. Error of the Estimate 1.259 a.067.066 1.02546 a. Predictors: (Constant), CON3 Perceived Invasiveness of SNS Ads ANOVA b Sum of Squares df Mean Square F Sig. 1 Regression 96.802 1 96.802 92.055.000 a Residual 1344.953 1279 1.052 Total 1441.755 1280 a. Predictors: (Constant), CON3 Perceived Invasiveness of SNS Ads Coefficients a Unstandardized Coefficients Standardized Coefficients B Std. Error Beta t Sig. 1 (Constant) 2.297.096 23.968.000 CON3 Perceived Invasiveness of SNS Ads.268.028.259 9.595.000 a. Dependent Variable: CON8 Privacy Concern

First, look at your model summary. Your R-square tells you the goodness of fit of the model. You can think of it as a percentage. Our R-square for this model is.066, which means that the X variable can explain about 6.6% of the change in Y. The next table, ANOVA, just tells us that the model can predict Y using X. The significance is.000, so we can reject the null hypothesis that The model has no predictive value. Finally, the most important table is the coefficients. The significance level of.000 indicates that we can reject the null hypothesis that X does not predict Y. The first coefficient, (Constant), is your intercept term. That is, before you account for the dependent variable(s) or, putting it another way, when X is zero this is the value of Y. In this case, the intercept is 2.297, so when X=0, Y will equal 2.297. Remember, that regression equations are in this format: Y = b 0 + b 1 x 1 Each of the other coefficients are b variables, or the slope of the line. For each 1-unit change in X, Y will change by b units. Since we only have one variable in this case, we just have a b1 (the slope) and a x1 (the value of X). In this case, X is CON3 (Perceived invasiveness). For prediction, use the unstandardized coefficient. We can plug the coefficients into the formula to predict Y using a value of X. Y = b 0 + b 1 x 1 Y = 2.297 +.028* CON3 If X=1, then Y= 2.297 + (.028*1) = 2.325 If X=2, then Y=2.297 + (.028*2) = 2.353 etc. So when do we use the standardized coefficient? It tells us the relative size of the influence of a variable. The standardized coefficient can be interpreted like the Pearson coefficient in bivariate associative analysis (a 0-1 scale with 1 being perfectly correlated). In multivariate regression, this allows us to compare the relative influence of multiple variables, as shown in the next example where we add some more variables to this analysis. Let s use the following variables: Dependent: CON8 (Privacy concern) Independent: CON6 (Perceived information security) Q2 (Age) (F4) How many months participating in SNS CON3 (Perceived invasiveness)

Entered/Removed b Entered Removed Method 1 F4 About how many months have you been participating in SNS sites, CON6 Perceived Information Security of SNS, CON3 Perceived Invasiveness of SNS Ads, Q2 Age a. Enter a. All requested variables entered. Summary R R Square Adjusted R Square Std. Error of the Estimate 1.293 a.086.083 1.01456 a. Predictors: (Constant), F4 About how many months have you been participating in SNS sites, CON6 Perceived Information Security of SNS, CON3 Perceived Invasiveness of SNS Ads, Q2 Age Notice that our R-square has increased to.083, so we have a better model with the additional variables. ANOVA b Sum of Squares df Mean Square F Sig. 1 Regression 118.566 4 29.641 28.797.000 a Residual 1267.112 1231 1.029 Total 1385.678 1235 a. Predictors: (Constant), F4 About how many months have you been participating in SNS sites, CON6 Perceived Information Security of SNS, CON3 Perceived Invasiveness of SNS Ads, Q2 Age

The ANOVA shows that the model has predictive value, since it is significant. (Null hypothesis of no predictive value is rejected) Coefficients a Unstandardized Coefficients Standardized Coefficients B Std. Error Beta t Sig. 1 (Constant) 2.251.164 13.686.000 CON3 Perceived Invasiveness of SNS Ads.253.028.244 8.905.000 Q2 Age.010.003.081 2.850.004 CON6 Perceived Information Security of SNS F4 About how many months have you been participating in SNS sites -.095.029 -.093-3.317.001.004.002.067 2.378.018 a. Dependent Variable: CON8 Privacy Concern Notice that all of the significance levels are <.05, so they are all significant. (Reject null hypothesis that they are not associated with the dependent variable). The unstandardized coefficients can be used to create an equation for Y. b 0 = (Constant) b 1 = CON3 b 2 = Age etc. Y = b 0 + b 1 x 1 + b 2 x 2 + b 3 x 3 + b 4 x 4 Y = (2.251) + (.253)*x 1 +.010*x 2 + (-.095)*x 3 + (.004)*x 4 For this example, we ll stop here, but to calculate Y you would just plug in the data from an individual observation for each of these variables.