Binary Logistic Regression



Similar documents
Chapter Seven. Multiple regression An introduction to multiple regression Performing a multiple regression on SPSS

SPSS Guide: Regression Analysis

LOGISTIC REGRESSION ANALYSIS

HYPOTHESIS TESTING: CONFIDENCE INTERVALS, T-TESTS, ANOVAS, AND REGRESSION

Ordinal Regression. Chapter

STATISTICA Formula Guide: Logistic Regression. Table of Contents

Directions for using SPSS

Examining Differences (Comparing Groups) using SPSS Inferential statistics (Part I) Dwayne Devonish

Students' Opinion about Universities: The Faculty of Economics and Political Science (Case Study)

Chapter 5 Analysis of variance SPSS Analysis of variance

Credit Risk Analysis Using Logistic Regression Modeling

Multiple Regression in SPSS This example shows you how to perform multiple regression. The basic command is regression : linear.

11. Analysis of Case-control Studies Logistic Regression

Overview Classes Logistic regression (5) 19-3 Building and applying logistic regression (6) 26-3 Generalizations of logistic regression (7)

This chapter will demonstrate how to perform multiple linear regression with IBM SPSS

Descriptive Statistics

Logistic Regression.

Multiple logistic regression analysis of cigarette use among high school students

Chapter 15. Mixed Models Overview. A flexible approach to correlated data.

Comparing a Multiple Regression Model Across Groups

Additional sources Compilation of sources:

Statistics in Retail Finance. Chapter 2: Statistical models of default

SUGI 29 Statistics and Data Analysis

A Study to Predict No Show Probability for a Scheduled Appointment at Free Health Clinic

How to set the main menu of STATA to default factory settings standards

SPSS Notes (SPSS version 15.0)

Finding Supporters. Political Predictive Analytics Using Logistic Regression. Multivariate Solutions

ASSIGNMENT 4 PREDICTIVE MODELING AND GAINS CHARTS

Testing and Interpreting Interactions in Regression In a Nutshell

Data Mining Part 5. Prediction

1/27/2013. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2

I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN

Is it statistically significant? The chi-square test

SPSS Explore procedure

Methods for Interaction Detection in Predictive Modeling Using SAS Doug Thompson, PhD, Blue Cross Blue Shield of IL, NM, OK & TX, Chicago, IL

Analyzing Intervention Effects: Multilevel & Other Approaches. Simplest Intervention Design. Better Design: Have Pretest

The Dummy s Guide to Data Analysis Using SPSS

EARLY VS. LATE ENROLLERS: DOES ENROLLMENT PROCRASTINATION AFFECT ACADEMIC SUCCESS?

Illustration (and the use of HLM)

Linear Models in STATA and ANOVA

Module 5: Introduction to Multilevel Modelling SPSS Practicals Chris Charlton 1 Centre for Multilevel Modelling

MULTIPLE REGRESSION WITH CATEGORICAL DATA

Applied Data Mining Analysis: A Step-by-Step Introduction Using Real-World Data Sets

Stepwise Regression. Chapter 311. Introduction. Variable Selection Procedures. Forward (Step-Up) Selection

Correlates of Academic Achievement for Master of Education Students at Open University Malaysia

Simple Linear Regression, Scatterplots, and Bivariate Correlation

SAS Software to Fit the Generalized Linear Model

This presentation was made at the California Association for Institutional Research Conference on November 19, 2010.

Module 4 - Multiple Logistic Regression

How to Get More Value from Your Survey Data

Εισαγωγή στην πολυεπίπεδη μοντελοποίηση δεδομένων με το HLM. Βασίλης Παυλόπουλος Τμήμα Ψυχολογίας, Πανεπιστήμιο Αθηνών

Multivariate Models of Student Success

II. DISTRIBUTIONS distribution normal distribution. standard scores

Free Trial - BIRT Analytics - IAAs

Data analysis process

SPSS ADVANCED ANALYSIS WENDIANN SETHI SPRING 2011

HLM software has been one of the leading statistical packages for hierarchical

Using An Ordered Logistic Regression Model with SAS Vartanian: SW 541

PROC LOGISTIC: Traps for the unwary Peter L. Flom, Independent statistical consultant, New York, NY

January 26, 2009 The Faculty Center for Teaching and Learning

Multiple Regression. Page 24

Introduction to Longitudinal Data Analysis

4. Multiple Regression in Practice

IBM SPSS Regression 20

Answer: C. The strength of a correlation does not change if units change by a linear transformation such as: Fahrenheit = 32 + (5/9) * Centigrade

The Chi-Square Test. STAT E-50 Introduction to Statistics

i SPSS Regression 17.0

ln(p/(1-p)) = α +β*age35plus, where p is the probability or odds of drinking

The first three steps in a logistic regression analysis with examples in IBM SPSS. Steve Simon P.Mean Consulting

SCHOOL OF HEALTH AND HUMAN SCIENCES DON T FORGET TO RECODE YOUR MISSING VALUES

Weight of Evidence Module

DEPARTMENT OF PSYCHOLOGY UNIVERSITY OF LANCASTER MSC IN PSYCHOLOGICAL RESEARCH METHODS ANALYSING AND INTERPRETING DATA 2 PART 1 WEEK 9

Gender Differences in Employed Job Search Lindsey Bowen and Jennifer Doyle, Furman University

S P S S Statistical Package for the Social Sciences

SPSS and AM statistical software example.

Step by Step Guide to Importing Genetic Data into JMP Genomics

IBM SPSS Statistics 20 Part 4: Chi-Square and ANOVA

Chapter 13 Introduction to Linear Regression and Correlation Analysis

SPSS TRAINING SESSION 3 ADVANCED TOPICS (PASW STATISTICS 17.0) Sun Li Centre for Academic Computing lsun@smu.edu.sg

Counting the Ways to Count in SAS. Imelda C. Go, South Carolina Department of Education, Columbia, SC

Data Mining: An Overview of Methods and Technologies for Increasing Profits in Direct Marketing. C. Olivia Rud, VP, Fleet Bank

MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS

Journal of Business & Economics Research November, 2004 Volume 2, Number 11

Doing Multiple Regression with SPSS. In this case, we are interested in the Analyze options so we choose that menu. If gives us a number of choices:

Introduction to Multilevel Modeling Using HLM 6. By ATS Statistical Consulting Group

Modeling Lifetime Value in the Insurance Industry

Developing Risk Adjustment Techniques Using the System for Assessing Health Care Quality in the

Data Mining III: Numeric Estimation

Multinomial and Ordinal Logistic Regression

MORTGAGE LENDER PROTECTION UNDER INSURANCE ARRANGEMENTS Irina Genriha Latvian University, Tel ,

Cluster analysis with SPSS: K-Means Cluster Analysis

SPSS Introduction. Yi Li

Co-Curricular Activities and Academic Performance -A Study of the Student Leadership Initiative Programs. Office of Institutional Research

Transcription:

Binary Logistic Regression Main Effects Model Logistic regression will accept quantitative, binary or categorical predictors and will code the latter two in various ways. Here s a simple model including a selection of variable types -- the criterion variable is vs. nonly aged college students and the predictors are gender, marital status, loneliness and stress. Analyze Regression Binary Logistic Move the criterion variable into the Dependent: window Move the predictors into the Covariate window Notice that you can select a subset of cases from here which gives the same result as using Data Select Cases. Click Categorical to specify the categorical variables. Specify the entry method -- here Enter means to add all variables to the model simultaneously. Both blockwise entry (adding all predictors in that block) and blockwise selection (selecting best predictors from that block and entering them one at a time) are possible Move each categorical predictor into the Categorical Covariates: window. The default coding is dummy codes ( Indicator coding ) with the highest coded group as the comparison group ( reference group ). I ve changed the comparison group for marital3 to be the lowest coded group () Highlight the categorical variable you want to specify, then choose whether the comparison group should be the First or the Last then click Change. Notice the specification of each predictor is given in the Categorical Covariates: window.

The SPSS output specifies the coding, etc. in the first part of the output. Dependent Variable Encoding Original Value non Internal Value 0 The coding for the criterion variable is given first -- The largest coded group is identified as the target GENDER Categorical Variables Codings single married other male female Parameter coding Frequency () (2) 242.000.000 2.000.000 42.000.000 80.000 225.000 The coding for all categorical predictors is specified. Marital3 Single will be the comparison group st parameter will compare married to single 2 nd parameter will compare other to single Gender female will be the comparison group the parameter will compare male to female 0 a. is included in the model. b. The cut value is.500 0 non Classification Table a,b non Correct 204 0 00.0 20 0.0 -.05.099.022.88.985 50.4 Block 0 The process is inherently stepwise -- for forming and testing nested hierarchical models. The first step is to compute and enter just the constant -- even if you ve specified only a single block of variables, as in this case. The classification table tells the # and % or cases correctly classified by the model. Regression weights and a test of the H0: b = 0 for the variables in the equation (only the constant for Block 0) is provided. 0 Variables Overall Statistics Variables not in the Equation () (2) Score df Sig. 5.38.023 263.57 2.000 63.823.000 47.559.000 27.890.000 8.92.003 277.780 5.000 The contribution of each predictor were it added alone into the equation on the next step is foretold.

Omnibus Tests of Model Coefficients Block Model Chi-square df Sig. Model Summary -2 Log Cox & Snell Nagelkerke likelihood R Square R Square 200.468.590.786 Block -- tests the contribution of the specific variable(s) entered on this step Block -- tests the contribution of all the variables entered with this block Model -- tests the fit of the whole model These are all the same for a model with a single set of predictors that are entered simultaneously. 2 R² values are presented to estimate the fit of the model to the data -- both are transformations of the -2log likelihood values. Classification Table a The reclassification table shows the accuracy of the model. a. The cut value is.500 non non Correct 98 6 97. 30 7 85. 9. If you are going to do classification predictions, asymmetries and/or accuracy can sometimes be improved by adjusting the cutoff value from the default of.5 (in the options window). a () (2) -.08.370.048.826.922 6.774 2.000 5.974.760 6.774.000 393.029 23.064 5793.703.000.997.0E+0.09.07 29.943.000.095 -.072.028 6.448.0.930-4.480.706 40.255.000.0 a. Variable(s) entered on step : GENDER,,,. Interpreting the model: Gender does not contribute to the model the negative B indicates that the target group (non) tends to have more of those coded 0 (females) than of these coded (males) - but not significantly, after controlling for the other predictors Marital does contribute to the model there is a test of the overall contribution of multi-category variables, as well as a test of the contribution of each parameter the positive B for married vs. single indicates more married () in the non () group, after controlling other vs. single does not contribute to the model, after controlling for the other predictors Loneliness does contribute to the model, with higher average loneliness for the non () group, after controlling.. Stress does contribute to the model, with lower average stress for the non () group, after controlling

A second step added the interactions of gender with loneliness and stress. Click Next to specify what will be entered on the second Block To inter an interaction highlight both terms of the interaction (by holding down Ctrl as you click each predictor) and then click the >a*b> button The interaction will be calculated as the product of the terms, or the term parameters in the case of categorical variables (following the specifications given earlier) Omnibus Tests of Model Coefficients Block Model Chi-square df Sig. 4.58 2.000 4.58 2.000 375.7 7.000 Model Summary -2 Log Cox & Snell Nagelkerke likelihood R Square R Square 86.23.634.89 Again and Block are the same because of the blockwise entry of the two interaction terms. The and Block X² tests tell us that the model was improved by the inclusion of these terms. Remember that tests the average contribution of the included terms and so you one or more of multiple included terms can be significant without a significant improvement to the model. As would be expected, the fit of the model improved, according to both of the R² calculations. Classification Table a a. The cut value is.500 non non Correct 20 3 98.5 2 89 94.0 96.3 The % correct reclassification also improved, especially for the non group.

a () (2) by by a. Variable(s) entered on step : GENDER *, GENDER *..259.325.038.845.296 6.736 2.000 5.983.762 6.736.000 396.78 23.072 5800.267.000.997.0E+0.093.023 6.509.000.098 -.064.038 2.803.094.938 -.005.033.022.883.995-2.79.057 4.099.000.882-4.64.933 24.742.000.00 Interpreting the model: Gender does not contribute to the model including the interactions either (though always check -- contributions can change with the adding or deleting of predictors -- as the colinearity mix changes!) Marital does contribute to the model -- with the same pattern as in the main effects model Loneliness does contribute to the model, with higher average loneliness for the non () group, after controlling.. Stress does not contribute to the model -- though it contributed to the main effect model The Gender * Loneliness interaction does not contribute to the model The Gender * Stress interaction does contribute to the model the negative weight for this interaction means the slope of the relationship between group and stress is less positive for the males () and more positive for the females (0)