General Regression Formulae ) (N-2) (1 - r 2 YX



Similar documents
One-Way Analysis of Variance: A Guide to Testing Differences Between Multiple Groups

" Y. Notation and Equations for Regression Lecture 11/4. Notation:

Part 2: Analysis of Relationship Between Two Variables

Regression step-by-step using Microsoft Excel

Regression Analysis: A Complete Example

Outline. Topic 4 - Analysis of Variance Approach to Regression. Partitioning Sums of Squares. Total Sum of Squares. Partitioning sums of squares

individualdifferences

One-Way Analysis of Variance

Section 13, Part 1 ANOVA. Analysis Of Variance

Univariate Regression

Additional sources Compilation of sources:

12: Analysis of Variance. Introduction

Study Guide for the Final Exam

2. Simple Linear Regression

One-Way Analysis of Variance (ANOVA) Example Problem

5. Linear Regression

1.5 Oneway Analysis of Variance

POLYNOMIAL AND MULTIPLE REGRESSION. Polynomial regression used to fit nonlinear (e.g. curvilinear) data into a least squares linear regression model.

SPSS Guide: Regression Analysis

Elementary Statistics Sample Exam #3

Estimation of σ 2, the variance of ɛ

Chapter 7: Simple linear regression Learning Objectives

Overview Classes Logistic regression (5) 19-3 Building and applying logistic regression (6) 26-3 Generalizations of logistic regression (7)

ANOVA ANOVA. Two-Way ANOVA. One-Way ANOVA. When to use ANOVA ANOVA. Analysis of Variance. Chapter 16. A procedure for comparing more than two groups

Simple Linear Regression Inference

Chapter 7. One-way ANOVA

Example: Boats and Manatees

1. What is the critical value for this 95% confidence interval? CV = z.025 = invnorm(0.025) = 1.96

Testing for Lack of Fit

Final Exam Practice Problem Answers

This chapter will demonstrate how to perform multiple linear regression with IBM SPSS

Introduction to Analysis of Variance (ANOVA) Limitations of the t-test

IAPRI Quantitative Analysis Capacity Building Series. Multiple regression analysis & interpreting results

Week 5: Multiple Linear Regression

Week TSX Index

Testing Group Differences using T-tests, ANOVA, and Nonparametric Measures

DEPARTMENT OF PSYCHOLOGY UNIVERSITY OF LANCASTER MSC IN PSYCHOLOGICAL RESEARCH METHODS ANALYSING AND INTERPRETING DATA 2 PART 1 WEEK 9

Multivariate Analysis of Variance (MANOVA): I. Theory

KSTAT MINI-MANUAL. Decision Sciences 434 Kellogg Graduate School of Management

PSYC 381 Statistics Arlo Clark-Foos, Ph.D.

Chapter 13 Introduction to Linear Regression and Correlation Analysis

17. SIMPLE LINEAR REGRESSION II

When to use Excel. When NOT to use Excel 9/24/2014

Hypothesis testing - Steps

1. The parameters to be estimated in the simple linear regression model Y=α+βx+ε ε~n(0,σ) are: a) α, β, σ b) α, β, ε c) a, b, s d) ε, 0, σ

Descriptive Statistics

MULTIPLE LINEAR REGRESSION ANALYSIS USING MICROSOFT EXCEL. by Michael L. Orlov Chemistry Department, Oregon State University (1996)

Correlational Research

A Primer on Forecasting Business Performance

Randomized Block Analysis of Variance

1 Theory: The General Linear Model

CHAPTER 13 SIMPLE LINEAR REGRESSION. Opening Example. Simple Regression. Linear Regression

Two-sample t-tests. - Independent samples - Pooled standard devation - The equal variance assumption

MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS

Unit 31 A Hypothesis Test about Correlation and Slope in a Simple Linear Regression

Comparing Nested Models

Statistics Review PSY379

An analysis method for a quantitative outcome and two categorical explanatory variables.

Simple Linear Regression, Scatterplots, and Bivariate Correlation

Section 1: Simple Linear Regression

1 Simple Linear Regression I Least Squares Estimation

1/27/2013. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2

MULTIPLE REGRESSION WITH CATEGORICAL DATA

Curve Fitting. Before You Begin

Basic Statistics and Data Analysis for Health Researchers from Foreign Countries

EXCEL Analysis TookPak [Statistical Analysis] 1. First of all, check to make sure that the Analysis ToolPak is installed. Here is how you do it:

Rank-Based Non-Parametric Tests

DATA ANALYSIS. QEM Network HBCU-UP Fundamentals of Education Research Workshop Gerunda B. Hughes, Ph.D. Howard University

Multiple Regression in SPSS This example shows you how to perform multiple regression. The basic command is regression : linear.

Multiple Linear Regression

Recall this chart that showed how most of our course would be organized:

Notes on Applied Linear Regression

Class 19: Two Way Tables, Conditional Distributions, Chi-Square (Text: Sections 2.5; 9.1)

Copyright 2007 by Laura Schultz. All rights reserved. Page 1 of 5

Rockefeller College University at Albany

Introduction to Regression and Data Analysis

Part II. Multiple Linear Regression

1 Basic ANOVA concepts

GLM I An Introduction to Generalized Linear Models

Multiple Regression. Page 24

LOGISTIC REGRESSION ANALYSIS

NCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( )

Econometrics Simple Linear Regression

UNDERSTANDING THE TWO-WAY ANOVA

Linear Models for Continuous Data

One-Way ANOVA using SPSS SPSS ANOVA procedures found in the Compare Means analyses. Specifically, we demonstrate

STATISTICS FOR PSYCHOLOGISTS

INTERPRETING THE ONE-WAY ANALYSIS OF VARIANCE (ANOVA)

1.1. Simple Regression in Excel (Excel 2010).

Calculating P-Values. Parkland College. Isela Guerra Parkland College. Recommended Citation

Regression III: Advanced Methods

Chapter 9. Two-Sample Tests. Effect Sizes and Power Paired t Test Calculation

MULTIPLE REGRESSION ANALYSIS OF MAIN ECONOMIC INDICATORS IN TOURISM. R, analysis of variance, Student test, multivariate analysis

Calculating, Interpreting, and Reporting Estimates of Effect Size (Magnitude of an Effect or the Strength of a Relationship)

Chapter 14: Repeated Measures Analysis of Variance (ANOVA)

The Statistics Tutor s Quick Guide to

Business Statistics. Successful completion of Introductory and/or Intermediate Algebra courses is recommended before taking Business Statistics.

Transcription:

General Regression Formulae Single Predictor Standardized Parameter Model: Z Yi = β Z Xi + ε i Single Predictor Standardized Statistical Model: Z Yi = β Z Xi Estimate of Beta (Beta-hat: β = r YX (1 Standard error of estimate: s Zy.Zx = ( Standard error of Beta: Se β = (1 - r YX (N- (3 There are two identical null hypotheses: Ho: β = 0 and Ho: ρ = 0 Both are tested with a t-statistic with (N - degrees of freedom (df which can be computed two ways. t (N - = r (N - (4 and t (N - = β - 0 Se β (5 Single Predictor Raw Score Parameter Model: Y i = α + βx i + ε i Single Predictor Raw Score Statistical Model: Y = a + b 1 X 1 Estimate of Beta (b: b=β s Y sx (6 Since Beta-hat and r are identical in the single predictor model r can be substituted. Estimate of the Y-intercept or Regression constant (a: a = Y - bx (7 Standard error of estimate: s Y.X = s Y (8 Standard error of b: Se b = s Y.X (N-1s X (9 There are two identical null hypotheses: Ho: β = 0 and Ho: ρ = 0 Both are tested with a t-statistic with (N - degrees of freedom (df which can be computed two ways. Again with formula (4 and with t (N - = b - 0 Se b (10

Two Predictor Standardized Parameter Model: Z Yi = β 1 Z X1i + β Z Xi + ε i Two Predictor Standardized Statistical Model: Z Yi = β 1 Z X1i + β Z Xi To calculate Beta-hat the correlation between the predictor variables must be taken into consideration β 1 = r Y1 - r Y r 1 1 - r 1 (11 and β = r Y - r Y1 r 1 1 - r 1 (1 Similar to formula (, the Standard error of estimate is: s Zy.Zy = 1 - R YX (13 In the two predictor case the Standard error of Beta-hat is the same for both variables: Se β = (1 - R Y.1 (N-3(1- r 1 (14 However, there is more than one null hypothesis that can be tested. First of all, one can test whether the overall model significantly improves prediction over the mean. Ho: β 1 = β = 0 This is tested with an F-statistic with two (number of predictors and (N-3 dfs: (R F (,N-3 = - 0/ (1 - R /(N - 3 (15 Multiple R has a general formula: K R Y = βj r Yj j =1 = β 1 r Y1 + β r Y +... + β k r Yk (16 One may also test whether each predictor makes a significant improvement in prediction over the other predictor(s. This is tested with a t-test with (N - k -1 degrees of freedom, where k equals the number of predictors (in this case k =. For any variable j: t (N - k -1 = β j - 0 Se βj (17 where Se βj = (1 - R Y.1... k (N-k-1(1- R j.1... k (18 This can also be tested with a more flexible F-statistic: F (kf - k R,N-k F -1 = (R F - RR/(kF - k R (1 - R F/(N - kf -1 (19

Two Predictor Raw Score Parameter Model: Y i = α + β 1 X 1i + β X i + ε i Two Predictor Raw Score Statistical Model: Y = a + b 1 X 1 + b X s For any variable j, the Estimate of Beta (bj: b j =β Y jsxj (0 Estimate of the Y-intercept or Regression constant (a: a = Y - b 1 X 1 - b X (1 Similar to formula (8, the Standard error of estimate: s Y.Y = s Y 1 - R ( Because of possible differences in variance across variables, each predictor variable has a different Standard error of b: s Y.Y (3 For any of the two variables denoted as j: Sebj = s j (N-1(1- r 1 Again, one can test whether the overall model significantly improves prediction over the mean. Ho: β 1 = β = 0, which is tested with the F-statistic in formula (15. Also similar to the standardized model, one may also test whether each predictor makes a significant improvement in prediction over the other predictor(s. This is tested with a t-test with (N - k -1 degrees of freedom, where k equals the number of predictors (in this case k =. For any variable j: t (N - k -1 = b j - 0 Se bj (4 Partial Correlations are used to statistically "control" the effects of all other predictors. Partial correlations remove the effect of control variables from variables of interest including the dependent variable. Some researchers use them instead of Beta-hat to interpret variable "importance." With one dependent variable (Y and two predictors, the general formula is: r Y1. = r Y1 - r Y r 1 1 - r 1 1 - r Y (5 Semi-Partial (sometimes referred to as Part correlation are an index of the "unique" correlation between variables. Semi-Partial correlations remove the effect of a variable from all other predictors but not the dependent variable. With one dependent variable (Y and predictors, the general formula is: r Y(1. = r Y1 - r Y r 1 1 - r 1 (6 Squaring Semi-partial correlations are useful because they give the "unique" contribution a variable makes to the R of a multiple regression model. For example with two predictors R can be decomposed as follows:

R Y.1 = r Y + r Y(1. and conversely, R Y.1 = r Y1 + r Y(.1

Source Table for Multiple Regression Although this process would be laborious, this is the conceptual derivation for the F-ratio in Multiple Regression Source Sum of Squares df Mean Squares F Regression (Y i - Y k SS R /k MS R /MS e (Explained Variance Residual (Y i - Y i N - k - 1 SS e /df e Total Variance (Y i - Y N - 1 s = SS T /N-1 where, N = total number of cases, k = number of predictors, Y = the mean of Y. Y i = each individual score on Y, and Y i = each individual predicted Y. Given, R = SS R /SS T Source Sum of Squares df MS F Regression R SS T k SS R /k (R /k (Explained Variance (1- R /(N - k - 1 Residual (1- R SS T N - k - 1 SS e /N - k - 1 Total Variance (Y i - Y N - 1 s = SS T /N-1 One-Way ANOVA Source Table When we extend Least Squares Regression Methodology to a continuous dependent variable Y and categorical independent variables, it is often referred to as the ANalysis Of Varaince (ANOVA. In the ANOVA, the predicted score, Y i, for each individual in the jth group is equal to their (jth group mean, Y i = Y j. Knowing this, the previous Source Tables simplify greatly. For the One-way (one categorical independent variable ANOVA, the Source Table is as follows: Source Sum of Squares df Mean Squares F Between Groups n j (Y j - Y * J - 1 SS B /J - 1 MS B /MS W (Explained Variance Within Groups Total Variance (Y i - Y j N - J SS W /df W (Y i - Y * N - 1 s = SS T /N-1 where, N = total number of cases, J = number of groups, Y * = the grand mean of Y across all groups. Y i = each individual score on Y, and Y j = the mean for group j. n j = the number of cases in group j. R = η = SS B /SS T.