Chapter 9. Multiple Linear Regression Analysis
|
|
- Esmond Maximillian Leonard
- 7 years ago
- Views:
Transcription
1 Chapter 9 Multiple Linear Regression Analysis In Chapter 8, we studied simple linear regression. In simple regression we deal with one independent variable and one dependent variable. In this chapter, we will study multiple regression in which we deal with more than one independent variable. We still only have one dependent variable. As an example, think about sales revenue as a dependent variable. It depends not just on advertising expense but other factors as well such as the product price, product quality, price of competing products, quality of competing products, brand loyalty etc. When we try to predict sales revenue using several factors (or variables) we make use of multiple regression models. This is just one example. In general, when we try to predict any dependent variable using several factors, we make use of multiple regression. Scatter Plots: Remember in simple regression we first drew a scatterplot of the data to give us a visual representation of the relationship between two variables? When we have two independent variables and one dependent variable, the scatterplot becomes a three-dimensional graph. The dependent variable is on the vertical axis and the two independent variables are on the two horizontal axes at the base. You can imagine a cube with a swarm of points in the space within the cube. But when we have three or more independent variables and a dependent variable, we cannot imagine a scatterplot because we would have to imagine a four-dimensional or a higher dimensional box, something you and I as human beings simply cannot do. But it is not absolutely necessary to have scatterplots. If you can draw scatterplots, it s great, but if we cannot, it s no big deal. We will not get a visual impression of the relationship, but we can still get the mathematical representation of the relationship using multiple regression techniques. The Population Model: Remember in simple regression we had a population model that looked like this: y = α + βx + ε. In multiple regression since there are several independent variables, the model is a little bit more complex. Suppose there are k independent variables x 1, x 2,, x k. The population model will look like this: y = α + β 1 x 1 + β 2 x β k.x k + ε The Expected value of y is given by E(y) = α + β 1 x 1 + β 2 x β k.x k Just like in simple regression, these models have the y-intercept (α), but instead of one slope coefficient, we have k slope coefficients, one corresponding to each of the k independent variables. Our job is to estimate α and the k slope coefficients. We do this using sample data. The Sample Model: The sample model looks like this:, where a, b 1, b 2, b k are estimates of α, β 1, β 2, β k respectively. As in simple regression, a is called the y-intercept and b 1, b 2, b k are the slope coefficients of the k independent variables. Interpretation of slope coefficients: The interpretation of the slope coefficients in multiple regression is similar to the interpretation in simple regression with one important difference. The difference is that we have to mention that other independent variables are constant. I will explain the difference with the help of an example. Here is how we would interpret b 1 : As x 1 increases by one unit, the estimated value of y increases by an average of b 1 units while other variables stay constant. It is important to mention while other variables stay constant. Only stating that y increases by an average of b 1 as x 1 increases by one unit would be incomplete. How to perform Multiple Regression in Excel? If you know how to perform Simple Regression in Excel (which you do), you already know how to perform multiple regression. The only difference is that when specifying the X-range, you specify multiple columns 1
2 instead of one column of data. The output also looks very much like that of a simple regression. The only difference is that in the third panel, you will see extra rows. In simple regression, you see two rows, one for the y-intercept and one for the independent variable. In multiple regression, if you have k independent variables, you will see k+1 rows. One row for the y-intercept and k rows for the k independent variables. Let s look at an example: Adv. Expense ('000s) Price Sales (millions) Figure 1: Data on Sales, Adv. Expense and Price Say we have collected some sample data on Sales, Advertising Expense and Prices as shown in Figure 1. We want to predict Sales using two independent variables Adv. Expense and Price. As you might guess, as price increases, sales will tend to decrease, and as advertising expense increases, sales will tend to increase. So we want to predict sales for a particular price and a particular level of advertising expense. For this type of problem, we use multiple regression. In this example, we have one dependent variable (Sales) and two independent variables (Adv. Expense and Price). Anytime we have two or more independent variable, we use multiple regression. Luckily, in Excel, it is just as easy to run multiple regression as it is to run simple regression. In fact, Excel does not even distinguish between the two. It just calls it Regression. If we select only one column of data for the X-Range, then we are running simple regression. If we select two or more columns of data for the X-Range, then we are running multiple regression. Let s look at the regression output of the above data, shown in Figure 2. Figure 2: Regression Output using Excel for data in Figure 1. 2
3 Using the regression output in Figure 2, we can write the sample model as follows: y-hat = X X 2, where y-hat is the estimated average Sales (in millions of dollars) and X 1 is the Advertising expense (in thousands of dollars) and X 2 is Price (in dollars). Please note the coefficient values. The coefficient of X 1 is How do we interpret this coefficient? For every extra one thousand dollars spent in advertising will increase sales by million or 7.11 thousand dollars at a given price level. The coefficient of X 2 is How to interpret this coefficient? For every extra one dollar increase in price, the total sales will decrease by million at a given level of advertising expense. Please note that since the sign of the coefficient for the price variable is negative, we should be careful in saying that an increase in the price will decrease sales. It is also important to state that the advertising expense stays constant at some level. Notice that in the regression output, the third panel had three rows of information - the first row for the y- intercept and two for the two independent variables. R-Square in Multiple Regression: The R-Square value has the same interpretation in multiple regression as in simple regression i.e. it represents the proportion of variation in the dependent variable that is explained by the independent variables. As we add more independent variables, the R-Square always gets better, i.e. adding an extra independent variable can only increase the proportion of the variation explained by the independent variables. Can R-Square ever decrease by adding an extra independent variable? No. R-Square can never decrease by adding an extra independent variable. Even if an independent variable has no correlation whatsoever with the dependent variable, even then the R-square value will not decrease by including this variable. It will only go up, even if very slightly. It is important to understand the consequences of this phenomenon. Because the R-square value keeps increasing as you keep adding independent variables and since the R-square value is often used as a benchmark to evaluate a model s explanatory power, there is a tendency for model builders to keep adding more independent variables in an effort to improve the R-square value. But this practice is not a very good one from the decision maker s point of view. To resolve this problem, statisticians have invented another measure called the Adjusted R-square. The Adjusted R-Square The adjusted R-square is adjusted for the number of independent variables. So if an unrelated independent variable is added to a model, it might increase the R-square value a little bit but it will decrease the adjusted R-square value. It is for this reason that when evaluating a multiple regression model that we look at the adjusted R-square value instead of the R-square value. In our example, the R-square value is and the adjusted R-square is The adjusted R-square value will always be smaller than the R-square value. Hypotheses for the Regression Coefficients In regression, we are interested in the statistical significance of the slope coefficients of the variables. We test the significance of the coefficients using the hypothesis testing procedure we have learnt in previous chapters. Let me explain this procedure with the help of the following examples: 3
4 Example 1: Test whether the variable advertising expense is significant. Null Hypothesis: H 0 : β 1 = 0 Alt. Hypothesis: H a : β 1 0 The Test Statistic: t-value = b 1 /s b1 = / = Critical t-value: two tailed t value for alpha of 0.05 for degree of freedom 5 = +/ The deg. Of freedom is n-3 = 8-3 = 5. If there are k independent variables, the degree of freedom is n (k+1). The p-value: corresponding to the test statistic is Decision: Since the test statistic is less than the critical value and correspondingly, since the p-value is larger than 0.05, we fail to reject the null hypothesis. Conclusion: At a significance level of 0.05, there is not sufficient evidence that the variable advertising expense is making any difference on Sales. Example 2: Test whether the variable Price is significant Null Hypothesis: H 0 : β 2 = 0 Alt. Hypothesis: H a : β 2 0 The Test Statistic: t-value = b 2 /s b2 = / = Critical t-value: two tailed t value for alpha of 0.05 for degree of freedom 5 = +/ The p-value: corresponding to the test statistic is Decision: Since the p-value is much smaller than 0.05, we reject the null hypothesis. Note that the t-statistic value is which is in the rejection region because the rejection region is below and above Conclusion: At a significance level of 0.05, there is sufficient evidence that the variable price is making a difference on Sales. An important insight: In the above two examples, we found one variable (price) to be significant while the other variable (advertising expense) to be not significant. I want you to note that for the significant variable, in this case, price, both the lower and the upper confidence limits have the same sign (in this case, negative). For the nonsignificant variable, in this case, advertising expense, the lower confidence limit is negative while the upper confidence limit is positive. You will always see that for non-significant variables, the lower and the upper confidence limits will have different signs. Because the lower limit is negative and the upper limit is positive, zero lies within the confidence interval. It implies that slope could be zero, which implies that the independent variable has no effect on the dependent variable. Which is why we say that the slope is not significant. 4
5 Hypotheses for the Overall Model: Besides performing hypotheses testing for individual variables, we can also perform hypothesis testing for testing the significance of the overall model. Let me explain with the help of an example: Null Hypothesis: H 0 : β 1 = β 2 = 0 Alt. Hypothesis: H a : At least one of the βs is not zero The Test Statistic: F-value = The p-value: corresponding to the test statistic is Decision: Since the since the p-value is smaller than 0.05, we reject the null hypothesis. Conclusion: At a significance level of 0.05, there is sufficient evidence that the model is significant i.e., advertising expense and price together have an effect on Sales. Note: For testing the significance of the overall model, the test statistic used is the F-statistic. For testing the significance of individual coefficients, the test statistic used is the t-statistic. You may recall that in ANOVA, we test hypotheses about several means. In ANOVA we use the F statistic. Similarly here, for the overall model, we are testing the significance of several betas and therefore we use the F statistic found in the ANOVA panel of the Regression output. Which Independent Variable is more significant? Whenever we run a multiple regression model involving several independent variables, an obvious question is which one of the many independent variables is the most statistically significant. The answer is rather straightforward. You basically compare the p-values of the coefficients of each variable. The variable whose slope coefficient has the least p-value is the most statistically significant. In our example, the p-value for Advertising expense is whereas the p-value for Price is Clearly Price is the more significant variable here. Please note that the absolute value of the slope coefficient does not tell us anything about the significance of the variable. Practical vs. Statistical Significance If the magnitude of the coefficient of the variable is very small, it implies that the dependent variable does not increase (or decrease) much if we increase the independent variable by one unit. Therefore there may be a tendency to treat that variable as insignificant from a practical point of view. However, one has to be cautious. Even if the magnitude of the coefficient is small, it may be practically significant because the units of the dependent variable may be a large unit. In our example, the unit of our dependent variable (Sales) is millions of dollars. So even though the magnitude of the coefficient is rather small (0.0071), since it is in millions, if we convert it to dollars, the coefficient value is really 7,112, which is fairly large and therefore practically significant because spending an extra one thousand dollars in advertising expense will increase sales by 7,112. But suppose the unit of the dependent variable is small and the magnitude of the coefficient is small, then there may not be a high practical significance for this variable even but as long as its p-value is smaller than the significance level (alpha) the coefficient is considered statistically significant. 5
6 Summary of the Chapter 1. In Multiple Regression, we have one dependent variable and two or more independent variables. 2. In Linear Multiple Regression, the population model is: y = α + β 1 x 1 + β 2 x β k.x k + ε 3. The sample regression model is: 4. In multiple regression, the Adjusted R-square is a better measure for the coefficient of determination than R-square because the adjusted R-square penalizes addition of more explanatory variable. 5. The independent variable with the least p-value is considered the most significant variable. 6. A high F-value of a model signifies a significant model. 7. A variable may be statistically very significant, but in practice, it may or may not have a high coefficient value to be of any practical use. 6
Session 7 Bivariate Data and Analysis
Session 7 Bivariate Data and Analysis Key Terms for This Session Previously Introduced mean standard deviation New in This Session association bivariate analysis contingency table co-variation least squares
More information1. What is the critical value for this 95% confidence interval? CV = z.025 = invnorm(0.025) = 1.96
1 Final Review 2 Review 2.1 CI 1-propZint Scenario 1 A TV manufacturer claims in its warranty brochure that in the past not more than 10 percent of its TV sets needed any repair during the first two years
More informationCourse Objective This course is designed to give you a basic understanding of how to run regressions in SPSS.
SPSS Regressions Social Science Research Lab American University, Washington, D.C. Web. www.american.edu/provost/ctrl/pclabs.cfm Tel. x3862 Email. SSRL@American.edu Course Objective This course is designed
More informationSimple linear regression
Simple linear regression Introduction Simple linear regression is a statistical method for obtaining a formula to predict values of one variable from another where there is a causal relationship between
More information" Y. Notation and Equations for Regression Lecture 11/4. Notation:
Notation: Notation and Equations for Regression Lecture 11/4 m: The number of predictor variables in a regression Xi: One of multiple predictor variables. The subscript i represents any number from 1 through
More information1) Write the following as an algebraic expression using x as the variable: Triple a number subtracted from the number
1) Write the following as an algebraic expression using x as the variable: Triple a number subtracted from the number A. 3(x - x) B. x 3 x C. 3x - x D. x - 3x 2) Write the following as an algebraic expression
More informationSimple Regression Theory II 2010 Samuel L. Baker
SIMPLE REGRESSION THEORY II 1 Simple Regression Theory II 2010 Samuel L. Baker Assessing how good the regression equation is likely to be Assignment 1A gets into drawing inferences about how close the
More informationSimple Linear Regression, Scatterplots, and Bivariate Correlation
1 Simple Linear Regression, Scatterplots, and Bivariate Correlation This section covers procedures for testing the association between two continuous variables using the SPSS Regression and Correlate analyses.
More informationChapter 10. Key Ideas Correlation, Correlation Coefficient (r),
Chapter 0 Key Ideas Correlation, Correlation Coefficient (r), Section 0-: Overview We have already explored the basics of describing single variable data sets. However, when two quantitative variables
More informationHypothesis testing - Steps
Hypothesis testing - Steps Steps to do a two-tailed test of the hypothesis that β 1 0: 1. Set up the hypotheses: H 0 : β 1 = 0 H a : β 1 0. 2. Compute the test statistic: t = b 1 0 Std. error of b 1 =
More informationClass 19: Two Way Tables, Conditional Distributions, Chi-Square (Text: Sections 2.5; 9.1)
Spring 204 Class 9: Two Way Tables, Conditional Distributions, Chi-Square (Text: Sections 2.5; 9.) Big Picture: More than Two Samples In Chapter 7: We looked at quantitative variables and compared the
More informationWe are often interested in the relationship between two variables. Do people with more years of full-time education earn higher salaries?
Statistics: Correlation Richard Buxton. 2008. 1 Introduction We are often interested in the relationship between two variables. Do people with more years of full-time education earn higher salaries? Do
More informationLinear Models in STATA and ANOVA
Session 4 Linear Models in STATA and ANOVA Page Strengths of Linear Relationships 4-2 A Note on Non-Linear Relationships 4-4 Multiple Linear Regression 4-5 Removal of Variables 4-8 Independent Samples
More informationThe correlation coefficient
The correlation coefficient Clinical Biostatistics The correlation coefficient Martin Bland Correlation coefficients are used to measure the of the relationship or association between two quantitative
More informationChapter 7: Simple linear regression Learning Objectives
Chapter 7: Simple linear regression Learning Objectives Reading: Section 7.1 of OpenIntro Statistics Video: Correlation vs. causation, YouTube (2:19) Video: Intro to Linear Regression, YouTube (5:18) -
More informationModule 3: Correlation and Covariance
Using Statistical Data to Make Decisions Module 3: Correlation and Covariance Tom Ilvento Dr. Mugdim Pašiƒ University of Delaware Sarajevo Graduate School of Business O ften our interest in data analysis
More informationPart 2: Analysis of Relationship Between Two Variables
Part 2: Analysis of Relationship Between Two Variables Linear Regression Linear correlation Significance Tests Multiple regression Linear Regression Y = a X + b Dependent Variable Independent Variable
More informationUnit 31 A Hypothesis Test about Correlation and Slope in a Simple Linear Regression
Unit 31 A Hypothesis Test about Correlation and Slope in a Simple Linear Regression Objectives: To perform a hypothesis test concerning the slope of a least squares line To recognize that testing for a
More informationFinal Exam Practice Problem Answers
Final Exam Practice Problem Answers The following data set consists of data gathered from 77 popular breakfast cereals. The variables in the data set are as follows: Brand: The brand name of the cereal
More information2. Simple Linear Regression
Research methods - II 3 2. Simple Linear Regression Simple linear regression is a technique in parametric statistics that is commonly used for analyzing mean response of a variable Y which changes according
More informationHow To Run Statistical Tests in Excel
How To Run Statistical Tests in Excel Microsoft Excel is your best tool for storing and manipulating data, calculating basic descriptive statistics such as means and standard deviations, and conducting
More informationPremaster Statistics Tutorial 4 Full solutions
Premaster Statistics Tutorial 4 Full solutions Regression analysis Q1 (based on Doane & Seward, 4/E, 12.7) a. Interpret the slope of the fitted regression = 125,000 + 150. b. What is the prediction for
More informationRegression Analysis: A Complete Example
Regression Analysis: A Complete Example This section works out an example that includes all the topics we have discussed so far in this chapter. A complete example of regression analysis. PhotoDisc, Inc./Getty
More informationThe Point-Slope Form
7. The Point-Slope Form 7. OBJECTIVES 1. Given a point and a slope, find the graph of a line. Given a point and the slope, find the equation of a line. Given two points, find the equation of a line y Slope
More informationModule 5: Multiple Regression Analysis
Using Statistical Data Using to Make Statistical Decisions: Data Multiple to Make Regression Decisions Analysis Page 1 Module 5: Multiple Regression Analysis Tom Ilvento, University of Delaware, College
More information2013 MBA Jump Start Program. Statistics Module Part 3
2013 MBA Jump Start Program Module 1: Statistics Thomas Gilbert Part 3 Statistics Module Part 3 Hypothesis Testing (Inference) Regressions 2 1 Making an Investment Decision A researcher in your firm just
More informationCHAPTER 13 SIMPLE LINEAR REGRESSION. Opening Example. Simple Regression. Linear Regression
Opening Example CHAPTER 13 SIMPLE LINEAR REGREION SIMPLE LINEAR REGREION! Simple Regression! Linear Regression Simple Regression Definition A regression model is a mathematical equation that descries the
More informationUnivariate Regression
Univariate Regression Correlation and Regression The regression line summarizes the linear relationship between 2 variables Correlation coefficient, r, measures strength of relationship: the closer r is
More informationCorrelation. What Is Correlation? Perfect Correlation. Perfect Correlation. Greg C Elvers
Correlation Greg C Elvers What Is Correlation? Correlation is a descriptive statistic that tells you if two variables are related to each other E.g. Is your related to how much you study? When two variables
More informationScatter Plots with Error Bars
Chapter 165 Scatter Plots with Error Bars Introduction The procedure extends the capability of the basic scatter plot by allowing you to plot the variability in Y and X corresponding to each point. Each
More informationUnit 26 Estimation with Confidence Intervals
Unit 26 Estimation with Confidence Intervals Objectives: To see how confidence intervals are used to estimate a population proportion, a population mean, a difference in population proportions, or a difference
More informationExercise 1.12 (Pg. 22-23)
Individuals: The objects that are described by a set of data. They may be people, animals, things, etc. (Also referred to as Cases or Records) Variables: The characteristics recorded about each individual.
More informationAnalysis of Variance ANOVA
Analysis of Variance ANOVA Overview We ve used the t -test to compare the means from two independent groups. Now we ve come to the final topic of the course: how to compare means from more than two populations.
More informationIntroduction to Quantitative Methods
Introduction to Quantitative Methods October 15, 2009 Contents 1 Definition of Key Terms 2 2 Descriptive Statistics 3 2.1 Frequency Tables......................... 4 2.2 Measures of Central Tendencies.................
More informationInteraction between quantitative predictors
Interaction between quantitative predictors In a first-order model like the ones we have discussed, the association between E(y) and a predictor x j does not depend on the value of the other predictors
More informationElements of a graph. Click on the links below to jump directly to the relevant section
Click on the links below to jump directly to the relevant section Elements of a graph Linear equations and their graphs What is slope? Slope and y-intercept in the equation of a line Comparing lines on
More informationChapter 5 Analysis of variance SPSS Analysis of variance
Chapter 5 Analysis of variance SPSS Analysis of variance Data file used: gss.sav How to get there: Analyze Compare Means One-way ANOVA To test the null hypothesis that several population means are equal,
More informationEconometrics Simple Linear Regression
Econometrics Simple Linear Regression Burcu Eke UC3M Linear equations with one variable Recall what a linear equation is: y = b 0 + b 1 x is a linear equation with one variable, or equivalently, a straight
More informationEQUATIONS and INEQUALITIES
EQUATIONS and INEQUALITIES Linear Equations and Slope 1. Slope a. Calculate the slope of a line given two points b. Calculate the slope of a line parallel to a given line. c. Calculate the slope of a line
More informationPart 1 Expressions, Equations, and Inequalities: Simplifying and Solving
Section 7 Algebraic Manipulations and Solving Part 1 Expressions, Equations, and Inequalities: Simplifying and Solving Before launching into the mathematics, let s take a moment to talk about the words
More informationLOGIT AND PROBIT ANALYSIS
LOGIT AND PROBIT ANALYSIS A.K. Vasisht I.A.S.R.I., Library Avenue, New Delhi 110 012 amitvasisht@iasri.res.in In dummy regression variable models, it is assumed implicitly that the dependent variable Y
More informationRegression step-by-step using Microsoft Excel
Step 1: Regression step-by-step using Microsoft Excel Notes prepared by Pamela Peterson Drake, James Madison University Type the data into the spreadsheet The example used throughout this How to is a regression
More informationCOMP6053 lecture: Relationship between two variables: correlation, covariance and r-squared. jn2@ecs.soton.ac.uk
COMP6053 lecture: Relationship between two variables: correlation, covariance and r-squared jn2@ecs.soton.ac.uk Relationships between variables So far we have looked at ways of characterizing the distribution
More informationPLOTTING DATA AND INTERPRETING GRAPHS
PLOTTING DATA AND INTERPRETING GRAPHS Fundamentals of Graphing One of the most important sets of skills in science and mathematics is the ability to construct graphs and to interpret the information they
More informationMultiple Linear Regression
Multiple Linear Regression A regression with two or more explanatory variables is called a multiple regression. Rather than modeling the mean response as a straight line, as in simple regression, it is
More informationSection 14 Simple Linear Regression: Introduction to Least Squares Regression
Slide 1 Section 14 Simple Linear Regression: Introduction to Least Squares Regression There are several different measures of statistical association used for understanding the quantitative relationship
More informationSimple Predictive Analytics Curtis Seare
Using Excel to Solve Business Problems: Simple Predictive Analytics Curtis Seare Copyright: Vault Analytics July 2010 Contents Section I: Background Information Why use Predictive Analytics? How to use
More informationMULTIPLE REGRESSION EXAMPLE
MULTIPLE REGRESSION EXAMPLE For a sample of n = 166 college students, the following variables were measured: Y = height X 1 = mother s height ( momheight ) X 2 = father s height ( dadheight ) X 3 = 1 if
More informationSimple Linear Regression Inference
Simple Linear Regression Inference 1 Inference requirements The Normality assumption of the stochastic term e is needed for inference even if it is not a OLS requirement. Therefore we have: Interpretation
More informationExample: Boats and Manatees
Figure 9-6 Example: Boats and Manatees Slide 1 Given the sample data in Table 9-1, find the value of the linear correlation coefficient r, then refer to Table A-6 to determine whether there is a significant
More informationNCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( )
Chapter 340 Principal Components Regression Introduction is a technique for analyzing multiple regression data that suffer from multicollinearity. When multicollinearity occurs, least squares estimates
More informationWhat does the number m in y = mx + b measure? To find out, suppose (x 1, y 1 ) and (x 2, y 2 ) are two points on the graph of y = mx + b.
PRIMARY CONTENT MODULE Algebra - Linear Equations & Inequalities T-37/H-37 What does the number m in y = mx + b measure? To find out, suppose (x 1, y 1 ) and (x 2, y 2 ) are two points on the graph of
More informationScatter Plot, Correlation, and Regression on the TI-83/84
Scatter Plot, Correlation, and Regression on the TI-83/84 Summary: When you have a set of (x,y) data points and want to find the best equation to describe them, you are performing a regression. This page
More informationSample Size and Power in Clinical Trials
Sample Size and Power in Clinical Trials Version 1.0 May 011 1. Power of a Test. Factors affecting Power 3. Required Sample Size RELATED ISSUES 1. Effect Size. Test Statistics 3. Variation 4. Significance
More informationBecause the slope is, a slope of 5 would mean that for every 1cm increase in diameter, the circumference would increase by 5cm.
Measurement Lab You will be graphing circumference (cm) vs. diameter (cm) for several different circular objects, and finding the slope of the line of best fit using the CapStone program. Write out or
More informationModule 5: Statistical Analysis
Module 5: Statistical Analysis To answer more complex questions using your data, or in statistical terms, to test your hypothesis, you need to use more advanced statistical tests. This module reviews the
More informationCOMPARISONS OF CUSTOMER LOYALTY: PUBLIC & PRIVATE INSURANCE COMPANIES.
277 CHAPTER VI COMPARISONS OF CUSTOMER LOYALTY: PUBLIC & PRIVATE INSURANCE COMPANIES. This chapter contains a full discussion of customer loyalty comparisons between private and public insurance companies
More informationStatistical tests for SPSS
Statistical tests for SPSS Paolo Coletti A.Y. 2010/11 Free University of Bolzano Bozen Premise This book is a very quick, rough and fast description of statistical tests and their usage. It is explicitly
More informationCALCULATIONS & STATISTICS
CALCULATIONS & STATISTICS CALCULATION OF SCORES Conversion of 1-5 scale to 0-100 scores When you look at your report, you will notice that the scores are reported on a 0-100 scale, even though respondents
More informationQUANTITATIVE METHODS BIOLOGY FINAL HONOUR SCHOOL NON-PARAMETRIC TESTS
QUANTITATIVE METHODS BIOLOGY FINAL HONOUR SCHOOL NON-PARAMETRIC TESTS This booklet contains lecture notes for the nonparametric work in the QM course. This booklet may be online at http://users.ox.ac.uk/~grafen/qmnotes/index.html.
More information1. The parameters to be estimated in the simple linear regression model Y=α+βx+ε ε~n(0,σ) are: a) α, β, σ b) α, β, ε c) a, b, s d) ε, 0, σ
STA 3024 Practice Problems Exam 2 NOTE: These are just Practice Problems. This is NOT meant to look just like the test, and it is NOT the only thing that you should study. Make sure you know all the material
More informationRelationships Between Two Variables: Scatterplots and Correlation
Relationships Between Two Variables: Scatterplots and Correlation Example: Consider the population of cars manufactured in the U.S. What is the relationship (1) between engine size and horsepower? (2)
More informationThe Dummy s Guide to Data Analysis Using SPSS
The Dummy s Guide to Data Analysis Using SPSS Mathematics 57 Scripps College Amy Gamble April, 2001 Amy Gamble 4/30/01 All Rights Rerserved TABLE OF CONTENTS PAGE Helpful Hints for All Tests...1 Tests
More informationX X X a) perfect linear correlation b) no correlation c) positive correlation (r = 1) (r = 0) (0 < r < 1)
CORRELATION AND REGRESSION / 47 CHAPTER EIGHT CORRELATION AND REGRESSION Correlation and regression are statistical methods that are commonly used in the medical literature to compare two or more variables.
More informationAssociation Between Variables
Contents 11 Association Between Variables 767 11.1 Introduction............................ 767 11.1.1 Measure of Association................. 768 11.1.2 Chapter Summary.................... 769 11.2 Chi
More informationChapter 13 Introduction to Linear Regression and Correlation Analysis
Chapter 3 Student Lecture Notes 3- Chapter 3 Introduction to Linear Regression and Correlation Analsis Fall 2006 Fundamentals of Business Statistics Chapter Goals To understand the methods for displaing
More informationKSTAT MINI-MANUAL. Decision Sciences 434 Kellogg Graduate School of Management
KSTAT MINI-MANUAL Decision Sciences 434 Kellogg Graduate School of Management Kstat is a set of macros added to Excel and it will enable you to do the statistics required for this course very easily. To
More informationThe Correlation Coefficient
The Correlation Coefficient Lelys Bravo de Guenni April 22nd, 2015 Outline The Correlation coefficient Positive Correlation Negative Correlation Properties of the Correlation Coefficient Non-linear association
More informationLinear functions Increasing Linear Functions. Decreasing Linear Functions
3.5 Increasing, Decreasing, Max, and Min So far we have been describing graphs using quantitative information. That s just a fancy way to say that we ve been using numbers. Specifically, we have described
More informationAnswer: C. The strength of a correlation does not change if units change by a linear transformation such as: Fahrenheit = 32 + (5/9) * Centigrade
Statistics Quiz Correlation and Regression -- ANSWERS 1. Temperature and air pollution are known to be correlated. We collect data from two laboratories, in Boston and Montreal. Boston makes their measurements
More informationStatistics courses often teach the two-sample t-test, linear regression, and analysis of variance
2 Making Connections: The Two-Sample t-test, Regression, and ANOVA In theory, there s no difference between theory and practice. In practice, there is. Yogi Berra 1 Statistics courses often teach the two-sample
More informationThis chapter will demonstrate how to perform multiple linear regression with IBM SPSS
CHAPTER 7B Multiple Regression: Statistical Methods Using IBM SPSS This chapter will demonstrate how to perform multiple linear regression with IBM SPSS first using the standard method and then using the
More informationIndependent samples t-test. Dr. Tom Pierce Radford University
Independent samples t-test Dr. Tom Pierce Radford University The logic behind drawing causal conclusions from experiments The sampling distribution of the difference between means The standard error of
More information2. What is the general linear model to be used to model linear trend? (Write out the model) = + + + or
Simple and Multiple Regression Analysis Example: Explore the relationships among Month, Adv.$ and Sales $: 1. Prepare a scatter plot of these data. The scatter plots for Adv.$ versus Sales, and Month versus
More informationReview Jeopardy. Blue vs. Orange. Review Jeopardy
Review Jeopardy Blue vs. Orange Review Jeopardy Jeopardy Round Lectures 0-3 Jeopardy Round $200 How could I measure how far apart (i.e. how different) two observations, y 1 and y 2, are from each other?
More informationCopyright 2007 by Laura Schultz. All rights reserved. Page 1 of 5
Using Your TI-83/84 Calculator: Linear Correlation and Regression Elementary Statistics Dr. Laura Schultz This handout describes how to use your calculator for various linear correlation and regression
More informationChapter 15. Mixed Models. 15.1 Overview. A flexible approach to correlated data.
Chapter 15 Mixed Models A flexible approach to correlated data. 15.1 Overview Correlated data arise frequently in statistical analyses. This may be due to grouping of subjects, e.g., students within classrooms,
More informationChapter 23. Inferences for Regression
Chapter 23. Inferences for Regression Topics covered in this chapter: Simple Linear Regression Simple Linear Regression Example 23.1: Crying and IQ The Problem: Infants who cry easily may be more easily
More informationCategorical Data Analysis
Richard L. Scheaffer University of Florida The reference material and many examples for this section are based on Chapter 8, Analyzing Association Between Categorical Variables, from Statistical Methods
More informationEstimation of σ 2, the variance of ɛ
Estimation of σ 2, the variance of ɛ The variance of the errors σ 2 indicates how much observations deviate from the fitted surface. If σ 2 is small, parameters β 0, β 1,..., β k will be reliably estimated
More informationT O P I C 1 2 Techniques and tools for data analysis Preview Introduction In chapter 3 of Statistics In A Day different combinations of numbers and types of variables are presented. We go through these
More informationDEPARTMENT OF PSYCHOLOGY UNIVERSITY OF LANCASTER MSC IN PSYCHOLOGICAL RESEARCH METHODS ANALYSING AND INTERPRETING DATA 2 PART 1 WEEK 9
DEPARTMENT OF PSYCHOLOGY UNIVERSITY OF LANCASTER MSC IN PSYCHOLOGICAL RESEARCH METHODS ANALYSING AND INTERPRETING DATA 2 PART 1 WEEK 9 Analysis of covariance and multiple regression So far in this course,
More informationNonlinear Regression Functions. SW Ch 8 1/54/
Nonlinear Regression Functions SW Ch 8 1/54/ The TestScore STR relation looks linear (maybe) SW Ch 8 2/54/ But the TestScore Income relation looks nonlinear... SW Ch 8 3/54/ Nonlinear Regression General
More informationABSORBENCY OF PAPER TOWELS
ABSORBENCY OF PAPER TOWELS 15. Brief Version of the Case Study 15.1 Problem Formulation 15.2 Selection of Factors 15.3 Obtaining Random Samples of Paper Towels 15.4 How will the Absorbency be measured?
More informationNotes on Applied Linear Regression
Notes on Applied Linear Regression Jamie DeCoster Department of Social Psychology Free University Amsterdam Van der Boechorststraat 1 1081 BT Amsterdam The Netherlands phone: +31 (0)20 444-8935 email:
More informationSIMPLE LINEAR CORRELATION. r can range from -1 to 1, and is independent of units of measurement. Correlation can be done on two dependent variables.
SIMPLE LINEAR CORRELATION Simple linear correlation is a measure of the degree to which two variables vary together, or a measure of the intensity of the association between two variables. Correlation
More informationUsing Excel for inferential statistics
FACT SHEET Using Excel for inferential statistics Introduction When you collect data, you expect a certain amount of variation, just caused by chance. A wide variety of statistical tests can be applied
More informationStudy Guide for the Final Exam
Study Guide for the Final Exam When studying, remember that the computational portion of the exam will only involve new material (covered after the second midterm), that material from Exam 1 will make
More informationLecture 11: Chapter 5, Section 3 Relationships between Two Quantitative Variables; Correlation
Lecture 11: Chapter 5, Section 3 Relationships between Two Quantitative Variables; Correlation Display and Summarize Correlation for Direction and Strength Properties of Correlation Regression Line Cengage
More informationFactors affecting online sales
Factors affecting online sales Table of contents Summary... 1 Research questions... 1 The dataset... 2 Descriptive statistics: The exploratory stage... 3 Confidence intervals... 4 Hypothesis tests... 4
More informationTwo-Sample T-Tests Assuming Equal Variance (Enter Means)
Chapter 4 Two-Sample T-Tests Assuming Equal Variance (Enter Means) Introduction This procedure provides sample size and power calculations for one- or two-sided two-sample t-tests when the variances of
More informationSection 1.1 Linear Equations: Slope and Equations of Lines
Section. Linear Equations: Slope and Equations of Lines Slope The measure of the steepness of a line is called the slope of the line. It is the amount of change in y, the rise, divided by the amount of
More informationA full analysis example Multiple correlations Partial correlations
A full analysis example Multiple correlations Partial correlations New Dataset: Confidence This is a dataset taken of the confidence scales of 41 employees some years ago using 4 facets of confidence (Physical,
More informationSensitivity Analysis 3.1 AN EXAMPLE FOR ANALYSIS
Sensitivity Analysis 3 We have already been introduced to sensitivity analysis in Chapter via the geometry of a simple example. We saw that the values of the decision variables and those of the slack and
More informationRegression and Correlation
Regression and Correlation Topics Covered: Dependent and independent variables. Scatter diagram. Correlation coefficient. Linear Regression line. by Dr.I.Namestnikova 1 Introduction Regression analysis
More informationA Review of Cross Sectional Regression for Financial Data You should already know this material from previous study
A Review of Cross Sectional Regression for Financial Data You should already know this material from previous study But I will offer a review, with a focus on issues which arise in finance 1 TYPES OF FINANCIAL
More informationSPSS Explore procedure
SPSS Explore procedure One useful function in SPSS is the Explore procedure, which will produce histograms, boxplots, stem-and-leaf plots and extensive descriptive statistics. To run the Explore procedure,
More information11. Analysis of Case-control Studies Logistic Regression
Research methods II 113 11. Analysis of Case-control Studies Logistic Regression This chapter builds upon and further develops the concepts and strategies described in Ch.6 of Mother and Child Health:
More informationTwo-Sample T-Tests Allowing Unequal Variance (Enter Difference)
Chapter 45 Two-Sample T-Tests Allowing Unequal Variance (Enter Difference) Introduction This procedure provides sample size and power calculations for one- or two-sided two-sample t-tests when no assumption
More informationMath 4310 Handout - Quotient Vector Spaces
Math 4310 Handout - Quotient Vector Spaces Dan Collins The textbook defines a subspace of a vector space in Chapter 4, but it avoids ever discussing the notion of a quotient space. This is understandable
More information