Coefficients of determination
|
|
|
- Logan Dalton
- 9 years ago
- Views:
Transcription
1 Coefficients of determination Jean-Marie Dufour McGill University First version: March 1983 Revised: February 2002, July 2011 his version: July 2011 Compiled: November 21, 2011, 11:05 his work was supported by the William Dow Chair in Political Economy (McGill University), the Bank of Canada (Research Fellowship), a Guggenheim Fellowship, a Konrad-Adenauer Fellowship (Alexander-von-Humboldt Foundation, Germany), the Canadian Network of Centres of Excellence [program on Mathematics of Information echnology and Complex Systems (MIACS)], the Natural Sciences and Engineering Research Council of Canada, the Social Sciences and Humanities Research Council of Canada, and the Fonds de recherche sur la société et la culture (Québec). William Dow Professor of Economics, McGill University, Centre interuniversitaire de recherche en analyse des organisations (CIRANO), and Centre interuniversitaire de recherche en économie quantitative (CIREQ). Mailing address: Department of Economics, McGill University, Leacock Building, Room 519, 855 Sherbrooke Street West, Montréal, Québec H3A 27, Canada. EL: (1) ; FAX: (1) ; [email protected]. Web page:
2 Contents 1. Coefficient of determination: R Significance tests and R Relation of R 2 with a Fisher statistic General relation between R 2 and Fisher tests Uncentered coefficient of determination: R Adjusted coefficient of determination: R Definition and basic properties Criterion for R 2 increase through the omission of an explanatory variable Generalized criterion for R 2 increase through the imposition of linear constraints 8 5. Notes on bibliography Chronological list of references 10 i
3 1. Coefficient of determination: R 2 Let y Xβ + ε be a model that satisfies the assumptions of the classical linear model, where y and ε are 1 vectors, X is a k matrix and β is k 1 coefficient vector. We wish to characterize to which extent the variables included in X (excluding the constant, if there is one) explain y. A first method consists in computing R 2, the coefficient of determination, or R R 2, the coefficient of multiple correlation. Let ŷ X ˆβ, ˆε y ŷ, y y t / i y/, (1.1) i (1,1,..., 1) the unit vector of dimension, (1.2) SS SSR SSE We can then define variance estimators as follows: (y t y) 2 (y iy) (y iy), (total sum of squares) (1.3) (ŷ t y) 2 (ŷ iy) (ŷ iy), (regression sum of squares) (1.4) (y t ŷ t ) 2 (y ŷ) (y ŷ) ˆε ˆε, (error sum of squares). (1.5) 1.1 Definition R 2 1 ( ˆV (ε)/ ˆV(y) ) 1 (SSE/SS). 1.2 Proposition R 2 1. PROOF his result is immediate on observing that SSE/SS 0. ˆV (y) SS/, (1.6) ˆV (ŷ) SSR/, (1.7) ˆV (ε) SSE/. (1.8) 1.3 Lemma y y ŷ ŷ+ ˆε ˆε. PROOF hence We have y ŷ+ ˆε and ŷ ˆε ˆε ŷ 0, (1.9) y y (ŷ+ ˆε) (ŷ+ ˆε) ŷ ŷ+ŷ ˆε + ˆε ŷ+ ˆε ˆε ŷ ŷ+ ˆε ˆε. 1
4 1.4 Proposition If one of the regressors is a constant, then SS SSR + SSE, ˆV (y) ˆV (ŷ)+ ˆV (ε). PROOF Let A I i(i i) 1 i I 1 ii. hen, A A A and Ay [I 1 ] ii y y iy. If one of the regressors is a constant, we have hence 1 i ˆε ˆε t 0 ŷ t 1 i ŷ 1 i (y ˆε) 1 i y y, Aˆε ˆε 1 ii ˆε ˆε, and, using the fact that Aˆε ˆε and ŷ ˆε 0, Aŷ ŷ 1 ii ŷ ŷ iy, SS (y iy) (y iy) y A Ay y Ay (ŷ+ ˆε) A(ŷ+ ˆε) ŷ Aŷ+ŷ Aˆε + ŷ Aˆε + ˆε Aˆε ŷ Aŷ+ ˆε ˆε (Aŷ) (Aŷ)+ ˆε ˆε SSR+SSE. 1.5 Proposition If one of the regressors is a constant, R 2 ˆV (ŷ) ˆV (y) SSR SS and 0 R 2 1. PROOF By the definition of R 2, we have R 2 1 and R 2 1 ˆV (ε) ˆV (y) ˆV (y) ˆV (ε) ˆV (ŷ) ˆV (y) ˆV (y) SSR SS 2
5 hence R Proposition If one of the regressors is a constant, the empirical correlation between y and ŷ is non-negative and equal to R 2. PROOF he empirical correlation between y and ŷ is defined by ˆρ(y,ŷ) Ĉ(y,ŷ) [ ˆV (y) ˆV (ŷ) ] 1/2 where Ĉ(y,ŷ) 1 (y t y)(ŷ t y) 1 (Ay) (Aŷ) and A I 1 ii. Since one of the regressors is a constant, and Aˆε ˆε, Ay Aŷ+ ˆε, ˆε (Aŷ) ˆε ŷ 0 Ĉ(y,ŷ) 1 (Aŷ+ ˆε) (Aŷ) 1 (Aŷ) (Aŷ) ˆV (ŷ), ˆρ(y, ŷ) [ ] 1/2 ˆV (ŷ) ˆV [ ˆV (y) ˆV (ŷ) ] 1/2 (ŷ) R ˆV 2 0. (y) 2. Significance tests and R Relation of R 2 with a Fisher statistic R 2 is descriptive statistic which measures the proportion of the variance of the dependent variable y explained by suggested explanatory variables (excluding the constant). However, R 2 can be related to a significance test (under the assumptions of the Gaussian classical linear model). Consider the model y t β 1 + β 2 X t2 + +β k X tk + ε t, t 1,...,. We wish to test the hypothesis that none of these variables (excluding the constant) should appear in the equation: H 0 : β 2 β 3 β k 0. 3
6 he Fisher statistic for H 0 is F (S ω S Ω )/q S Ω /( k) F (q, k) where q k 1, S Ω is the error sum of squares from the estimation of the unconstrained model Ω : y Xβ + ε, where X [i,x 2,..., X k ] and S ω s the error sum of squares from the estimation of the constrained model ω : y iβ 1 + ε, where i (1,1,..., 1). We see easily that S Ω ( y X ˆβ ) ( y X ˆβ ) SSE, ˆβ 1 S ω ( i i ) 1 i y 1 (y iy) (y iy) SS y t y, (under ω) and F (SS SSE)/(k 1) SSE/( k) [ ] 1 SSE SS /(k 1) /( k) SSE SS R 2 /(k 1) (1 R 2 F (k 1, k). )/( k) As R 2 increases, F increases General relation between R 2 and Fisher tests Consider the general linear hypothesis H 0 : Cβ r where C : q k, β : k 1, r : q 1 and rank(c) q. he values of R 2 for the constrained and unconstrained models are respectively: hence R S ω SS, R2 1 1 S Ω SS, S ω ( 1 R 2 0) SS, SΩ ( 1 R 2 1) SS. 4
7 he Fisher statistic for testing H 0 may thus be written F (S ( ω S Ω )/q R 2 S Ω /( k) 1 R 2 ) 0 /q ( ) 1 R 2 1 /( k) ( ) k R 2 1 R 2 0 q 1 R 2. 1 If R 2 1 R2 0 is large, we tend to reject H 0. If H 0 : β 2 β 3 β k 0, then q k 1, S ω SS, R and the formula for F above gets reduced of the one given in section Uncentered coefficient of determination: R 2 Since R 2 can take negative values when the model does not contain a constant, R 2 has little meaning in this case. In such situations, we can instead use a coefficient where the values of y t are not centered around the mean. 3.1 Definition R 2 1 (ˆε ˆε/y y ). R 2 is called the uncentered coefficient of determination on uncentered R 2 and R R 2 the uncentered coefficient of multiple correlation. 3.2 Proposition 0 R 2 1. PROOF his follows directly from Lemma 1.3: y y ŷ ŷ+ ˆε ˆε. 4. Adjusted coefficient of determination: R Definition and basic properties An unattractive property of the R 2 coefficient comes form the fact that R 2 cannot decrease when explanatory variables are added to the model, even if these have no relevance. Consequently, choosing to maximize R 2 can be misleading. It seems desirable to penalize models that contain too many variables. Since R 2 1 ˆV (ε) ˆV (y), 5
8 where ˆV (ε) SSE 1 ˆε t 2, ˆV (y) SS 1 (y t y) 2, heil (1961, p. 213) suggested to replace ˆV (ε) and ˆV (y) by unbiased estimators : s 2 SSE k 1 k s 2 y SS ˆε 2 t, (y t y) Definition R 2 adjusted for degrees of freedom is defined by R 2 1 s2 s 2 y 1 1 k ( ) SSE. SS 4.2 Proposition R ( k 1 R 2 ) R 2 k 1 ( k 1 R 2 ). PROOF R k ( ) SSE SS 1 k+ k 1 k 1 ( 1 R 2) k 1 k 1 1 ( 1 R 2 ) k ( 1 R 2 ) ( 1 1+ k 1 ) (1 R 2 ) k ( 1 R 2 ) R 2 k 1 ( 1 R 2 ). Q.E.D. k 4.3 Proposition R 2 R 2 1. PROOF he result follows from the fact that 1 R 2 0 and (4.2). 4.4 Proposition R 2 R 2 iff (k 1 or R 2 1). 4.5 Proposition R 2 0 iff R 2 k 1 1. R 2 can be negative even if R 2 0. If the number of explanatory variables is increased, R 2 and k both increase, so that R 2 can increase or decrease. 6
9 4.6 Remark When several models are compared on the basis of R 2 or R 2, it is important to have the same dependent variable. When the dependent variable (y) is the same, maximizing R 2 is equivalent to minimizing the standard error of the regression [ 1 s k ˆε 2 t ] 1/ Criterion for R 2 increase through the omission of an explanatory variable Consider the two models: y t β 1 X t1 + +β k 1 X t(k 1) + ε t, t 1,...,, (4.1) y t β 1 X t1,+ +β k 1 X t(k 1) + β k X tk + ε t, t 1,...,. (4.2) We can then show that the value of R 2 associated with the restricted model (4.1) is larger than the one of model (4.2) if the t statistic for testing β k 0 is smaller than 1 (in absolute value). 4.7 Proposition If R 2 k 1 and R 2 k are the values of R 2 for models (4.1) and (4.2), then ( ) 1 R 2 R 2 k R 2 k ( k 1 t 2 ( k+ 1) k 1 ) (4.3) where t k is the Student t statistic for testing β k 0 in model (4.2), and R 2 k R 2 k 1 iff t 2 k 1 iff t k 1. If furthermore R 2 k < 1, then R 2 k R 2 k 1 iff t k 1. PROOF By definition, R 2 k 1 s2 k s 2 y and R 2 k 1 1 s2 k 1 s 2 y where s 2 k SS k/( k) and s 2 k 1 SS k 1/( k+ 1). SS k and SS k 1 are the sums of squared errors for the models with k and k 1 explanatory variables. Since tk 2 is the Fisher statistic for testing β k 0, we have t 2 k (SS k 1 SS k ) SS k /( k) 7
10 for s 2 k 1 s2 y ( ) 1 R 2 k 1 [ ( k+ 1)s 2 k 1 ( k)s 2 k] s 2 k ( ( k+ 1) 1 R 2 k 1 ) ( ) ( k) 1 R 2 k 1 R 2 k ( 1 R 2 ) k 1 ( k+ 1) 1 R 2 ( k) k ( ) and s 2 k s2 y 1 R 2 k. Consequently, ( ) [ 1 R 2 k 1 1 R 2 tk 2 +( k)] k k+ 1 and R 2 k R 2 k 1 ( ) ( ) 1 R 2 k 1 1 R 2 k ( ) [ 1 R 2 t 2 ] k +( k) k k+ 1 1 ( ) [ 1 R 2 tk 2 1 ] k. k Generalized criterion for R 2 increase through the imposition of linear constraints We will now study when the imposition of q linearly independent constraints H 0 : Cβ r will raise or decrease R 2, where C : q k, r : q 1 and rank(c) q. Let R 2 H 0 and R 2 be the values of R 2 for the constrained (by H 0 ) and unconstrained models, similarly, s 2 0 and s2 are the values of the corresponding unbiased estimators of the error variance. 4.8 Proposition Let F be the Fisher statistic for testing H 0. hen s 2 0 s 2 qs 2 (F 1) k+ q 8
11 and s 2 0 s 2 iff F 1. PROOF If SS 0 and SS are the sum of squared errors for the constrained and unconstrained models, we have: s 2 0 SS 0 and s 2 SS k+ q k. he F statistic may then be written hence F (SS 0 SS)/q SS/( k) [ ( k+ q)s 2 0 ( k)s 2] qs 2 k+ q q s 2 0 s 2[qF +( k)] ( k)+q s 2 0 s 2 2 q(f 1) s ( k)+q,, ( s 2) 0 s 2 k q and s 2 0 s 2 iff F Proposition Let F be the Fisher statistic for testing H 0. hen q (1 R 2) R 2 R 2 H 0 (F 1) k+ q and PROOF hus, By definition, R 2 H 0 R 2 iff F 1. R 2 H 0 1 s2 0 s 2 y, R 2 1 s2 s 2 y. R 2 R 2 H 0 s2 s 2 0 s 2 y q k+ q ( s 2 s 2 y ) (F 1) 9
12 q (1 R 2) k+ q (F 1) hence R 2 H 0 R 2 iff F 1. On taking q 1, we get property (4.3). If we test an hypothesis of the type H 0 : β k β k+1 β k+l 0, it is possible that F > 1, while all the statistics t i, i k,..., k+l are smaller than 1. his means that R 2 increases when we omit one explanatory variable at a time, but decreases when they are all excluded from the regression. Further, it is also possible that F < 1, but t i > 1 for all i: R 2 increases when all the explanatory variables are simultaneously excluded, but decreases when only one is excluded. 5. Notes on bibliography he notion of R 2 was proposed by heil (1961, p. 213). Several authors have presented detailed discussions of the different concepts of multiple correlation: for example, heil (1971, Chap. 4), Schmidt (1976) and Maddala (1977, Sections 8.1, 8.2, 8.3, 8.9). he R 2 concept is criticized by Pesaran (1974). he mean and bias of R 2 were studied by Cramer (1987) in the Gaussian case, and by Srivastava, Srivastava and Ullah (1995) in some non-gaussian cases. 6. Chronological list of references 1. heil (1961, p. 213) _ he R 2 nation was proposed in this book. 2. heil (1971, Chap. 4) _ Detailed discussion of R 2, R 2 and partial correlation. 3. Pesaran (1974) _ Critique of R Schmidt (1976) 5. Maddala (1977, Sections 8.1, 8.2, 8.3, 8.9) _ Discussion of R 2 and R 2 along with their relation with hypothesis tests. 6. Hendry and Marshall (1983) 7. Cramer (1987) 8. Ohtani and Hasegawa (1993) 10
13 9. Srivastava et al. (1995) 11
14 References Cramer, J. S. (1987), Mean and variance of R 2 in small and moderate samples, Econometric Reviews 35, Hendry, D. F. and Marshall, R. C. (1983), On high and low R 2 contributions, Oxford Bulletin of Economics and Statistics 45, Maddala, G. S. (1977), Econometrics, McGraw-Hill, New York. Ohtani, L. and Hasegawa, H. (1993), On small-sample properties of R 2 in a linear regression model with multivariate t errors and proxy variables, Econometric heory 9, Pesaran, M. H. (1974), On the general problem of model selection, Review of Economic Studies 41, Schmidt, P. (1976), Econometrics, Marcel Dekker, New York. Srivastava, A. K., Srivastava, V. K. and Ullah, A. (1995), he coefficient of determination and its adjusted version in linear regression models, Econometric Reviews 14, heil, H. (1961), Economic Forecasts and Policy, 2nd Edition, North-Holland, Amsterdam. heil, H. (1971), Principles of Econometrics, John Wiley & Sons, New York. 12
Introduction to time series analysis
Introduction to time series analysis Jean-Marie Dufour First version: December 1998 Revised: January 2003 This version: January 8, 2008 Compiled: January 8, 2008, 6:12pm This work was supported by the
Properties of moments of random variables
Properties of moments of rom variables Jean-Marie Dufour Université de Montréal First version: May 1995 Revised: January 23 This version: January 14, 23 Compiled: January 14, 23, 1:5pm This work was supported
Econometrics Simple Linear Regression
Econometrics Simple Linear Regression Burcu Eke UC3M Linear equations with one variable Recall what a linear equation is: y = b 0 + b 1 x is a linear equation with one variable, or equivalently, a straight
Mean squared error matrix comparison of least aquares and Stein-rule estimators for regression coefficients under non-normal disturbances
METRON - International Journal of Statistics 2008, vol. LXVI, n. 3, pp. 285-298 SHALABH HELGE TOUTENBURG CHRISTIAN HEUMANN Mean squared error matrix comparison of least aquares and Stein-rule estimators
Simple Linear Regression Inference
Simple Linear Regression Inference 1 Inference requirements The Normality assumption of the stochastic term e is needed for inference even if it is not a OLS requirement. Therefore we have: Interpretation
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written
Introduction to General and Generalized Linear Models
Introduction to General and Generalized Linear Models General Linear Models - part I Henrik Madsen Poul Thyregod Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby
Solución del Examen Tipo: 1
Solución del Examen Tipo: 1 Universidad Carlos III de Madrid ECONOMETRICS Academic year 2009/10 FINAL EXAM May 17, 2010 DURATION: 2 HOURS 1. Assume that model (III) verifies the assumptions of the classical
IAPRI Quantitative Analysis Capacity Building Series. Multiple regression analysis & interpreting results
IAPRI Quantitative Analysis Capacity Building Series Multiple regression analysis & interpreting results How important is R-squared? R-squared Published in Agricultural Economics 0.45 Best article of the
MATHEMATICAL METHODS OF STATISTICS
MATHEMATICAL METHODS OF STATISTICS By HARALD CRAMER TROFESSOK IN THE UNIVERSITY OF STOCKHOLM Princeton PRINCETON UNIVERSITY PRESS 1946 TABLE OF CONTENTS. First Part. MATHEMATICAL INTRODUCTION. CHAPTERS
Chapter 3: The Multiple Linear Regression Model
Chapter 3: The Multiple Linear Regression Model Advanced Econometrics - HEC Lausanne Christophe Hurlin University of Orléans November 23, 2013 Christophe Hurlin (University of Orléans) Advanced Econometrics
2. Linear regression with multiple regressors
2. Linear regression with multiple regressors Aim of this section: Introduction of the multiple regression model OLS estimation in multiple regression Measures-of-fit in multiple regression Assumptions
Testing for Granger causality between stock prices and economic growth
MPRA Munich Personal RePEc Archive Testing for Granger causality between stock prices and economic growth Pasquale Foresti 2006 Online at http://mpra.ub.uni-muenchen.de/2962/ MPRA Paper No. 2962, posted
Outline. Topic 4 - Analysis of Variance Approach to Regression. Partitioning Sums of Squares. Total Sum of Squares. Partitioning sums of squares
Topic 4 - Analysis of Variance Approach to Regression Outline Partitioning sums of squares Degrees of freedom Expected mean squares General linear test - Fall 2013 R 2 and the coefficient of correlation
Standard errors of marginal effects in the heteroskedastic probit model
Standard errors of marginal effects in the heteroskedastic probit model Thomas Cornelißen Discussion Paper No. 320 August 2005 ISSN: 0949 9962 Abstract In non-linear regression models, such as the heteroskedastic
Chapter 4: Vector Autoregressive Models
Chapter 4: Vector Autoregressive Models 1 Contents: Lehrstuhl für Department Empirische of Wirtschaftsforschung Empirical Research and und Econometrics Ökonometrie IV.1 Vector Autoregressive Models (VAR)...
Quadratic forms Cochran s theorem, degrees of freedom, and all that
Quadratic forms Cochran s theorem, degrees of freedom, and all that Dr. Frank Wood Frank Wood, [email protected] Linear Regression Models Lecture 1, Slide 1 Why We Care Cochran s theorem tells us
Part 2: Analysis of Relationship Between Two Variables
Part 2: Analysis of Relationship Between Two Variables Linear Regression Linear correlation Significance Tests Multiple regression Linear Regression Y = a X + b Dependent Variable Independent Variable
Statistical Tests for Multiple Forecast Comparison
Statistical Tests for Multiple Forecast Comparison Roberto S. Mariano (Singapore Management University & University of Pennsylvania) Daniel Preve (Uppsala University) June 6-7, 2008 T.W. Anderson Conference,
PROBABILITY AND STATISTICS. Ma 527. 1. To teach a knowledge of combinatorial reasoning.
PROBABILITY AND STATISTICS Ma 527 Course Description Prefaced by a study of the foundations of probability and statistics, this course is an extension of the elements of probability and statistics introduced
171:290 Model Selection Lecture II: The Akaike Information Criterion
171:290 Model Selection Lecture II: The Akaike Information Criterion Department of Biostatistics Department of Statistics and Actuarial Science August 28, 2012 Introduction AIC, the Akaike Information
5. Linear Regression
5. Linear Regression Outline.................................................................... 2 Simple linear regression 3 Linear model............................................................. 4
Example: Boats and Manatees
Figure 9-6 Example: Boats and Manatees Slide 1 Given the sample data in Table 9-1, find the value of the linear correlation coefficient r, then refer to Table A-6 to determine whether there is a significant
Airport Planning and Design. Excel Solver
Airport Planning and Design Excel Solver Dr. Antonio A. Trani Professor of Civil and Environmental Engineering Virginia Polytechnic Institute and State University Blacksburg, Virginia Spring 2012 1 of
Practical Guide to the Simplex Method of Linear Programming
Practical Guide to the Simplex Method of Linear Programming Marcel Oliver Revised: April, 0 The basic steps of the simplex algorithm Step : Write the linear programming problem in standard form Linear
5. Multiple regression
5. Multiple regression QBUS6840 Predictive Analytics https://www.otexts.org/fpp/5 QBUS6840 Predictive Analytics 5. Multiple regression 2/39 Outline Introduction to multiple linear regression Some useful
The Loss in Efficiency from Using Grouped Data to Estimate Coefficients of Group Level Variables. Kathleen M. Lang* Boston College.
The Loss in Efficiency from Using Grouped Data to Estimate Coefficients of Group Level Variables Kathleen M. Lang* Boston College and Peter Gottschalk Boston College Abstract We derive the efficiency loss
On Marginal Effects in Semiparametric Censored Regression Models
On Marginal Effects in Semiparametric Censored Regression Models Bo E. Honoré September 3, 2008 Introduction It is often argued that estimation of semiparametric censored regression models such as the
Univariate Regression
Univariate Regression Correlation and Regression The regression line summarizes the linear relationship between 2 variables Correlation coefficient, r, measures strength of relationship: the closer r is
D-optimal plans in observational studies
D-optimal plans in observational studies Constanze Pumplün Stefan Rüping Katharina Morik Claus Weihs October 11, 2005 Abstract This paper investigates the use of Design of Experiments in observational
Weight of Evidence Module
Formula Guide The purpose of the Weight of Evidence (WoE) module is to provide flexible tools to recode the values in continuous and categorical predictor variables into discrete categories automatically,
Regression III: Advanced Methods
Lecture 16: Generalized Additive Models Regression III: Advanced Methods Bill Jacoby Michigan State University http://polisci.msu.edu/jacoby/icpsr/regress3 Goals of the Lecture Introduce Additive Models
Chapter 6: Multivariate Cointegration Analysis
Chapter 6: Multivariate Cointegration Analysis 1 Contents: Lehrstuhl für Department Empirische of Wirtschaftsforschung Empirical Research and und Econometrics Ökonometrie VI. Multivariate Cointegration
FULLY MODIFIED OLS FOR HETEROGENEOUS COINTEGRATED PANELS
FULLY MODIFIED OLS FOR HEEROGENEOUS COINEGRAED PANELS Peter Pedroni ABSRAC his chapter uses fully modified OLS principles to develop new methods for estimating and testing hypotheses for cointegrating
ECON 142 SKETCH OF SOLUTIONS FOR APPLIED EXERCISE #2
University of California, Berkeley Prof. Ken Chay Department of Economics Fall Semester, 005 ECON 14 SKETCH OF SOLUTIONS FOR APPLIED EXERCISE # Question 1: a. Below are the scatter plots of hourly wages
2. What is the general linear model to be used to model linear trend? (Write out the model) = + + + or
Simple and Multiple Regression Analysis Example: Explore the relationships among Month, Adv.$ and Sales $: 1. Prepare a scatter plot of these data. The scatter plots for Adv.$ versus Sales, and Month versus
The Power of the KPSS Test for Cointegration when Residuals are Fractionally Integrated
The Power of the KPSS Test for Cointegration when Residuals are Fractionally Integrated Philipp Sibbertsen 1 Walter Krämer 2 Diskussionspapier 318 ISNN 0949-9962 Abstract: We show that the power of the
Lasso on Categorical Data
Lasso on Categorical Data Yunjin Choi, Rina Park, Michael Seo December 14, 2012 1 Introduction In social science studies, the variables of interest are often categorical, such as race, gender, and nationality.
MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.
MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column
Simultaneous Prediction of Actual and Average Values of Study Variable Using Stein-rule Estimators
Shalabh & Christian Heumann Simultaneous Prediction of Actual and Average Values of Study Variable Using Stein-rule Estimators Technical Report Number 104, 2011 Department of Statistics University of Munich
Data Mining and Data Warehousing. Henryk Maciejewski. Data Mining Predictive modelling: regression
Data Mining and Data Warehousing Henryk Maciejewski Data Mining Predictive modelling: regression Algorithms for Predictive Modelling Contents Regression Classification Auxiliary topics: Estimation of prediction
A Subset-Continuous-Updating Transformation on GMM Estimators for Dynamic Panel Data Models
Article A Subset-Continuous-Updating Transformation on GMM Estimators for Dynamic Panel Data Models Richard A. Ashley 1, and Xiaojin Sun 2,, 1 Department of Economics, Virginia Tech, Blacksburg, VA 24060;
Introduction to Matrix Algebra
Psychology 7291: Multivariate Statistics (Carey) 8/27/98 Matrix Algebra - 1 Introduction to Matrix Algebra Definitions: A matrix is a collection of numbers ordered by rows and columns. It is customary
Unit 31 A Hypothesis Test about Correlation and Slope in a Simple Linear Regression
Unit 31 A Hypothesis Test about Correlation and Slope in a Simple Linear Regression Objectives: To perform a hypothesis test concerning the slope of a least squares line To recognize that testing for a
REGRESSION MODEL OF SALES VOLUME FROM WHOLESALE WAREHOUSE
REGRESSION MODEL OF SALES VOLUME FROM WHOLESALE WAREHOUSE Diana Santalova Faculty of Transport and Mechanical Engineering Riga Technical University 1 Kalku Str., LV-1658, Riga, Latvia E-mail: [email protected]
Association Between Variables
Contents 11 Association Between Variables 767 11.1 Introduction............................ 767 11.1.1 Measure of Association................. 768 11.1.2 Chapter Summary.................... 769 11.2 Chi
Linear Programming Notes V Problem Transformations
Linear Programming Notes V Problem Transformations 1 Introduction Any linear programming problem can be rewritten in either of two standard forms. In the first form, the objective is to maximize, the material
Stochastic Inventory Control
Chapter 3 Stochastic Inventory Control 1 In this chapter, we consider in much greater details certain dynamic inventory control problems of the type already encountered in section 1.3. In addition to the
Date: April 12, 2001. Contents
2 Lagrange Multipliers Date: April 12, 2001 Contents 2.1. Introduction to Lagrange Multipliers......... p. 2 2.2. Enhanced Fritz John Optimality Conditions...... p. 12 2.3. Informative Lagrange Multipliers...........
Section 6.1 - Inner Products and Norms
Section 6.1 - Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,
PART 1: INSTRUCTOR INFORMATION, COURSE DESCRIPTION AND TEACHING METHODS
PART 1: INSTRUCTOR INFORMATION, COURSE DESCRIPTION AND TEACHING METHODS 1.1 General information Full course title: Quantitative Methods in Economics Type of course: Compulsory Level of course B.A. Year
DEPARTMENT OF PSYCHOLOGY UNIVERSITY OF LANCASTER MSC IN PSYCHOLOGICAL RESEARCH METHODS ANALYSING AND INTERPRETING DATA 2 PART 1 WEEK 9
DEPARTMENT OF PSYCHOLOGY UNIVERSITY OF LANCASTER MSC IN PSYCHOLOGICAL RESEARCH METHODS ANALYSING AND INTERPRETING DATA 2 PART 1 WEEK 9 Analysis of covariance and multiple regression So far in this course,
Causal Forecasting Models
CTL.SC1x -Supply Chain & Logistics Fundamentals Causal Forecasting Models MIT Center for Transportation & Logistics Causal Models Used when demand is correlated with some known and measurable environmental
Chapter 13 Introduction to Nonlinear Regression( 非 線 性 迴 歸 )
Chapter 13 Introduction to Nonlinear Regression( 非 線 性 迴 歸 ) and Neural Networks( 類 神 經 網 路 ) 許 湘 伶 Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) LR Chap 10 1 / 35 13 Examples
1 Another method of estimation: least squares
1 Another method of estimation: least squares erm: -estim.tex, Dec8, 009: 6 p.m. (draft - typos/writos likely exist) Corrections, comments, suggestions welcome. 1.1 Least squares in general Assume Y i
What is Linear Programming?
Chapter 1 What is Linear Programming? An optimization problem usually has three essential ingredients: a variable vector x consisting of a set of unknowns to be determined, an objective function of x to
POLYNOMIAL AND MULTIPLE REGRESSION. Polynomial regression used to fit nonlinear (e.g. curvilinear) data into a least squares linear regression model.
Polynomial Regression POLYNOMIAL AND MULTIPLE REGRESSION Polynomial regression used to fit nonlinear (e.g. curvilinear) data into a least squares linear regression model. It is a form of linear regression
MULTIPLE REGRESSIONS ON SOME SELECTED MACROECONOMIC VARIABLES ON STOCK MARKET RETURNS FROM 1986-2010
Advances in Economics and International Finance AEIF Vol. 1(1), pp. 1-11, December 2014 Available online at http://www.academiaresearch.org Copyright 2014 Academia Research Full Length Research Paper MULTIPLE
Review Jeopardy. Blue vs. Orange. Review Jeopardy
Review Jeopardy Blue vs. Orange Review Jeopardy Jeopardy Round Lectures 0-3 Jeopardy Round $200 How could I measure how far apart (i.e. how different) two observations, y 1 and y 2, are from each other?
GENERALIZED INTEGER PROGRAMMING
Professor S. S. CHADHA, PhD University of Wisconsin, Eau Claire, USA E-mail: [email protected] Professor Veena CHADHA University of Wisconsin, Eau Claire, USA E-mail: [email protected] GENERALIZED INTEGER
The sum of digits of polynomial values in arithmetic progressions
The sum of digits of polynomial values in arithmetic progressions Thomas Stoll Institut de Mathématiques de Luminy, Université de la Méditerranée, 13288 Marseille Cedex 9, France E-mail: [email protected]
Nominal and ordinal logistic regression
Nominal and ordinal logistic regression April 26 Nominal and ordinal logistic regression Our goal for today is to briefly go over ways to extend the logistic regression model to the case where the outcome
Multiple Linear Regression
Multiple Linear Regression A regression with two or more explanatory variables is called a multiple regression. Rather than modeling the mean response as a straight line, as in simple regression, it is
Session 7 Bivariate Data and Analysis
Session 7 Bivariate Data and Analysis Key Terms for This Session Previously Introduced mean standard deviation New in This Session association bivariate analysis contingency table co-variation least squares
Section 1: Simple Linear Regression
Section 1: Simple Linear Regression Carlos M. Carvalho The University of Texas McCombs School of Business http://faculty.mccombs.utexas.edu/carlos.carvalho/teaching/ 1 Regression: General Introduction
Least-Squares Intersection of Lines
Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a
3.1 Least squares in matrix form
118 3 Multiple Regression 3.1 Least squares in matrix form E Uses Appendix A.2 A.4, A.6, A.7. 3.1.1 Introduction More than one explanatory variable In the foregoing chapter we considered the simple regression
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
Multiple Linear Regression in Data Mining
Multiple Linear Regression in Data Mining Contents 2.1. A Review of Multiple Linear Regression 2.2. Illustration of the Regression Process 2.3. Subset Selection in Linear Regression 1 2 Chap. 2 Multiple
Statistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
Multi-variable Calculus and Optimization
Multi-variable Calculus and Optimization Dudley Cooke Trinity College Dublin Dudley Cooke (Trinity College Dublin) Multi-variable Calculus and Optimization 1 / 51 EC2040 Topic 3 - Multi-variable Calculus
Least Squares Estimation
Least Squares Estimation SARA A VAN DE GEER Volume 2, pp 1041 1045 in Encyclopedia of Statistics in Behavioral Science ISBN-13: 978-0-470-86080-9 ISBN-10: 0-470-86080-4 Editors Brian S Everitt & David
UNIVERSITY OF WAIKATO. Hamilton New Zealand
UNIVERSITY OF WAIKATO Hamilton New Zealand Can We Trust Cluster-Corrected Standard Errors? An Application of Spatial Autocorrelation with Exact Locations Known John Gibson University of Waikato Bonggeun
Linear Programming. March 14, 2014
Linear Programming March 1, 01 Parts of this introduction to linear programming were adapted from Chapter 9 of Introduction to Algorithms, Second Edition, by Cormen, Leiserson, Rivest and Stein [1]. 1
3. Regression & Exponential Smoothing
3. Regression & Exponential Smoothing 3.1 Forecasting a Single Time Series Two main approaches are traditionally used to model a single time series z 1, z 2,..., z n 1. Models the observation z t as a
Some probability and statistics
Appendix A Some probability and statistics A Probabilities, random variables and their distribution We summarize a few of the basic concepts of random variables, usually denoted by capital letters, X,Y,
This chapter will demonstrate how to perform multiple linear regression with IBM SPSS
CHAPTER 7B Multiple Regression: Statistical Methods Using IBM SPSS This chapter will demonstrate how to perform multiple linear regression with IBM SPSS first using the standard method and then using the
160 CHAPTER 4. VECTOR SPACES
160 CHAPTER 4. VECTOR SPACES 4. Rank and Nullity In this section, we look at relationships between the row space, column space, null space of a matrix and its transpose. We will derive fundamental results
Causes of Inflation in the Iranian Economy
Causes of Inflation in the Iranian Economy Hamed Armesh* and Abas Alavi Rad** It is clear that in the nearly last four decades inflation is one of the important problems of Iranian economy. In this study,
MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS
MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS MSR = Mean Regression Sum of Squares MSE = Mean Squared Error RSS = Regression Sum of Squares SSE = Sum of Squared Errors/Residuals α = Level of Significance
MULTIPLE REGRESSION WITH CATEGORICAL DATA
DEPARTMENT OF POLITICAL SCIENCE AND INTERNATIONAL RELATIONS Posc/Uapp 86 MULTIPLE REGRESSION WITH CATEGORICAL DATA I. AGENDA: A. Multiple regression with categorical variables. Coding schemes. Interpreting
A SURVEY ON CONTINUOUS ELLIPTICAL VECTOR DISTRIBUTIONS
A SURVEY ON CONTINUOUS ELLIPTICAL VECTOR DISTRIBUTIONS Eusebio GÓMEZ, Miguel A. GÓMEZ-VILLEGAS and J. Miguel MARÍN Abstract In this paper it is taken up a revision and characterization of the class of
Chapter 6. Linear Programming: The Simplex Method. Introduction to the Big M Method. Section 4 Maximization and Minimization with Problem Constraints
Chapter 6 Linear Programming: The Simplex Method Introduction to the Big M Method In this section, we will present a generalized version of the simplex method that t will solve both maximization i and
Multiple Regression in SPSS This example shows you how to perform multiple regression. The basic command is regression : linear.
Multiple Regression in SPSS This example shows you how to perform multiple regression. The basic command is regression : linear. In the main dialog box, input the dependent variable and several predictors.
Robust procedures for Canadian Test Day Model final report for the Holstein breed
Robust procedures for Canadian Test Day Model final report for the Holstein breed J. Jamrozik, J. Fatehi and L.R. Schaeffer Centre for Genetic Improvement of Livestock, University of Guelph Introduction
Empirical Model-Building and Response Surfaces
Empirical Model-Building and Response Surfaces GEORGE E. P. BOX NORMAN R. DRAPER Technische Universitat Darmstadt FACHBEREICH INFORMATIK BIBLIOTHEK Invortar-Nf.-. Sachgsbiete: Standort: New York John Wiley
Section 14 Simple Linear Regression: Introduction to Least Squares Regression
Slide 1 Section 14 Simple Linear Regression: Introduction to Least Squares Regression There are several different measures of statistical association used for understanding the quantitative relationship
Business Statistics. Successful completion of Introductory and/or Intermediate Algebra courses is recommended before taking Business Statistics.
Business Course Text Bowerman, Bruce L., Richard T. O'Connell, J. B. Orris, and Dawn C. Porter. Essentials of Business, 2nd edition, McGraw-Hill/Irwin, 2008, ISBN: 978-0-07-331988-9. Required Computing
From the help desk: Swamy s random-coefficients model
The Stata Journal (2003) 3, Number 3, pp. 302 308 From the help desk: Swamy s random-coefficients model Brian P. Poi Stata Corporation Abstract. This article discusses the Swamy (1970) random-coefficients
Chapter 4: Statistical Hypothesis Testing
Chapter 4: Statistical Hypothesis Testing Christophe Hurlin November 20, 2015 Christophe Hurlin () Advanced Econometrics - Master ESA November 20, 2015 1 / 225 Section 1 Introduction Christophe Hurlin
LOGISTIC REGRESSION. Nitin R Patel. where the dependent variable, y, is binary (for convenience we often code these values as
LOGISTIC REGRESSION Nitin R Patel Logistic regression extends the ideas of multiple linear regression to the situation where the dependent variable, y, is binary (for convenience we often code these values
Department of Economics
Department of Economics On Testing for Diagonality of Large Dimensional Covariance Matrices George Kapetanios Working Paper No. 526 October 2004 ISSN 1473-0278 On Testing for Diagonality of Large Dimensional
Computer Handholders Investment Software Research Paper Series TAILORING ASSET ALLOCATION TO THE INDIVIDUAL INVESTOR
Computer Handholders Investment Software Research Paper Series TAILORING ASSET ALLOCATION TO THE INDIVIDUAL INVESTOR David N. Nawrocki -- Villanova University ABSTRACT Asset allocation has typically used
One-Way Analysis of Variance: A Guide to Testing Differences Between Multiple Groups
One-Way Analysis of Variance: A Guide to Testing Differences Between Multiple Groups In analysis of variance, the main research question is whether the sample means are from different populations. The
CHAPTER 13 SIMPLE LINEAR REGRESSION. Opening Example. Simple Regression. Linear Regression
Opening Example CHAPTER 13 SIMPLE LINEAR REGREION SIMPLE LINEAR REGREION! Simple Regression! Linear Regression Simple Regression Definition A regression model is a mathematical equation that descries the
Module 5: Multiple Regression Analysis
Using Statistical Data Using to Make Statistical Decisions: Data Multiple to Make Regression Decisions Analysis Page 1 Module 5: Multiple Regression Analysis Tom Ilvento, University of Delaware, College
Teaching model: C1 a. General background: 50% b. Theory-into-practice/developmental 50% knowledge-building: c. Guided academic activities:
1. COURSE DESCRIPTION Degree: Double Degree: Derecho y Finanzas y Contabilidad (English teaching) Course: STATISTICAL AND ECONOMETRIC METHODS FOR FINANCE (Métodos Estadísticos y Econométricos en Finanzas
Factor Analysis. Factor Analysis
Factor Analysis Principal Components Analysis, e.g. of stock price movements, sometimes suggests that several variables may be responding to a small number of underlying forces. In the factor model, we
