Topic 1 Simultaneous Equations I: Introduction

Size: px
Start display at page:

Download "Topic 1 Simultaneous Equations I: Introduction"

Transcription

1 ECONOMETRICS II Topic 1 Simultaneous Equations I: Introduction These slides are copyright 2010 by Tavis Barr. This work is licensed under a Creative Commons Attribution- ShareAlike 3.0 Unported License. See for further information.

2 Topic Outline Introduction Systems of Exogenous Regressors Seemingly Unrelated Regressions Singular Systems Identification Through a Likelihood Function Identification of a Linear System The Rank and Order Conditions Identification Through Error Term Restrictions

3 Introduction So far, we have dealt with systems of the form y = f(x) + ε Here, y is a vector of observations of one variable, whereas X is a matrix of observations of several variables Mostly, this has taken the linear form y = Xβ + ε

4 Introduction In this topic, we will consider systems of the form f m (Y,X,ε m ) = 0, for m = {1,2,...M} Here, Y is a matrix of observations of several variables (M in number) just like X (still k in number) In general, we will have M equations describing the relationship between X and Y

5 Introduction In this topic, we will consider systems of the form f m (Y,X,ε m ) = 0, for m = {1,2,...M} Mostly, this will take the linear form YΓ = XΒ + ε

6 Introduction When we write the system equations in matrix format, we assume that they are stacked and sorted by equation, i.e., X, Y, 1 =0 f1 f 2 X, Y, 2 f M X, Y, M =0

7 Introduction In this topic, we will consider systems of the form f m (Y,X,ε m ) = 0, for m = {1,2,...M} Calling Y endogenous and X exogenous is essentially saying that y m is stochastic given X and Y ~m We will consider ε m as the stochastic component that determines y m (or some deterministic function of it) given the other variables

8 Systems of Exogenous Regressors Suppose we have a system of the form: 1 y1=x1 1 y 2 =X 2 2 M 2 y M =X M M y does not appear on the right-hand side The variables X may or may not be the same in each equation

9 Systems of Exogenous Regressors Suppose we have a system of the form: 1 y1=x1 1 y 2 =X 2 2 M 2 y M =X M M ε has covariance Σ I N, Σ is unrestricted Such a system is called a system of seemingly unrelated regressions

10 Systems of Exogenous Regressors Suppose we have a system of the form: Example: y1=x1 1 1 y 2 =X 2 2 M 2 where y M =X M M E =0 E ' = I N y is industry output for M industries X includes usage of labor, capital, energy, etc. We have N time periods for each industry

11 Systems of Exogenous Regressors Suppose we have a system of the form: y1=x1 1 1 y 2 =X 2 2 M 2 where y M =X M M E =0 E ' = I N Within each equation, the model does not violate the Gauss-Markov assumptions Therefore, OLS works fine

12 Systems of Exogenous Regressors Suppose we have a system of the form: y1=x1 1 1 y 2 =X 2 2 M 2 where y M =X M M E =0 E ' = I N Suppose we want to try to improve on OLS by using a GLS estimator: = X ' I N 1 X 1 X ' I N 1 y

13 Systems of Exogenous Regressors Suppose we want to try to improve on OLS by using a GLS estimator If the exogenous variables in all equations are the same, then this reduces to: [ 11 X ' X 1 12 X ' X 1 1M 1][ X ' X X m=1 21 X ' X 1 22 X ' X 1 2M X ' X M X ' m=1 M1 X ' X 1 M2 X ' X 1 MM X ' M X X ' m=1 ' M 1 1m y m m] 1 2m y m 1 Mm y

14 Systems of Exogenous Regressors If the exogenous variables in all equations are the same, GLS reduces to: [ 11 X ' X 1 12 X ' X 1 1M 1][ X ' X X m=1 21 X ' X 1 22 X ' X 1 2M X ' X M X ' m=1 M1 X ' X 1 M2 X ' X 1 MM X ' M X X ' m=1 Consider the top row of the product: M l=1 1l X ' X 1 M m=1 lm 1 X ' y m ' M 1 1m y m m] 1 2m y m 1 Mm y

15 Systems of Exogenous Regressors Consider the top row of the product: M l=1 1l X ' X 1 M = X ' X 1 X ' l=1 M m=1 M 1l m=1 = X ' X 1 X ' y 1 lm 1 X ' y m lm 1 y m In other words, it's identical to equation-byequation OLS

16 Systems of Exogenous Regressors If exogenous variables are the same in all equations, SUR is identical to equation-byequation OLS If error terms are unrelated across equations, this is obviously also true Evidence suggests that there may be improvements when exogenous variables are less related across equations

17 Systems of Exogenous Regressors Sometimes we have a system of equations where the regressors have to add up to a constant across the equations If a coefficient each equation is a cost share for one input, cost shares have to add up to one Similarly, if coefficient in each equation represents consumer expenditure shares on each good, expenditure shares must sum to one Such a system of equations is called a singular system because Σ will be singular

18 Systems of Exogenous Regressors In a system of equations where β is constrained, Σ will be singular Example: Translog Cost Function We have cost C(Y,p) = Σ m=1,m p m x(y,p) Shepard's Lemma tells us x i = C/ p i In logs, this derivative is C/ p i = p i x i / C= s i If we have CRS, then this is invariant to Y Suppose we take a Taylor series approximation of C(p), in logs, around p=0

19 Systems of Exogenous Regressors Example: Translog Cost Function Shepard's Lemma says lnc/ lnp i = p i x i / C= s i Suppose we take a Taylor series approximation of C(p), in logs, around p=0: M lnc ln m=1 C ln p m lnp m 1 2 l=1 M M m=1 Here the RHS is a total derivative lnc/ lnp i 2 lnc ln p l ln p ln p ln p l m m We can write this in coefficients as: M M M ln C /Y = m ln p 1 m lm m=1 l=1 m=1 2 ln p ln p l m

20 Systems of Exogenous Regressors Example: Translog Cost Function Shepard's Lemma says lnc/ lnp i = p i x i / C= s i We approximate the total derivative lnc/ lnp i : M ln C /Y = m=1 M M m ln p m l=1 m=1 lm 1 2 ln p l ln p m Another way of stating this: M M s m = m ln p m l=1 m=1 lm 1 2 ln p l This gives us M equations of cost shares

21 Systems of Exogenous Regressors Example: Translog Cost Function Another way of writing the cost function: M M s m = m ln p m l=1 m=1 lm 1 2 ln p l But if these are really cost shares, then: M m=1 M m =1 ; m=1 M lm =0 ; m=1 ml =0 Therefore, system is constrained and covariance matrix is singular

22 Systems of Exogenous Regressors How to solve a singular system? Estimate M-1 equations, infer the last equation's coefficients from other equations Better to do by ML; then we can substitute M 1 M =1 m m=1 directly into the likelihood function for the Mth equation, and we don't have to drop it

23 Systems of Exogenous Regressors The bottom line: Having multiple equations does not in itself create problems for OLS There may be interesting challenges from cross-equation restrictions What makes the remaining material in this topic different is not multiple equations, but multiple endogenous variables

24 Mostly, we will study a system of equations in its linear form YΓ = XΒ + ε We assume that both Β and Γ are nonsingular This implies that there is a unique transformation Γ -1 such that Y = -XΒΓ -1 +εγ -1

25 With a linear system, there will be a unique transformation Γ -1 : Y = -XΒΓ -1 + εγ -1 In other words, there is a unique error term εγ -1 that determines y m given X and Y ~m Therefore, a linear model (and many nonlinear models as well) requires that there are as many equations as endogenous variables

26 Remember the general definition of identification: A model with parameters θ is identified if, given a empirical relationship with a known unique solution f(x, y,β) = 0, there is a one-to-one function g such that θ = g(β) It is underidentified if there exists more than one such g for some admissible values of β

27 Remember the general definition of identification: A model with parameters θ is identified if, given a empirical relationship with a known unique solution f(x, y,β) = 0, there is a one-to-one function g such that θ = g(β) It is overidentified if there exists no such g for some admissible values of β

28 Here, with a system of multiple equations, we would say it is identified if a one-to-one function g: θ = g(β) exists given the system of equations with a known unique solution f m (Y,X,ε m ) = 0, for m = {1,2,...M}

29 Here, with a system of multiple equations, we would say it is identified if a one-to-one function g: c = g(β) exists given the system of equations with a known unique solution f m (Y,X,ε m ) = 0, for m = {1,2,...M} Example: If f is linear, then we have the reduced form y m = XΠ m + ω m, m = {1,2,...M}

30 Example: If f is linear, then we have the reduced form y m = XΠ m + ω m, m = {1,2,...M} The question then becomes: Is there a unique transformation g such that g(y m ) = g(xπ m ) + g(ω m ) Obviously, without further restrictions on g, it is not unique

31 An example of an unidentified model: y 1 = y 2 β 1 x 1 β 2 ε 1 y 2 = x 1 β 3 ε 2 In this case, the first equation would show us: y 1 = x 1 β 3 ε 2 β 1 x 1 β 2 ε 1

32 An example of an unidentified model: y 1 = y 2 β 1 x 1 β 2 ε 1 y 2 = x 1 β 3 ε 2 which implies y 1 = x 1 β 3 ε 2 β 1 x 1 β 2 ε 1 The fact that y 1 contains x 1 is not a problem But y 2 includes ε 2, which might be correlated with ε 1

33 An example of an unidentified model: y 1 = y 2 β 1 x 1 β 2 ε 1 y 2 = x 1 β 3 ε 2 which implies y 1 = x 1 β 3 ε 2 β 1 x 1 β 2 ε 1 Here, y 2 includes ε 2, which might be correlated with ε 1 In general, identifying a system means distinguishing the error processes that generate each variable

34 Another example of an unidentified model: q = β 0 p β 1 ε 1 q = β 2 p β 3 ε 2 This is meant to represent a supply and demand system, with one equation representing supply and the other representing demand Using OLS, both equations will result in the same coefficients

35 Another example of an unidentified model: q = β 0 p β 1 ε 1 q = β 2 p β 3 ε 2 Using OLS, both equations will result in the same coefficients To identify which is which, we need to separate the supply shocks from the demand shocks

36 Identification via the likelihood function is more straight-forward to understand, though usually more difficult to implement Let f m (Y,X,ε m ) = 0, for m = {1,2,...M} be a system of equations Let L(θ;Y,X) be the likelihood function for the parameters of the model Then, the system is identified if, at the maximum likelihood L *, there is a unique θ such that L(θ;Y,X) = L *

37 Let L(θ;Y,X) be the likelihood function for the parameters of the model Then, the system is identified if, at the maximum likelihood L *, there is a unique θ such that L(θ;Y,X) = L * This generalizes to any pseudo-likelihood function L whereby the maximum of L(θ;Y,X) over possible values of θ is a consistent estimator of θ

38 A system is identified if, at the maximum likelihood L *, there is a unique θ such that L(θ;Y,X) = L * Notice that the likelihood function, given Y and X, is derived from the probability of observing the error terms given the parameters of the model So this requirement in one sense states that the error terms, as specified by the likelihood function, are clarified enough to identify one model from the other

39 Identification of a linear system: We have a non-identified reduced form model: y m = XΠ m + ω m We then have a (hopefully) identified, structural form model: y m = YΓ + XΒ + ε m

40 Identification of a linear system: We then have a (hopefully) identified, structural form model: y m = YΓ + XΒ + ε m We will sometimes use the equivalent model YΓ * = XΒ + ε m where Γ * = I - Γ

41 Identification of a linear system: We are attempting to transform an observable reduced form model into an identified structural form model: y m = XΠ m + ω m Y m Γ * m = X m Β m + ε m This implies that there must be a unique transformation Γ from the reduced form to the structural form such that Β m = ΠΓ * m

42 We are attempting to transform an observable reduced form model into an identified structural form model: y m = XΠ m + ω m y m = YΓ + XΒ + ε m Note the implicit assumption that γ mm = 0 and that y m appears with a coefficient of 1 in the mth equation We can dispense with this assumption, but not without imposing an equivalently strong normalization to replace it

43 We need a unique transformation Γ that gets us from the reduced form to the structural form: y m = XΠ m + ω m y m = YΓ + XΒ + ε m This is an M x M matrix with M(M-1) unknown terms We can identify it either through restrictions on Γ and Β, or indirectly through restrictions on Σ = ΓΩΓ, the structural covariance matrix

44 Identification of a linear system through exclusion (γ mo, β mj = 0 ): We need a unique transformation Γ that gets us from the reduced form to the structural form y m = XΠ m + ω m y m = YΓ + XΒ + ε m or YΓ * + XΠΓ * + ωγ * Consider the mth row of Γ. Suppose we declare that for a set of M * m variables y* m, the coefficient γ mo = 0

45 Identification of a linear system through exclusion (γ mo, β mj = 0 ): We need a unique transformation Γ that gets us from the reduced form to the structural form y m = XΠ m + ω m y m = YΓ + XΒ + ε m or y m = YΓ * + XΠΓ * + ωγ * Declaring β mj = 0 helps by creating restrictions on linear combinations of elements of Π How many exclusions do we need to identify the model?

46 Identification of a linear system through exclusion (γ mo, β mj = 0 ): Some notation: Label Description No. Y m Endogenous variables included M m in the mth equation Y * m Endogenous variables excluded M* m from the mth equation X m Exogenous variables included k m in the mth equation X * m Endogenous variables excluded k* m from the mth equation

47 Consider the matrix cross-product ΠΓ * for the mth equation when the coefficient on the jth exogenous variable is set to zero: * γ m o o: y 0 Y m π o j =β m j =0 Here each oth element of Γ * is being m multiplied by the jth variable in each oth equation; the sum equals the coefficient of the jth variable in Β m

48 Consider the matrix cross-product ΠΓ * for the mth equation when the coefficient on the jth exogenous variable is set to zero: * γ m o o: y 0 Y m π o j =β m j =0 This is an equation with M m unknowns, one for each non-excluded endogenous variable To solve for these unknowns, we need to exclude M m exogenous variables from the mth equation

49 Consider the matrix cross-product ΠΓ m, partitioned between excluded and nonexcluded exogenous variables: Π :x X j j m * Π j :x j X m = Γ * m B j 0 The bottom half of the system has k * m equations and M m unknowns Once this system of equations is solved for Γ * m, we can use the top half to derive Β m

50 Consider the matrix cross-product ΠΓ m, partitioned between excluded and nonexcluded exogenous variables: Π :x X j j m * Π j :x j X m = Γ * m B j 0 First condition for identification: k* m M m, i.e., the number of exogenous variables excluded is at least as great as the number of endogenous variables included This is known as the order condition

51 Consider the matrix cross-product ΠΓ m, partitioned between excluded and nonexcluded exogenous variables: Π :x X j j m * Π j :x j X m = Γ * m B j 0 Second condition for identification: The portion of Π corresponding to variables in X * has to have full rank m This is known as the rank condition

52 Considering just exclusion restrictions, we have the following three cases: 1. k* m < M m or the rank condition is not met. The equation is unidentified and the structural form cannot be estimated. 2. k* m = M m and the rank condition is met. The equation is exactly identified. 3.k* m > M m and the rank condition is met. The equation is overidentified and some of the restrictions can be tested.

53 We have just seen how a linear system can be identified based on exclusions from Β and Γ Obviously, linear restrictions such as β mj = 1 will have the same effect since they turn β mj into a known element and provide an equation in Γ * m We can also use non-linear restrictions; in this case, the rank and order conditions do not necessarily apply since we don t necessarily need M equations to solve for M unknowns

54 Additionally, we can identify based on restrictions on Σ Remember that Σ = Γ* ΩΓ * and that Ω is estimable A restriction on an element of Σ therefore acts as an equation in Γ * : σ ij = Γ * i Ω i Γ* j

55 Example: A triangular system. Suppose that: y1 y 2 y X y X y 1 32 y X y M = 0 M1 y 1 M2 y 2 M3 y M X M = M Note that this system is not identified because there are no excluded exogenous variables

56 Triangular systems appear in macroeconomics from time to time. Someone will argue that some variables are not structurally affected by others Example: A money (m) demand equation, as a function of inflation (π) and interest rates (r): t 1 r t m t t = 0 21 t 1 31 t 1 X 1 2 X 2 32 m t 3 X 3 41 t 1 42 m t 43 r t 1 4 X = 4

57 Example: A triangular system. Suppose that y1 y 2 y X1 21 y X y 1 32 y X y M = 0 M1 y 1 M2 y 2 M3 y M X M = 1 M Now, suppose that we declare that Σ is diagonal Consider a transformation of the above system by a matrix F into another admissible system It must be that F is upper triangular, since it has to produce 0's above the diagonal for the y's

58 Example: A triangular system. Suppose that y1 y 2 y X1 21 y X y 1 32 y X y M = 0 M1 y 1 M2 y 2 M3 y M X M = 1 M Now, the transformed cross-equation covariance matrix, Σ, also has to be diagonal Because Σij =0 for i j, and Σ 12 =F 2 'ΣF 1 = σ 11 f 12 = 0, it must be that f 12 = 0

59 Example: A triangular system. Suppose that y1 y 2 y X1 21 y X y 1 32 y X 3 0 y M = 0 M1 y 1 M2 y 2 M3 y M X M = 1 M 3 Now, the transformed cross-equation covariance matrix, Σ, also has to be diagonal Recursively, we can show that if F j is diagonal, then F j+1 is also diagonal

60 Example: A triangular system. Suppose that y1 y 2 y X1 21 y X y 1 32 y X 3 0 y M = 0 M1 y 1 M2 y 2 M3 y M X M = 1 M 3 Recursively, we can show that if F j is diagonal, then F j+1 is also diagonal But the diagonal elements have to be one to produce y m by itself in the mth equation, so F must be the identity matrix

61 Another way to think about identification: There are M m endogenous variables in the mth equation, each of which is potentially correlated with the error term For each one, we need an instrumental variable that is correlated with the endogenous variable, but not correlated with the error term By excluding a variable from each equation, we are assuming it is uncorrelated with the error term

62 Another way to think about identification: There are M m endogenous variables in the mth equation, each of which is potentially correlated with the error term For each one, we need an instrumental variable that is correlated with the endogenous variable, but not correlated with the error term The rank condition implies that the excluded variables will be correlated with the endogenous variables

63 Back to the supply/demand example: Suppose we are looking at the foreign demand for Ethiopian coffee We hypothesize that the supply is affected by the weather in Kaffa (w), but the demand is not: q s = β 11 β 12 p β 13 w ε 1 q d = β 21 β 22 p ε 2 In this case, the demand equation is identified, but the supply equation is not We are effectively using the weather as an instrumental variable for the price

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written

More information

SYSTEMS OF REGRESSION EQUATIONS

SYSTEMS OF REGRESSION EQUATIONS SYSTEMS OF REGRESSION EQUATIONS 1. MULTIPLE EQUATIONS y nt = x nt n + u nt, n = 1,...,N, t = 1,...,T, x nt is 1 k, and n is k 1. This is a version of the standard regression model where the observations

More information

Chapter 1. Vector autoregressions. 1.1 VARs and the identi cation problem

Chapter 1. Vector autoregressions. 1.1 VARs and the identi cation problem Chapter Vector autoregressions We begin by taking a look at the data of macroeconomics. A way to summarize the dynamics of macroeconomic data is to make use of vector autoregressions. VAR models have become

More information

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +

More information

2x + y = 3. Since the second equation is precisely the same as the first equation, it is enough to find x and y satisfying the system

2x + y = 3. Since the second equation is precisely the same as the first equation, it is enough to find x and y satisfying the system 1. Systems of linear equations We are interested in the solutions to systems of linear equations. A linear equation is of the form 3x 5y + 2z + w = 3. The key thing is that we don t multiply the variables

More information

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a

More information

INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition)

INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition) INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition) Abstract Indirect inference is a simulation-based method for estimating the parameters of economic models. Its

More information

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1. MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column

More information

Solution to Homework 2

Solution to Homework 2 Solution to Homework 2 Olena Bormashenko September 23, 2011 Section 1.4: 1(a)(b)(i)(k), 4, 5, 14; Section 1.5: 1(a)(b)(c)(d)(e)(n), 2(a)(c), 13, 16, 17, 18, 27 Section 1.4 1. Compute the following, if

More information

Solving Linear Systems, Continued and The Inverse of a Matrix

Solving Linear Systems, Continued and The Inverse of a Matrix , Continued and The of a Matrix Calculus III Summer 2013, Session II Monday, July 15, 2013 Agenda 1. The rank of a matrix 2. The inverse of a square matrix Gaussian Gaussian solves a linear system by reducing

More information

Solución del Examen Tipo: 1

Solución del Examen Tipo: 1 Solución del Examen Tipo: 1 Universidad Carlos III de Madrid ECONOMETRICS Academic year 2009/10 FINAL EXAM May 17, 2010 DURATION: 2 HOURS 1. Assume that model (III) verifies the assumptions of the classical

More information

Econometrics Simple Linear Regression

Econometrics Simple Linear Regression Econometrics Simple Linear Regression Burcu Eke UC3M Linear equations with one variable Recall what a linear equation is: y = b 0 + b 1 x is a linear equation with one variable, or equivalently, a straight

More information

Chapter 4: Vector Autoregressive Models

Chapter 4: Vector Autoregressive Models Chapter 4: Vector Autoregressive Models 1 Contents: Lehrstuhl für Department Empirische of Wirtschaftsforschung Empirical Research and und Econometrics Ökonometrie IV.1 Vector Autoregressive Models (VAR)...

More information

To give it a definition, an implicit function of x and y is simply any relationship that takes the form:

To give it a definition, an implicit function of x and y is simply any relationship that takes the form: 2 Implicit function theorems and applications 21 Implicit functions The implicit function theorem is one of the most useful single tools you ll meet this year After a while, it will be second nature to

More information

Notes on Determinant

Notes on Determinant ENGG2012B Advanced Engineering Mathematics Notes on Determinant Lecturer: Kenneth Shum Lecture 9-18/02/2013 The determinant of a system of linear equations determines whether the solution is unique, without

More information

5.5. Solving linear systems by the elimination method

5.5. Solving linear systems by the elimination method 55 Solving linear systems by the elimination method Equivalent systems The major technique of solving systems of equations is changing the original problem into another one which is of an easier to solve

More information

Lecture 3: Finding integer solutions to systems of linear equations

Lecture 3: Finding integer solutions to systems of linear equations Lecture 3: Finding integer solutions to systems of linear equations Algorithmic Number Theory (Fall 2014) Rutgers University Swastik Kopparty Scribe: Abhishek Bhrushundi 1 Overview The goal of this lecture

More information

Direct Methods for Solving Linear Systems. Matrix Factorization

Direct Methods for Solving Linear Systems. Matrix Factorization Direct Methods for Solving Linear Systems Matrix Factorization Numerical Analysis (9th Edition) R L Burden & J D Faires Beamer Presentation Slides prepared by John Carroll Dublin City University c 2011

More information

IDENTIFICATION IN A CLASS OF NONPARAMETRIC SIMULTANEOUS EQUATIONS MODELS. Steven T. Berry and Philip A. Haile. March 2011 Revised April 2011

IDENTIFICATION IN A CLASS OF NONPARAMETRIC SIMULTANEOUS EQUATIONS MODELS. Steven T. Berry and Philip A. Haile. March 2011 Revised April 2011 IDENTIFICATION IN A CLASS OF NONPARAMETRIC SIMULTANEOUS EQUATIONS MODELS By Steven T. Berry and Philip A. Haile March 2011 Revised April 2011 COWLES FOUNDATION DISCUSSION PAPER NO. 1787R COWLES FOUNDATION

More information

Introduction to Matrix Algebra

Introduction to Matrix Algebra Psychology 7291: Multivariate Statistics (Carey) 8/27/98 Matrix Algebra - 1 Introduction to Matrix Algebra Definitions: A matrix is a collection of numbers ordered by rows and columns. It is customary

More information

1 Solving LPs: The Simplex Algorithm of George Dantzig

1 Solving LPs: The Simplex Algorithm of George Dantzig Solving LPs: The Simplex Algorithm of George Dantzig. Simplex Pivoting: Dictionary Format We illustrate a general solution procedure, called the simplex algorithm, by implementing it on a very simple example.

More information

Simple Regression Theory II 2010 Samuel L. Baker

Simple Regression Theory II 2010 Samuel L. Baker SIMPLE REGRESSION THEORY II 1 Simple Regression Theory II 2010 Samuel L. Baker Assessing how good the regression equation is likely to be Assignment 1A gets into drawing inferences about how close the

More information

v w is orthogonal to both v and w. the three vectors v, w and v w form a right-handed set of vectors.

v w is orthogonal to both v and w. the three vectors v, w and v w form a right-handed set of vectors. 3. Cross product Definition 3.1. Let v and w be two vectors in R 3. The cross product of v and w, denoted v w, is the vector defined as follows: the length of v w is the area of the parallelogram with

More information

Module 3: Correlation and Covariance

Module 3: Correlation and Covariance Using Statistical Data to Make Decisions Module 3: Correlation and Covariance Tom Ilvento Dr. Mugdim Pašiƒ University of Delaware Sarajevo Graduate School of Business O ften our interest in data analysis

More information

8.1. Cramer s Rule for Solving Simultaneous Linear Equations. Introduction. Prerequisites. Learning Outcomes. Learning Style

8.1. Cramer s Rule for Solving Simultaneous Linear Equations. Introduction. Prerequisites. Learning Outcomes. Learning Style Cramer s Rule for Solving Simultaneous Linear Equations 8.1 Introduction The need to solve systems of linear equations arises frequently in engineering. The analysis of electric circuits and the control

More information

What s New in Econometrics? Lecture 8 Cluster and Stratified Sampling

What s New in Econometrics? Lecture 8 Cluster and Stratified Sampling What s New in Econometrics? Lecture 8 Cluster and Stratified Sampling Jeff Wooldridge NBER Summer Institute, 2007 1. The Linear Model with Cluster Effects 2. Estimation with a Small Number of Groups and

More information

Solving Systems of Linear Equations

Solving Systems of Linear Equations LECTURE 5 Solving Systems of Linear Equations Recall that we introduced the notion of matrices as a way of standardizing the expression of systems of linear equations In today s lecture I shall show how

More information

Chapter 6: Multivariate Cointegration Analysis

Chapter 6: Multivariate Cointegration Analysis Chapter 6: Multivariate Cointegration Analysis 1 Contents: Lehrstuhl für Department Empirische of Wirtschaftsforschung Empirical Research and und Econometrics Ökonometrie VI. Multivariate Cointegration

More information

Standard errors of marginal effects in the heteroskedastic probit model

Standard errors of marginal effects in the heteroskedastic probit model Standard errors of marginal effects in the heteroskedastic probit model Thomas Cornelißen Discussion Paper No. 320 August 2005 ISSN: 0949 9962 Abstract In non-linear regression models, such as the heteroskedastic

More information

Increasing for all. Convex for all. ( ) Increasing for all (remember that the log function is only defined for ). ( ) Concave for all.

Increasing for all. Convex for all. ( ) Increasing for all (remember that the log function is only defined for ). ( ) Concave for all. 1. Differentiation The first derivative of a function measures by how much changes in reaction to an infinitesimal shift in its argument. The largest the derivative (in absolute value), the faster is evolving.

More information

The VAR models discussed so fare are appropriate for modeling I(0) data, like asset returns or growth rates of macroeconomic time series.

The VAR models discussed so fare are appropriate for modeling I(0) data, like asset returns or growth rates of macroeconomic time series. Cointegration The VAR models discussed so fare are appropriate for modeling I(0) data, like asset returns or growth rates of macroeconomic time series. Economic theory, however, often implies equilibrium

More information

a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.

a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2. Chapter 1 LINEAR EQUATIONS 1.1 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,..., a n, b are given

More information

Least Squares Estimation

Least Squares Estimation Least Squares Estimation SARA A VAN DE GEER Volume 2, pp 1041 1045 in Encyclopedia of Statistics in Behavioral Science ISBN-13: 978-0-470-86080-9 ISBN-10: 0-470-86080-4 Editors Brian S Everitt & David

More information

8 Square matrices continued: Determinants

8 Square matrices continued: Determinants 8 Square matrices continued: Determinants 8. Introduction Determinants give us important information about square matrices, and, as we ll soon see, are essential for the computation of eigenvalues. You

More information

Abstract: We describe the beautiful LU factorization of a square matrix (or how to write Gaussian elimination in terms of matrix multiplication).

Abstract: We describe the beautiful LU factorization of a square matrix (or how to write Gaussian elimination in terms of matrix multiplication). MAT 2 (Badger, Spring 202) LU Factorization Selected Notes September 2, 202 Abstract: We describe the beautiful LU factorization of a square matrix (or how to write Gaussian elimination in terms of matrix

More information

Sensitivity Analysis 3.1 AN EXAMPLE FOR ANALYSIS

Sensitivity Analysis 3.1 AN EXAMPLE FOR ANALYSIS Sensitivity Analysis 3 We have already been introduced to sensitivity analysis in Chapter via the geometry of a simple example. We saw that the values of the decision variables and those of the slack and

More information

Linear Programming Notes V Problem Transformations

Linear Programming Notes V Problem Transformations Linear Programming Notes V Problem Transformations 1 Introduction Any linear programming problem can be rewritten in either of two standard forms. In the first form, the objective is to maximize, the material

More information

2. Linear regression with multiple regressors

2. Linear regression with multiple regressors 2. Linear regression with multiple regressors Aim of this section: Introduction of the multiple regression model OLS estimation in multiple regression Measures-of-fit in multiple regression Assumptions

More information

Typical Linear Equation Set and Corresponding Matrices

Typical Linear Equation Set and Corresponding Matrices EWE: Engineering With Excel Larsen Page 1 4. Matrix Operations in Excel. Matrix Manipulations: Vectors, Matrices, and Arrays. How Excel Handles Matrix Math. Basic Matrix Operations. Solving Systems of

More information

Using row reduction to calculate the inverse and the determinant of a square matrix

Using row reduction to calculate the inverse and the determinant of a square matrix Using row reduction to calculate the inverse and the determinant of a square matrix Notes for MATH 0290 Honors by Prof. Anna Vainchtein 1 Inverse of a square matrix An n n square matrix A is called invertible

More information

6. Cholesky factorization

6. Cholesky factorization 6. Cholesky factorization EE103 (Fall 2011-12) triangular matrices forward and backward substitution the Cholesky factorization solving Ax = b with A positive definite inverse of a positive definite matrix

More information

15.062 Data Mining: Algorithms and Applications Matrix Math Review

15.062 Data Mining: Algorithms and Applications Matrix Math Review .6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop

More information

1 Review of Least Squares Solutions to Overdetermined Systems

1 Review of Least Squares Solutions to Overdetermined Systems cs4: introduction to numerical analysis /9/0 Lecture 7: Rectangular Systems and Numerical Integration Instructor: Professor Amos Ron Scribes: Mark Cowlishaw, Nathanael Fillmore Review of Least Squares

More information

SYSTEMS OF EQUATIONS AND MATRICES WITH THE TI-89. by Joseph Collison

SYSTEMS OF EQUATIONS AND MATRICES WITH THE TI-89. by Joseph Collison SYSTEMS OF EQUATIONS AND MATRICES WITH THE TI-89 by Joseph Collison Copyright 2000 by Joseph Collison All rights reserved Reproduction or translation of any part of this work beyond that permitted by Sections

More information

Marketing Mix Modelling and Big Data P. M Cain

Marketing Mix Modelling and Big Data P. M Cain 1) Introduction Marketing Mix Modelling and Big Data P. M Cain Big data is generally defined in terms of the volume and variety of structured and unstructured information. Whereas structured data is stored

More information

Can we rely upon fiscal policy estimates in countries with a tax evasion of 15% of GDP?

Can we rely upon fiscal policy estimates in countries with a tax evasion of 15% of GDP? Can we rely upon fiscal policy estimates in countries with a tax evasion of 15% of GDP? Raffaella Basile, Ministry of Economy and Finance, Dept. of Treasury Bruno Chiarini, University of Naples Parthenope,

More information

Operation Count; Numerical Linear Algebra

Operation Count; Numerical Linear Algebra 10 Operation Count; Numerical Linear Algebra 10.1 Introduction Many computations are limited simply by the sheer number of required additions, multiplications, or function evaluations. If floating-point

More information

CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES. From Exploratory Factor Analysis Ledyard R Tucker and Robert C.

CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES. From Exploratory Factor Analysis Ledyard R Tucker and Robert C. CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES From Exploratory Factor Analysis Ledyard R Tucker and Robert C MacCallum 1997 180 CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES In

More information

Chapter 3: The Multiple Linear Regression Model

Chapter 3: The Multiple Linear Regression Model Chapter 3: The Multiple Linear Regression Model Advanced Econometrics - HEC Lausanne Christophe Hurlin University of Orléans November 23, 2013 Christophe Hurlin (University of Orléans) Advanced Econometrics

More information

Estimating an ARMA Process

Estimating an ARMA Process Statistics 910, #12 1 Overview Estimating an ARMA Process 1. Main ideas 2. Fitting autoregressions 3. Fitting with moving average components 4. Standard errors 5. Examples 6. Appendix: Simple estimators

More information

NOTES ON LINEAR TRANSFORMATIONS

NOTES ON LINEAR TRANSFORMATIONS NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all

More information

Linear Algebra Notes

Linear Algebra Notes Linear Algebra Notes Chapter 19 KERNEL AND IMAGE OF A MATRIX Take an n m matrix a 11 a 12 a 1m a 21 a 22 a 2m a n1 a n2 a nm and think of it as a function A : R m R n The kernel of A is defined as Note

More information

Factor Analysis. Principal components factor analysis. Use of extracted factors in multivariate dependency models

Factor Analysis. Principal components factor analysis. Use of extracted factors in multivariate dependency models Factor Analysis Principal components factor analysis Use of extracted factors in multivariate dependency models 2 KEY CONCEPTS ***** Factor Analysis Interdependency technique Assumptions of factor analysis

More information

Review of Fundamental Mathematics

Review of Fundamental Mathematics Review of Fundamental Mathematics As explained in the Preface and in Chapter 1 of your textbook, managerial economics applies microeconomic theory to business decision making. The decision-making tools

More information

Penalized regression: Introduction

Penalized regression: Introduction Penalized regression: Introduction Patrick Breheny August 30 Patrick Breheny BST 764: Applied Statistical Modeling 1/19 Maximum likelihood Much of 20th-century statistics dealt with maximum likelihood

More information

Lecture 15. Endogeneity & Instrumental Variable Estimation

Lecture 15. Endogeneity & Instrumental Variable Estimation Lecture 15. Endogeneity & Instrumental Variable Estimation Saw that measurement error (on right hand side) means that OLS will be biased (biased toward zero) Potential solution to endogeneity instrumental

More information

1 Teaching notes on GMM 1.

1 Teaching notes on GMM 1. Bent E. Sørensen January 23, 2007 1 Teaching notes on GMM 1. Generalized Method of Moment (GMM) estimation is one of two developments in econometrics in the 80ies that revolutionized empirical work in

More information

DETERMINANTS TERRY A. LORING

DETERMINANTS TERRY A. LORING DETERMINANTS TERRY A. LORING 1. Determinants: a Row Operation By-Product The determinant is best understood in terms of row operations, in my opinion. Most books start by defining the determinant via formulas

More information

Factor analysis. Angela Montanari

Factor analysis. Angela Montanari Factor analysis Angela Montanari 1 Introduction Factor analysis is a statistical model that allows to explain the correlations between a large number of observed correlated variables through a small number

More information

Linear Programming. March 14, 2014

Linear Programming. March 14, 2014 Linear Programming March 1, 01 Parts of this introduction to linear programming were adapted from Chapter 9 of Introduction to Algorithms, Second Edition, by Cormen, Leiserson, Rivest and Stein [1]. 1

More information

by the matrix A results in a vector which is a reflection of the given

by the matrix A results in a vector which is a reflection of the given Eigenvalues & Eigenvectors Example Suppose Then So, geometrically, multiplying a vector in by the matrix A results in a vector which is a reflection of the given vector about the y-axis We observe that

More information

Basic numerical skills: EQUATIONS AND HOW TO SOLVE THEM. x + 5 = 7 2 + 5-2 = 7-2 5 + (2-2) = 7-2 5 = 5. x + 5-5 = 7-5. x + 0 = 20.

Basic numerical skills: EQUATIONS AND HOW TO SOLVE THEM. x + 5 = 7 2 + 5-2 = 7-2 5 + (2-2) = 7-2 5 = 5. x + 5-5 = 7-5. x + 0 = 20. Basic numerical skills: EQUATIONS AND HOW TO SOLVE THEM 1. Introduction (really easy) An equation represents the equivalence between two quantities. The two sides of the equation are in balance, and solving

More information

From the help desk: Demand system estimation

From the help desk: Demand system estimation The Stata Journal (2002) 2, Number 4, pp. 403 410 From the help desk: Demand system estimation Brian P. Poi Stata Corporation Abstract. This article provides an example illustrating how to use Stata to

More information

1 Introduction to Matrices

1 Introduction to Matrices 1 Introduction to Matrices In this section, important definitions and results from matrix algebra that are useful in regression analysis are introduced. While all statements below regarding the columns

More information

Solution of Linear Systems

Solution of Linear Systems Chapter 3 Solution of Linear Systems In this chapter we study algorithms for possibly the most commonly occurring problem in scientific computing, the solution of linear systems of equations. We start

More information

Statistical Machine Learning

Statistical Machine Learning Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes

More information

1 2 3 1 1 2 x = + x 2 + x 4 1 0 1

1 2 3 1 1 2 x = + x 2 + x 4 1 0 1 (d) If the vector b is the sum of the four columns of A, write down the complete solution to Ax = b. 1 2 3 1 1 2 x = + x 2 + x 4 1 0 0 1 0 1 2. (11 points) This problem finds the curve y = C + D 2 t which

More information

UNDERIDENTIFICATION?

UNDERIDENTIFICATION? UNDERIDENTIFICATION? Manuel Arellano, Lars Peter Hansen and Enrique Sentana CEMFI Working Paper No. 0905 August 2009 CEMFI Casado del Alisal 5; 28014 Madrid Tel. (34) 914 290 551 Fax (34) 914 291 056 Internet:

More information

1 Short Introduction to Time Series

1 Short Introduction to Time Series ECONOMICS 7344, Spring 202 Bent E. Sørensen January 24, 202 Short Introduction to Time Series A time series is a collection of stochastic variables x,.., x t,.., x T indexed by an integer value t. The

More information

Do R&D or Capital Expenditures Impact Wage Inequality? Evidence from the IT Industry in Taiwan ROC

Do R&D or Capital Expenditures Impact Wage Inequality? Evidence from the IT Industry in Taiwan ROC Lai, Journal of International and Global Economic Studies, 6(1), June 2013, 48-53 48 Do R&D or Capital Expenditures Impact Wage Inequality? Evidence from the IT Industry in Taiwan ROC Yu-Cheng Lai * Shih

More information

Notes on Applied Linear Regression

Notes on Applied Linear Regression Notes on Applied Linear Regression Jamie DeCoster Department of Social Psychology Free University Amsterdam Van der Boechorststraat 1 1081 BT Amsterdam The Netherlands phone: +31 (0)20 444-8935 email:

More information

Section 10.4 Vectors

Section 10.4 Vectors Section 10.4 Vectors A vector is represented by using a ray, or arrow, that starts at an initial point and ends at a terminal point. Your textbook will always use a bold letter to indicate a vector (such

More information

MATH 551 - APPLIED MATRIX THEORY

MATH 551 - APPLIED MATRIX THEORY MATH 55 - APPLIED MATRIX THEORY FINAL TEST: SAMPLE with SOLUTIONS (25 points NAME: PROBLEM (3 points A web of 5 pages is described by a directed graph whose matrix is given by A Do the following ( points

More information

problem arises when only a non-random sample is available differs from censored regression model in that x i is also unobserved

problem arises when only a non-random sample is available differs from censored regression model in that x i is also unobserved 4 Data Issues 4.1 Truncated Regression population model y i = x i β + ε i, ε i N(0, σ 2 ) given a random sample, {y i, x i } N i=1, then OLS is consistent and efficient problem arises when only a non-random

More information

Math 4310 Handout - Quotient Vector Spaces

Math 4310 Handout - Quotient Vector Spaces Math 4310 Handout - Quotient Vector Spaces Dan Collins The textbook defines a subspace of a vector space in Chapter 4, but it avoids ever discussing the notion of a quotient space. This is understandable

More information

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B KITCHENS The equation 1 Lines in two-dimensional space (1) 2x y = 3 describes a line in two-dimensional space The coefficients of x and y in the equation

More information

FIXED EFFECTS AND RELATED ESTIMATORS FOR CORRELATED RANDOM COEFFICIENT AND TREATMENT EFFECT PANEL DATA MODELS

FIXED EFFECTS AND RELATED ESTIMATORS FOR CORRELATED RANDOM COEFFICIENT AND TREATMENT EFFECT PANEL DATA MODELS FIXED EFFECTS AND RELATED ESTIMATORS FOR CORRELATED RANDOM COEFFICIENT AND TREATMENT EFFECT PANEL DATA MODELS Jeffrey M. Wooldridge Department of Economics Michigan State University East Lansing, MI 48824-1038

More information

Introduction to the Economic Growth course

Introduction to the Economic Growth course Economic Growth Lecture Note 1. 03.02.2011. Christian Groth Introduction to the Economic Growth course 1 Economic growth theory Economic growth theory is the study of what factors and mechanisms determine

More information

Epipolar Geometry. Readings: See Sections 10.1 and 15.6 of Forsyth and Ponce. Right Image. Left Image. e(p ) Epipolar Lines. e(q ) q R.

Epipolar Geometry. Readings: See Sections 10.1 and 15.6 of Forsyth and Ponce. Right Image. Left Image. e(p ) Epipolar Lines. e(q ) q R. Epipolar Geometry We consider two perspective images of a scene as taken from a stereo pair of cameras (or equivalently, assume the scene is rigid and imaged with a single camera from two different locations).

More information

ECONOMETRIC METHODS IN MARKETING RESEARCH

ECONOMETRIC METHODS IN MARKETING RESEARCH Abstract ECONOMETRIC METHODS IN MARKETING RESEARCH Nada Pleli, Šemso Tanković Faculty of Economics, Kennedyjev trg 6, Zagreb, Croatia This article deals with contemporary econometric approach to marketing

More information

Department of Economics

Department of Economics Department of Economics On Testing for Diagonality of Large Dimensional Covariance Matrices George Kapetanios Working Paper No. 526 October 2004 ISSN 1473-0278 On Testing for Diagonality of Large Dimensional

More information

8.2. Solution by Inverse Matrix Method. Introduction. Prerequisites. Learning Outcomes

8.2. Solution by Inverse Matrix Method. Introduction. Prerequisites. Learning Outcomes Solution by Inverse Matrix Method 8.2 Introduction The power of matrix algebra is seen in the representation of a system of simultaneous linear equations as a matrix equation. Matrix algebra allows us

More information

Multiple regression - Matrices

Multiple regression - Matrices Multiple regression - Matrices This handout will present various matrices which are substantively interesting and/or provide useful means of summarizing the data for analytical purposes. As we will see,

More information

Solving DEs by Separation of Variables.

Solving DEs by Separation of Variables. Solving DEs by Separation of Variables. Introduction and procedure Separation of variables allows us to solve differential equations of the form The steps to solving such DEs are as follows: dx = gx).

More information

What is Linear Programming?

What is Linear Programming? Chapter 1 What is Linear Programming? An optimization problem usually has three essential ingredients: a variable vector x consisting of a set of unknowns to be determined, an objective function of x to

More information

Solving Systems of Linear Equations

Solving Systems of Linear Equations LECTURE 5 Solving Systems of Linear Equations Recall that we introduced the notion of matrices as a way of standardizing the expression of systems of linear equations In today s lecture I shall show how

More information

Monotonicity Hints. Abstract

Monotonicity Hints. Abstract Monotonicity Hints Joseph Sill Computation and Neural Systems program California Institute of Technology email: joe@cs.caltech.edu Yaser S. Abu-Mostafa EE and CS Deptartments California Institute of Technology

More information

Centre for Central Banking Studies

Centre for Central Banking Studies Centre for Central Banking Studies Technical Handbook No. 4 Applied Bayesian econometrics for central bankers Andrew Blake and Haroon Mumtaz CCBS Technical Handbook No. 4 Applied Bayesian econometrics

More information

Exploratory Factor Analysis

Exploratory Factor Analysis Introduction Principal components: explain many variables using few new variables. Not many assumptions attached. Exploratory Factor Analysis Exploratory factor analysis: similar idea, but based on model.

More information

Figure 1.1 Vector A and Vector F

Figure 1.1 Vector A and Vector F CHAPTER I VECTOR QUANTITIES Quantities are anything which can be measured, and stated with number. Quantities in physics are divided into two types; scalar and vector quantities. Scalar quantities have

More information

General Framework for an Iterative Solution of Ax b. Jacobi s Method

General Framework for an Iterative Solution of Ax b. Jacobi s Method 2.6 Iterative Solutions of Linear Systems 143 2.6 Iterative Solutions of Linear Systems Consistent linear systems in real life are solved in one of two ways: by direct calculation (using a matrix factorization,

More information

State Space Time Series Analysis

State Space Time Series Analysis State Space Time Series Analysis p. 1 State Space Time Series Analysis Siem Jan Koopman http://staff.feweb.vu.nl/koopman Department of Econometrics VU University Amsterdam Tinbergen Institute 2011 State

More information

Constrained optimization.

Constrained optimization. ams/econ 11b supplementary notes ucsc Constrained optimization. c 2010, Yonatan Katznelson 1. Constraints In many of the optimization problems that arise in economics, there are restrictions on the values

More information

Comments on Housing Price Booms and the Current Account

Comments on Housing Price Booms and the Current Account Comments on Housing Price Booms and the Current Account Lars Peter Hansen October 3, 2011 1 Introduction Adam et al. (2011) propose and analyze an interesting model of house price dynamics. It is pitched

More information

A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution

A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution PSYC 943 (930): Fundamentals of Multivariate Modeling Lecture 4: September

More information

Chapter 4: Statistical Hypothesis Testing

Chapter 4: Statistical Hypothesis Testing Chapter 4: Statistical Hypothesis Testing Christophe Hurlin November 20, 2015 Christophe Hurlin () Advanced Econometrics - Master ESA November 20, 2015 1 / 225 Section 1 Introduction Christophe Hurlin

More information

7 Gaussian Elimination and LU Factorization

7 Gaussian Elimination and LU Factorization 7 Gaussian Elimination and LU Factorization In this final section on matrix factorization methods for solving Ax = b we want to take a closer look at Gaussian elimination (probably the best known method

More information

CHAPTER 14 ORDINAL MEASURES OF CORRELATION: SPEARMAN'S RHO AND GAMMA

CHAPTER 14 ORDINAL MEASURES OF CORRELATION: SPEARMAN'S RHO AND GAMMA CHAPTER 14 ORDINAL MEASURES OF CORRELATION: SPEARMAN'S RHO AND GAMMA Chapter 13 introduced the concept of correlation statistics and explained the use of Pearson's Correlation Coefficient when working

More information

Linear and quadratic Taylor polynomials for functions of several variables.

Linear and quadratic Taylor polynomials for functions of several variables. ams/econ 11b supplementary notes ucsc Linear quadratic Taylor polynomials for functions of several variables. c 010, Yonatan Katznelson Finding the extreme (minimum or maximum) values of a function, is

More information

Manifold Learning Examples PCA, LLE and ISOMAP

Manifold Learning Examples PCA, LLE and ISOMAP Manifold Learning Examples PCA, LLE and ISOMAP Dan Ventura October 14, 28 Abstract We try to give a helpful concrete example that demonstrates how to use PCA, LLE and Isomap, attempts to provide some intuition

More information