# The Classical Linear Regression Model

Save this PDF as:

Size: px
Start display at page:

## Transcription

1 The Classical Linear Regression Model 1 September 2004 A. A brief review of some basic concepts associated with vector random variables Let y denote an n x l vector of random variables, i.e., y = (y 1, y 2,..., y n)'. 1. The expected value of y is defined by (1) This is sometimes denoted by the vector, so that = E(y).

2 2 2. The variance of the vector y can be defined as follows: (2)

3 3 3. The n x l vector of random variables, y, is said to be distributed as a multivariate normal with mean vector and variance covariance matrix (denoted y N(, )) if the density of y is given by (3) 2 Consider the special case where n = 1: y = y 1, = 1, =. (4) is just the normal density for a single random variable.

4 4 B. The Linear Regression Model 1. structure Consider the model defined by (5) If we want to include an intercept, define x tl = 1 for all t and obtain (6) 2. intuition for the model a. graphical analysis Consider a linear model of the relationship between y and x where x t1 = 1 and the only other variable is x 2t. Assume that there are n observations on each variable and that is not a constant but a random error that varies with t. We then have t The n observations (y 1, x 1),..., (y n, x n) are graphically depicted as (7) While a straight line obviously does not go through all the points, one could eyeball an upward sloping line that would seem to catch the gist of the relationship. b. population regression line The linear model postulates that the points y depend on x in a linear fashion, but that actual

5 observations will deviate from this linear relationship to random disturbances (errors). Thus the population regression line might fit through the scatter of points as follows. 5 The true population regression line is unknown because 1 and 2 are unknown parameters. The observations don't lie on the population line because of the random error term. If were a nonzero constant, it would be like shifting the intercept ( 1) of the line. It is usually assumed that the error terms have an expected value of zero or so that the "average" value of y corresponding to any given x lies on the population regression line. c. sample regression line A variety of techniques can be used to estimate the unknown parameters ( 1, 2) and thus estimate the unknown population regression line. The most common technique for a linear model such as this is ordinary least squares. The estimated regression line is referred to as the sample regression line. It is given by (8) (9) where e t is now an estimated disturbance or error term. The estimated error is also sometimes denoted. The residual e t is the vertical distance from the y t to the sample regression line. The sample regression line is not synonymous with the population regression line due to differences in and. The residual e is also different from and they may have different properties. In t t

6 6 the same way that are estimators of the 's, the sample regression line is an estimator of the population regression line. 3. assumptions for the model A number of assumptions can be made concerning the model. Different assumptions lead to different sampling properties for the random variables and for estimated parameters. The standard assumptions are usually some form of the following. a. y t = x t2 + x 3 t t t (linear model) (1) b. E( t x t1, x t2,...x tk) = 0 t (zero mean) (2) 2 c. Var( t x t1,...,x tk) = t (homoskedasticity) (3) d. E( ) = 0 t s (no autocorrelation) (4) t s e. x ti is a known constant (x's nonstochastic) (5a) No x i is a linear combination of the other x's 2 f. t N(0, ) (normality) (6) Rewriting 6 for each t (t = 1, 2,..., n) we obtain (5b)

7 7 (10) The system of equations (10) is equivalent to the matrix representation (11) where the matrices y and X are defined as follows: (12) The columns contain the n observations on individual variables, while the rows may represent observations for all variables at a given point in time. The parameter vector and the error are represented as (13) We can also write the model as (14)

8 8 where xt is the tth row of the matrix X. Assumptions 2-4 and 6 can be written much more compactly as (15) Thus the model can be summarized in terms of five assumptions as (16) Assumption V as written implies II and III. C. Discussion of the assumptions of the model 1. linearity The functional form is linear. This means that y is a linear function of x and, and depends on no other variables. 2. zero mean This simply implies that on the average each point y is on the regression line as in equation 8.

9 9 If the expected value were not zero, the intercept would simply adjust to account for this fact since and can be combined. 1 t 3. covariance matrix which is an identity matrix multiplied by a scalar 2 Assumption III implies that the covariance matrix of the error vector is a constant multiplied by the identity matrix. In general this covariance may be any positive definite matrix. Different assumptions about this matrix will lead to different properties of various estimators. This assumption is best discussed in two parts, homoskedasticity (constant on the diagonal) and no autocorrelation (all off-diagonal elements are zero). a. homoskedasticity Heteroskedasticity (as compared to homoskedasticity) is the case where the diagonal terms of 2 the covariance matrix are not all equal, i.e. Var ( t) for all t. This can be shown in two dimensions as follows where there is more variability with higher levels of x in this particular case.

10 10 With heteroskedasticity alone the covariance matrix is given by (17) This model will have k + n parameters and cannot be estimated using n observations unless some assumptions (restrictions) about the parameters are made. b. autocorrelation Autocorrelation is the case where the off-diagonal elements of the covariance matrix are not zero, i.e. Cov ( t, s) 0 for t s. If there is no autocorrelation, the errors are independently drawn so that the realization in one period (or experiment) does not depend on the realization in other periods. With no autocorrelation, the errors have no discernible pattern.

11 11 In the case above, positive levels of tend to be associated with positive levels and so on. With autocorrelation alone is given by (18) This model will have k (n(n-1)/2) parameters and cannot be estimated using n observations unless some assumptions (restrictions) about the parameters are made. 4. non-stochastic x, or x uncorrelated with The idea is that the realization of the error term is not dependent on the value of the x variable. If the x's are controlled as in an experiment, this is not a problem. It is also not a problem if x naturally occurs independently of the error term. For economic data, this is often a problem as the value of an x may be correlated with the error in the underlying model. For example, an individual firm may have access to information not available to the economic modeler and take this into account when making economic decisions. If differences in responses y, to a given x, differ in a cross section of firms because of this unknown information, the error postulated by the modeler that occurs for any firm may be correlated with this unknown information and also with the decision variable x of the firm. f. normality of the error terms

12 12 The other assumptions already guarantee that the error term has zero mean and a constant 2 variance. This assumption is necessary in order to perform statistical tests concerning the 2 estimated parameters using the normal distribution and related tests involving, t and F distributions. D. Estimation of the linear model using ordinary least squares 1. The least squares estimator The basic model can be written as (19) where is an nx1 vector of predicted values for the dependent variable. The sum of squared errors is defined by (20) The least squares estimator is defined as the which minimizes. A necessary condition for to be a minimum is that (21) This gives the normal equations which can then be solved to obtain the least squares estimator (22) For this solution to minimize the sums of squares

13 13 (23) must be a positive definite matrix. To see that this is true let q = c X Xc for some arbitrary nonzero vector c. Now let v = Xc be a linear combination of the columns of X. Now write q as follows (24) Unless every element of v is zero, q will be strictly positive. But if an element of v could be zero, then v would be a linear combination of the columns of X that equals zero, which contradicts the assumption that X is of full rank. Because c is arbitrary, q is positive for every non-zero c, which establishes that X X is positive definite. 2. Algebraic aspects of the least squares solution Rewrite the normal equations as follows (25) For every column of x k of X, this implies xk e = 0. If the first column of X is a column of 1s, then the following implications hold. a. The least squares residuals sum to zero. This follows from. b. The regression hyperplane passes through the means of the data. The first normal equation implies c. The mean of the fitted/predicted values from the regressions equals mean of the actual values. This follows from a because and the first normal equation implies that the sum of the y s is equal to the sum of the.

14 14 E. The fundamental matrices of regression analysis 1. M - the residual matrix The residuals from the least squares regression can be expressed as (26) The matrix M X is symmetric and idempotent. When the context is obvious, we will often drop the subscript X on M. To show that it is symmetric find its transpose. (27) To show that it is idempotent, multiply it by itself. (28) Furthermore MX = 0 as can be seen by multiplying (29) We also can express the sum of squares in terms of M and also y

15 15 (30) We can similarly express the sum of squares in terms of the population residuals as follows (31) 2. P - the projection matrix Consider a representation of the predicted value of y (32) Again when the context is obvious, we will drop the subscript X on P. P is symmetric and idempotent. When the vector y is premultiplied by P, the result is the predicted valued of y from a regression of y on X. Additionally (33) These equations imply that the projection of X onto itself is just X. The second expression implies that M and P are orthogonal. 3. A n The deviation transformation matrix Consider the matrix A n below which transforms a vector or matrix to deviations from the mean.

16 16 (34) When n is obvious it will be dropped and we will write just A. As an example consider the following. (35) Notice that A is symmetric (obvious) and idempotent.

17 17 (36) Given that A is symmetric and idempotent, the following sum of squares can be written in matrix form using a quadratic form in A. (37) Further note that the matrix A nullifies the first column of the X matrix if it is a column of ones. (38) Similarly if we premultiply A by x1. (39) So if we write the regression model partitioning X as [1 X 2 ] where X 2 is the matrix with columns x 2, x... we obtain 3 (40)

18 18 This implies that we can write the model in deviation form where each variable is written as a deviation from its mean. The vector 2 will contain all the slope coefficients from the original model. 4. Partitioning y We can partition the vector y into two orthogonal parts, the predicted value and the error vector e. To do this write y as follows and then substitute for and e. (41) To see orthogonality, write e as y M and then premultiply Py by this. This will be zero because MP = Partitioning the sum of squares of y Write out y y as follows (42) 6. Partitioning the sum of squares of e Write out e e as follows (43) where X e = 0 from equation 25. F. Partitioned regression 1. General case Suppose a regression involves two sets of variables X 1 and X 2. We can then write y as follows (44)

19 19 The matrix X 1 has n rows and k 1 columns, the matrix X 2 has n rows and k 2 = k - k 1 columns. Similarly has k elements and so on. The normal equations are as follows 1 1 (45) They imply two sets of equations, one the as the right hand side, the other with as the right hand side. We can solve the first system for as a function of X, y and. Doing so we obtain, (46) This can be interpreted as the regression of y on X 1 minus a correction vector. Notice that if X 2 is orthogonal to X, there is no correction necessary. 1 Once we have as a function of X, y and, we can solve the second system of equations for. To begin write out the system substituting for and then solve (47) The matrix M 1 is a residual maker matrix like M in the standard model. In this case it gives the residual from a regression on the columns of X 1. M1y is the vector of the residuals from the regression of y on X 1, M1X 2 is the vector of residuals in the regression of the corresponding column of X on X. Given that M is symmetric and idempotent we can rewrite equation 47 as 2 1 1

20 20 (48) where (49) Notice that because M 1 is idempotent, we get the same result whether we regress y or y* on. This result is know as the Frisch-Waugh Theorem and is as follows. Theorem 1: In the linear least squares regression of the vector y on two sets of variables X 1 and X 2, the subvector is the set of coefficients obtained when the residuals from a regression of y on X 1 alone are regressed on the set of residuals obtained when each column of X 2 is regressed on X 1. This process is commonly called partialing out or netting our the effect of X 1. As an application consider the case in which X is i, a column of ones in the first column of X. The solution for 1 this case will then be the slopes in a regression with a constant term. The coefficient in a regression -1 of any variable z on i is given by [i i] i z =. The fitted values are and the residuals are. This then implies that the slopes in a multiple regression that contains a constant term are obtained by transforming the data to deviations from their means, and then regressing the variable y in deviation form on the explanatory variables, also in deviation form. 2. Addition of a single variable to a multiple regression Consider the two alternative models in (50) The second model postulates that the variable z should be added to the model. We refer to the first model as the short model and the second model as the long model. We can also write the long model as (51)

21 21 Consider computing c in the long model from an auxiliary regression of y and z on X. Using equations 48 and 49, define the following variables (52) y* is the vector of residuals from the regression of y on X, while z* is the vector of residuals of z on Z. Now compute c from a regression of y* on z*, i.e. (53) In effect to obtain c in the long regression, we regress the residuals from the short regression on the residuals from the regression of z on X. 3. Relation of the long and short regression models a. statement of long model Regress y on [X 1 X 2]. The following expressions all hold. (54) b. statement of short model Regress y on X 1. The following expressions all hold. (55) c. auxiliary regressions

22 22 Regress each column of X 2 on X 1. There will be k 2 of these regressions. (56) is the residual left after regressing x j on X 1. In some sense it is the part of x j that is purged of X 1 (i.e., cannot be explained by X 1). We can write this system of regressions in matrix form as follows (57) The matrix F is a matrix of coefficients from regressing the columns of X 2 on the columns of X 1. We can also put the residual vectors in an n by k 2 matrix as follows (58) Now calculate using equation 55 and substituting for y from equation 54 (59) The last step follows because from equation 55. Equation 59 says that the coefficients on X 1 in the short regression are a mixture of the coefficients in the long regression on X 1 and X 2 in the long regression with regression coefficients of the regression of X 2 on X1 serving as weights in that mixture. We can also find a relationship between the residuals in the short and long regressions.

23 23 (60) The last step follows because from equation 55. Equation 60 says that the residuals from the short regression equal the residuals from the long regression plus a mixture of the elements of, with the elements of serving as weights in that mixture. We can calculate the sum of squared errors from the short regression using equation 60 and simplifying. (61) Equation 61 says that sum of squared residuals from the short regression exceeds the sum of squared residuals from the long regression by the non-negative quantity v v where. So shortening a regression cannot improve the fit. The two sums of squared residuals are equal iff and only if v = 0, that is iff equal, that is (because rank = k 2) iff = 0. If = 0, then. If, then F = C1X 2 = 0 and. If F = 0, then from 59,. But in this case, e* e. d. example of trend removal A popular specification for economic time series takes the form where t indexes the observations, x 1 = 1, x 2 = t, and x 3,...,x k are conventional explanatory variables. Here x 2 allows for a linear trend in the expected value of the dependent variable. The question arises: should the trend term be included in the regression, or should the variables first be detrended and then used without the trend terms included? In the first volume of Econometrica, Frisch and Waugh (1933) concluded that it does not matter. The residual regression results apply. Let X 1 = (x 1, x 2) and X 2 = (x 3,...,x k). The coefficients on the latter set will be the same whether we include X 1 in the regression along with X 2, or alternatively first detrend X 2 (by calculating the residuals from the regression of X on X ) and then use only those detrended explanatory variables. 2 1

24 24 G. Goodness of Fit 1. introduction and intuition The least squares estimator is constructed so that the values of chosen minimize the sum of squares of the elements of the residual vector e. These estimates also choose in a way to minimize the distance from y to the column space of the matrix X, i.e., is the projection of y on the column space of X. In this sense, is a best estimate. But the computation of says nothing about how good of an estimate it is. Consider the diagram below. In the left panel the line is close of the points, and most of them are quite close to the line in a visual sense. In the right hand panel the line seems to minimize the sum of squared distances, but a large number of the points distant are from the line. But this sample regression line (which is conditioned on x) seems to fit the data better than as a predictor of the expected value of y.

25 25 One way to think about the value of using X as a predictor of y is to consider whether the deviation of any given y from the mean of the y s is more largely accounted for by deviations of the associated x from its mean than by the residual. The following diagram when there is only one x variable is useful. If the regression has a constant term (x 1t = 1), then the mean of the predicted values of y will be the

26 26 same as the mean of y, and the regression line will pass through the mean of y and x. Writing out the regression for the tth observation and subtracting from both sides we obtain the following. (62) At issue is how much of the difference between y t and is explained by as compared to e t. Equation 62 gives us some idea of the relative magnitudes of this difference for an individual observation. We seek an overall measure of impact. Given that the sum of either side of equation 62 is zero, an alternative metric is to consider the sums of squares of the various parts of the decomposition. To start with this process multiply both sides of equation 19 by the idempotent matrix A which transforms observations into deviations from sample means. (63) The vector e already has a mean of zero so Ae = e A = e A = e. Also. The normal equations imply that X e = 0. Using this information multiply equation 63 by y and simplify. (64) Note the similarity of this equation to equation 43 which reflects sums of squares rather than sums of squared deviations. We refer to the sum of squares of the deviations from the mean of y as SST or sum of squares total. We refer to the sum of squares of the error vector as SSE or sum of squared errors. We then refer to the sum of squares of the deviations of the predicted value of y from their mean as SSR or sum of squares due to regression. Specifically

27 27 (65) From equation 31, SSE = e e = y MXy and from equation 37, SST = y Ay. We can then rewrite 65 in the following useful form (66) Now consider the following measure of goodness of fit of the regression model (67) 2 R will lie between 0 and 1. It measures the proportion of the total variation in y that is accounted 2 for by variation in the regressors. It equals zero if the regression is a horizontal line. If R = 1, then the values of y and x all lie on the same hyperplane. 2 To see the impact of adding regressors to a model, consider writing R using the information in equation 61 as follows. (68) 2 Now determine R for the regression on y on X 1 using equation 55 (which we repeat here for

28 28 convenience) and then manipulating it like we did in equation 64. (69) Multiply both sides of 69 by y A as follows (70) Now divide both sides of equation 70 by y Ay to obtain (71) where equation 68 as is the coefficient of determination for the regression of y on X. We can then write 1 (72)

29 29 So R will go up as we add variables to the model. If is zero, then R will not rise Analysis of Variance (ANOVA) We have just decomposed the total sum of squares (SST) into two components: sum of squared errors (SSE) sum of squares explained by regression (SSR). This decomposition is commonly summarized in the form of an analysis of variance (ANOVA) table, where MSE adjusts sum of squares for degrees of freedom. Source of Variation SS d.f MSE Model Error SSR SSE k - 1 n - k Total SST n - 1 k = number of coefficients in model SSR/(k - 1) SSE/(n - k) Problems with R and an alternative, adjusted R a. adjusted R 2 2 R is the fraction of total sum of squares "explained" by the model. Note that increasing the number of independent variables in the model will not change SST (which only depend on y), 2 but from equation 61 will generally decrease SSE; hence, increase R. This is true even if the additional variables are not statistically significant in explaining y. This is clear from equatoin This has provided the motivation for considering the adjusted R, instead of the R. (73) where k = the number of x variables and 's (coefficients) in the model. This expression penalizes regressions with more explanatory variables. 2 b. Problems with R is there is not a constant(intercept) term in the model 2 2 A second difficulty with R concerns the constant term in the model. The proof that 0 R 1 (in equation 64) requires X to contain a column of 1s. If not, then Ae e and e AX 0, and therefore the terms are not equal to zero, and e Ae is not equal to e e. Consequently, when we compute

30 30 the result is unpredictable. It will never be higher and can be far lower than the same figure computed for the regression with a constant term included. H. The Geometry of Least Squares 1. Introduction and idea The essential elements of a linear regression are a regressand y and a matrix of regressors X = [x 1, x 2,... x k ]. The regressand y is n 1 and the matrix X is n k. The regressand y and each of the n regressors x 1, x 2,... x k can be thought of as a point in n-dimensional Euclidean space R. If the k n regressors are linearly independent, they span a k dimensional subspace of R. This space can be n denoted S(X), the column space of the matrix X. This subspace consists of all points z in R that n can be written as linear combinations of the columns of X, i.e., all points z in R such that z = X for some k vector. The dimension of S(X) will be equal to the rank of X. We can write this in summation notation as (75) where x 1 is the first column of X and so on. As long as the columns of X are linearly independent this representation of the point z will be unique. As a matter of notation we write S(x 1, x 2) for the n subspace of R that is jointly spanned by the vectors x 1 and x 2. Similarly, S(X, U) is the subspace spanned by columns of the matrices X and U. If the matrix X is transformed by any rank preserving linear transformation, the subspace spanned by this transformation is the same as the subspace spanned by X itself. For example, we can postmultiply the matrix X by a non-singular matrix D without changing the subspace spanned by it. If X * = X and X = XA then (76) n n The orthogonal complement of S(X) in R, which is denoted S (X), is denoted set all points u in R, such that, for any z S(X), w z = 0. Every point in S (X) is orthogonal to every point in S(X). The dimension of S (X) is n-k. Consider the diagram below where n =2 and k = 1. The arrow ends at the point X. All points along the line through X belong to S(X). All points which are perpendicular to the line through the origin intersecting the line through X are elements of S (X).

31 31 Notice that S(X) and S (X) would be the same if X were any point along the straight line through X except for the origin. This illustrates that S(X) is invariant to any nonsingular transformation. 2. Least squares As we have seen, any point in S(X) can be represented by a vector of the form X for some k-vector. If one wants to find the point in S(X) that is closest to any given vector y, the problem is that of minimizing, with respect to the choice of, the distance between y and X. Minimizing this distance is equivalent to minimizing the square of this distance. Thus, solving the problem (77) where x denotes Euclidean distance will find the point in S(X) that is closest to y. This squared distance can be written equivalently as (78) where y t and xt represent the tth element of y and the tth row of respectively. This distance is the sum of square residuals of the linear regression of y on X. Consider the diagram below where is an element of S(X) and is the closest element of S(X) to y. The distance from y to is the residual in the regression. In the diagram it is the line translated so that its origin is at instead of at zero. The right angle formed by and S(X) is the key geometric feature of least squares. At any other point in S(X), does not form a right with S(X), and, as a consequence, is further from y.

32 32 is a scalar in this example. The derivative of (78) with respect to is given by (79) These normal equations say that the vector must be orthogonal to all the columns of X.. Given that the columns of X are linearly independent, X X is of full rank and can be inverted. We therefore obtain as (80) Even if X X is not of full rank, the fitted values are uniquely defined, because is the point in S(X) that is closest to y. But if when X X is not of full rank, will not unique. If we substitute for in we obtain (81) X As discussed previously, P is the n n matrix that projects the vector y orthogonally onto S(X). In the diagram below, Py is the vector from the origin to the point in S(X) closest to y.

33 33 Associated with any linear subspace such as S(X), is a matrix P X that projects points onto that subspace. When the interpretation is obvious we write P for P X. Also associated with the subspace n S(X) is a matrix M X which projects any point in R onto the orthogonal complement S (X). This matrix is given by (82) We say that S(X) is the range of P and S (X) is the range of M. Note that X X If we multiply both sides of (83) by z we obtain (83) (84) n This implies that any point z R can be written as the sum of its projection onto S(X) and its projection onto S (X). Any set of vectors that form a basis for S(X) along with vectors forming a n basis for S (X) together form a basis for R. So writing z as a linear combination of its projection and the error vector associated with the that projection is just writing z as a function of a particular basis, forming two sub sums before determining z itself. The fact that P X and M X are idempotent is now made obvious by their geometric interpretation. The projection of a vector that is already in S(X) or S (X) is obviously just itself. Similarly we can obtain a geometric interpretation of equation 33, which we repeat here (85) Equation 85 must hold because P projects onto S(X) and M projects onto S (X). The only point X X

34 that belongs to both S(X) and S (X) is the origin or zero vector. We sometimes call M X the annihilation matrix because it annihilates not only P X but all point in S(X). Similarly, P X annihilates not must M X but all points in S (X). Also notice that if we transform the matrix X with a nonsingular matrix D that P does not change. X 34 (86) X Because the transformation does not change S(X), it does not change P. This implies that is any regression we can multiply any or all of the regressors by a non-zero constant without changing the predicted (fitted) values or the ordinary least squares (OLS) residuals. If the regressors include a constant term, then we can add any constant amount to any or all of the regressors without affecting the fitted values or the OLS residuals.

35 35 3. Example problem Consider the following problem in which n = 2 and k = 1. The y vector is [2, 4] and the single column of X is [6, 2]. The goal is to project y onto the line through the origin and (6,2). We first define some of the useful quantities for this problem. (87) Now consider the least squares estimate of (88) So our least squares fitted value of y is given by (89) We can also compute using the projection matrix P, i.e., as can be easily verified my multiplication. The vector of residuals is given by

36 36 (90) In the diagram below we can see that the subspace spanned by the matrix X is the line through the origin and (6, 2). This line is S(X), the column space of X. The point y is not in this space (not on the line). The fitted value of y is (3,1). This is the point in S(X) that is closest to y. Notice that the line between (3,1) and (2,4) is perpendicular to the line that represents S(X). Also notice that the vector of residuals represents how far we must move horizontally and vertically to get from to y. We move negative one units horizontally and then 3 units vertically to get from to y. This is represented by the vector (-1, 3), which is an element of S (X). The distance from the origin to (-1,3) is the same as the distance from to y. To see this compute the length of the vector from the origin to (-1, 3), and compare it to the distance from to y.

37 37 (91) Now consider P y and M y for this problem. X x (92) The matrix P X projects y onto S(X) and the matrix M X projects y onto S (X). Consider a different 2 vector in R, say y = [4, 3]. P and M will project it into the same subspaces. X X (93) The matrix P clearly projects y onto S(X) while matrix M projects y onto S (X). X X

38 38

39 39 2 Consider a third vector in R, say y = [5, 5]. (94) The matrix P X clearly projects y onto S(X) while matrix M X projects y onto S (X), in fact on the same point as in the original example.

40 40 Ignore this page Now define a new variable w = Cy as follows (95) Note that CC' is the identity matrix (96) Now rewrite SST as (97)

41 41

### Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written

### MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a

### 3.1 Least squares in matrix form

118 3 Multiple Regression 3.1 Least squares in matrix form E Uses Appendix A.2 A.4, A.6, A.7. 3.1.1 Introduction More than one explanatory variable In the foregoing chapter we considered the simple regression

### 1 Introduction to Matrices

1 Introduction to Matrices In this section, important definitions and results from matrix algebra that are useful in regression analysis are introduced. While all statements below regarding the columns

### α = u v. In other words, Orthogonal Projection

Orthogonal Projection Given any nonzero vector v, it is possible to decompose an arbitrary vector u into a component that points in the direction of v and one that points in a direction orthogonal to v

### MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +

### Introduction to Matrix Algebra

Psychology 7291: Multivariate Statistics (Carey) 8/27/98 Matrix Algebra - 1 Introduction to Matrix Algebra Definitions: A matrix is a collection of numbers ordered by rows and columns. It is customary

### The basic unit in matrix algebra is a matrix, generally expressed as: a 11 a 12. a 13 A = a 21 a 22 a 23

(copyright by Scott M Lynch, February 2003) Brief Matrix Algebra Review (Soc 504) Matrix algebra is a form of mathematics that allows compact notation for, and mathematical manipulation of, high-dimensional

Lecture 5: Linear least-squares Regression III: Advanced Methods William G. Jacoby Department of Political Science Michigan State University http://polisci.msu.edu/jacoby/icpsr/regress3 Simple Linear Regression

### Linear Dependence Tests

Linear Dependence Tests The book omits a few key tests for checking the linear dependence of vectors. These short notes discuss these tests, as well as the reasoning behind them. Our first test checks

### Linear Algebra Review. Vectors

Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka kosecka@cs.gmu.edu http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa Cogsci 8F Linear Algebra review UCSD Vectors The length

### SYSTEMS OF REGRESSION EQUATIONS

SYSTEMS OF REGRESSION EQUATIONS 1. MULTIPLE EQUATIONS y nt = x nt n + u nt, n = 1,...,N, t = 1,...,T, x nt is 1 k, and n is k 1. This is a version of the standard regression model where the observations

### Recall that two vectors in are perpendicular or orthogonal provided that their dot

Orthogonal Complements and Projections Recall that two vectors in are perpendicular or orthogonal provided that their dot product vanishes That is, if and only if Example 1 The vectors in are orthogonal

### Review Jeopardy. Blue vs. Orange. Review Jeopardy

Review Jeopardy Blue vs. Orange Review Jeopardy Jeopardy Round Lectures 0-3 Jeopardy Round \$200 How could I measure how far apart (i.e. how different) two observations, y 1 and y 2, are from each other?

### Introduction to General and Generalized Linear Models

Introduction to General and Generalized Linear Models General Linear Models - part I Henrik Madsen Poul Thyregod Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby

### Introduction to Matrix Algebra I

Appendix A Introduction to Matrix Algebra I Today we will begin the course with a discussion of matrix algebra. Why are we studying this? We will use matrix algebra to derive the linear regression model

### 15.062 Data Mining: Algorithms and Applications Matrix Math Review

.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop

### Vector Spaces II: Finite Dimensional Linear Algebra 1

John Nachbar September 2, 2014 Vector Spaces II: Finite Dimensional Linear Algebra 1 1 Definitions and Basic Theorems. For basic properties and notation for R N, see the notes Vector Spaces I. Definition

### Orthogonal Diagonalization of Symmetric Matrices

MATH10212 Linear Algebra Brief lecture notes 57 Gram Schmidt Process enables us to find an orthogonal basis of a subspace. Let u 1,..., u k be a basis of a subspace V of R n. We begin the process of finding

### Linear Algebra Notes

Linear Algebra Notes Chapter 19 KERNEL AND IMAGE OF A MATRIX Take an n m matrix a 11 a 12 a 1m a 21 a 22 a 2m a n1 a n2 a nm and think of it as a function A : R m R n The kernel of A is defined as Note

### Linear Algebra Notes for Marsden and Tromba Vector Calculus

Linear Algebra Notes for Marsden and Tromba Vector Calculus n-dimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of

### Section 1.1. Introduction to R n

The Calculus of Functions of Several Variables Section. Introduction to R n Calculus is the study of functional relationships and how related quantities change with each other. In your first exposure to

### Systems of Linear Equations

Systems of Linear Equations Beifang Chen Systems of linear equations Linear systems A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where a, a,, a n and

### Vector algebra Christian Miller CS Fall 2011

Vector algebra Christian Miller CS 354 - Fall 2011 Vector algebra A system commonly used to describe space Vectors, linear operators, tensors, etc. Used to build classical physics and the vast majority

### Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013

Notes on Orthogonal and Symmetric Matrices MENU, Winter 201 These notes summarize the main properties and uses of orthogonal and symmetric matrices. We covered quite a bit of material regarding these topics,

### Solutions to Math 51 First Exam January 29, 2015

Solutions to Math 5 First Exam January 29, 25. ( points) (a) Complete the following sentence: A set of vectors {v,..., v k } is defined to be linearly dependent if (2 points) there exist c,... c k R, not

### Orthogonal Projections

Orthogonal Projections and Reflections (with exercises) by D. Klain Version.. Corrections and comments are welcome! Orthogonal Projections Let X,..., X k be a family of linearly independent (column) vectors

### MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column

### MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.

MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix. Nullspace Let A = (a ij ) be an m n matrix. Definition. The nullspace of the matrix A, denoted N(A), is the set of all n-dimensional column

### Inner Product Spaces

Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and

### Random Vectors and the Variance Covariance Matrix

Random Vectors and the Variance Covariance Matrix Definition 1. A random vector X is a vector (X 1, X 2,..., X p ) of jointly distributed random variables. As is customary in linear algebra, we will write

### Chapter 3: The Multiple Linear Regression Model

Chapter 3: The Multiple Linear Regression Model Advanced Econometrics - HEC Lausanne Christophe Hurlin University of Orléans November 23, 2013 Christophe Hurlin (University of Orléans) Advanced Econometrics

### Lectures notes on orthogonal matrices (with exercises) 92.222 - Linear Algebra II - Spring 2004 by D. Klain

Lectures notes on orthogonal matrices (with exercises) 92.222 - Linear Algebra II - Spring 2004 by D. Klain 1. Orthogonal matrices and orthonormal sets An n n real-valued matrix A is said to be an orthogonal

### We call this set an n-dimensional parallelogram (with one vertex 0). We also refer to the vectors x 1,..., x n as the edges of P.

Volumes of parallelograms 1 Chapter 8 Volumes of parallelograms In the present short chapter we are going to discuss the elementary geometrical objects which we call parallelograms. These are going to

### Quadratic forms Cochran s theorem, degrees of freedom, and all that

Quadratic forms Cochran s theorem, degrees of freedom, and all that Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 1 Why We Care Cochran s theorem tells us

### 3. INNER PRODUCT SPACES

. INNER PRODUCT SPACES.. Definition So far we have studied abstract vector spaces. These are a generalisation of the geometric spaces R and R. But these have more structure than just that of a vector space.

### Inner product. Definition of inner product

Math 20F Linear Algebra Lecture 25 1 Inner product Review: Definition of inner product. Slide 1 Norm and distance. Orthogonal vectors. Orthogonal complement. Orthogonal basis. Definition of inner product

### 9 Multiplication of Vectors: The Scalar or Dot Product

Arkansas Tech University MATH 934: Calculus III Dr. Marcel B Finan 9 Multiplication of Vectors: The Scalar or Dot Product Up to this point we have defined what vectors are and discussed basic notation

### Linear Codes. In the V[n,q] setting, the terms word and vector are interchangeable.

Linear Codes Linear Codes In the V[n,q] setting, an important class of codes are the linear codes, these codes are the ones whose code words form a sub-vector space of V[n,q]. If the subspace of V[n,q]

### Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Chapter 3 Linear Least Squares Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction

### Linear Algebra: Vectors

A Linear Algebra: Vectors A Appendix A: LINEAR ALGEBRA: VECTORS TABLE OF CONTENTS Page A Motivation A 3 A2 Vectors A 3 A2 Notational Conventions A 4 A22 Visualization A 5 A23 Special Vectors A 5 A3 Vector

### Simple Regression Theory II 2010 Samuel L. Baker

SIMPLE REGRESSION THEORY II 1 Simple Regression Theory II 2010 Samuel L. Baker Assessing how good the regression equation is likely to be Assignment 1A gets into drawing inferences about how close the

### Lecture L3 - Vectors, Matrices and Coordinate Transformations

S. Widnall 16.07 Dynamics Fall 2009 Lecture notes based on J. Peraire Version 2.0 Lecture L3 - Vectors, Matrices and Coordinate Transformations By using vectors and defining appropriate operations between

### Math 215 HW #6 Solutions

Math 5 HW #6 Solutions Problem 34 Show that x y is orthogonal to x + y if and only if x = y Proof First, suppose x y is orthogonal to x + y Then since x, y = y, x In other words, = x y, x + y = (x y) T

### Mathematics Course 111: Algebra I Part IV: Vector Spaces

Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 1996-7 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are

### Solving Simultaneous Equations and Matrices

Solving Simultaneous Equations and Matrices The following represents a systematic investigation for the steps used to solve two simultaneous linear equations in two unknowns. The motivation for considering

### NOTES ON LINEAR TRANSFORMATIONS

NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all

### Chapter 6. Orthogonality

6.3 Orthogonal Matrices 1 Chapter 6. Orthogonality 6.3 Orthogonal Matrices Definition 6.4. An n n matrix A is orthogonal if A T A = I. Note. We will see that the columns of an orthogonal matrix must be

### Euclidean Geometry. We start with the idea of an axiomatic system. An axiomatic system has four parts:

Euclidean Geometry Students are often so challenged by the details of Euclidean geometry that they miss the rich structure of the subject. We give an overview of a piece of this structure below. We start

### Biggar High School Mathematics Department. National 5 Learning Intentions & Success Criteria: Assessing My Progress

Biggar High School Mathematics Department National 5 Learning Intentions & Success Criteria: Assessing My Progress Expressions & Formulae Topic Learning Intention Success Criteria I understand this Approximation

### Chapter 19. General Matrices. An n m matrix is an array. a 11 a 12 a 1m a 21 a 22 a 2m A = a n1 a n2 a nm. The matrix A has n row vectors

Chapter 9. General Matrices An n m matrix is an array a a a m a a a m... = [a ij]. a n a n a nm The matrix A has n row vectors and m column vectors row i (A) = [a i, a i,..., a im ] R m a j a j a nj col

### Matrix Differentiation

1 Introduction Matrix Differentiation ( and some other stuff ) Randal J. Barnes Department of Civil Engineering, University of Minnesota Minneapolis, Minnesota, USA Throughout this presentation I have

### Torgerson s Classical MDS derivation: 1: Determining Coordinates from Euclidean Distances

Torgerson s Classical MDS derivation: 1: Determining Coordinates from Euclidean Distances It is possible to construct a matrix X of Cartesian coordinates of points in Euclidean space when we know the Euclidean

### Figure 1.1 Vector A and Vector F

CHAPTER I VECTOR QUANTITIES Quantities are anything which can be measured, and stated with number. Quantities in physics are divided into two types; scalar and vector quantities. Scalar quantities have

### Adding vectors We can do arithmetic with vectors. We ll start with vector addition and related operations. Suppose you have two vectors

1 Chapter 13. VECTORS IN THREE DIMENSIONAL SPACE Let s begin with some names and notation for things: R is the set (collection) of real numbers. We write x R to mean that x is a real number. A real number

### Vector and Matrix Norms

Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a non-empty

### Similarity and Diagonalization. Similar Matrices

MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that

### Computer Graphics Prof. Sukhendu Das Dept. of Computer Science and Engineering Indian Institute of Technology, Madras Lecture 7 Transformations in 2-D

Computer Graphics Prof. Sukhendu Das Dept. of Computer Science and Engineering Indian Institute of Technology, Madras Lecture 7 Transformations in 2-D Welcome everybody. We continue the discussion on 2D

### Module 3: Correlation and Covariance

Using Statistical Data to Make Decisions Module 3: Correlation and Covariance Tom Ilvento Dr. Mugdim Pašiƒ University of Delaware Sarajevo Graduate School of Business O ften our interest in data analysis

### CONTROLLABILITY. Chapter 2. 2.1 Reachable Set and Controllability. Suppose we have a linear system described by the state equation

Chapter 2 CONTROLLABILITY 2 Reachable Set and Controllability Suppose we have a linear system described by the state equation ẋ Ax + Bu (2) x() x Consider the following problem For a given vector x in

### NCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( )

Chapter 340 Principal Components Regression Introduction is a technique for analyzing multiple regression data that suffer from multicollinearity. When multicollinearity occurs, least squares estimates

### MATH 551 - APPLIED MATRIX THEORY

MATH 55 - APPLIED MATRIX THEORY FINAL TEST: SAMPLE with SOLUTIONS (25 points NAME: PROBLEM (3 points A web of 5 pages is described by a directed graph whose matrix is given by A Do the following ( points

### Diagonal, Symmetric and Triangular Matrices

Contents 1 Diagonal, Symmetric Triangular Matrices 2 Diagonal Matrices 2.1 Products, Powers Inverses of Diagonal Matrices 2.1.1 Theorem (Powers of Matrices) 2.2 Multiplying Matrices on the Left Right by

### Solving Systems of Linear Equations

LECTURE 5 Solving Systems of Linear Equations Recall that we introduced the notion of matrices as a way of standardizing the expression of systems of linear equations In today s lecture I shall show how

### Unified Lecture # 4 Vectors

Fall 2005 Unified Lecture # 4 Vectors These notes were written by J. Peraire as a review of vectors for Dynamics 16.07. They have been adapted for Unified Engineering by R. Radovitzky. References [1] Feynmann,

### ELEC-E8104 Stochastics models and estimation, Lecture 3b: Linear Estimation in Static Systems

Stochastics models and estimation, Lecture 3b: Linear Estimation in Static Systems Minimum Mean Square Error (MMSE) MMSE estimation of Gaussian random vectors Linear MMSE estimator for arbitrarily distributed

### October 3rd, 2012. Linear Algebra & Properties of the Covariance Matrix

Linear Algebra & Properties of the Covariance Matrix October 3rd, 2012 Estimation of r and C Let rn 1, rn, t..., rn T be the historical return rates on the n th asset. rn 1 rṇ 2 r n =. r T n n = 1, 2,...,

### 2. Simple Linear Regression

Research methods - II 3 2. Simple Linear Regression Simple linear regression is a technique in parametric statistics that is commonly used for analyzing mean response of a variable Y which changes according

### Helpsheet. Giblin Eunson Library MATRIX ALGEBRA. library.unimelb.edu.au/libraries/bee. Use this sheet to help you:

Helpsheet Giblin Eunson Library ATRIX ALGEBRA Use this sheet to help you: Understand the basic concepts and definitions of matrix algebra Express a set of linear equations in matrix notation Evaluate determinants

### Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares

Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares Many economic models involve endogeneity: that is, a theoretical relationship does not fit

### Diagonalisation. Chapter 3. Introduction. Eigenvalues and eigenvectors. Reading. Definitions

Chapter 3 Diagonalisation Eigenvalues and eigenvectors, diagonalisation of a matrix, orthogonal diagonalisation fo symmetric matrices Reading As in the previous chapter, there is no specific essential

### T ( a i x i ) = a i T (x i ).

Chapter 2 Defn 1. (p. 65) Let V and W be vector spaces (over F ). We call a function T : V W a linear transformation form V to W if, for all x, y V and c F, we have (a) T (x + y) = T (x) + T (y) and (b)

### Factorization Theorems

Chapter 7 Factorization Theorems This chapter highlights a few of the many factorization theorems for matrices While some factorization results are relatively direct, others are iterative While some factorization

### 1 Eigenvalues and Eigenvectors

Math 20 Chapter 5 Eigenvalues and Eigenvectors Eigenvalues and Eigenvectors. Definition: A scalar λ is called an eigenvalue of the n n matrix A is there is a nontrivial solution x of Ax = λx. Such an x

### ISOMETRIES OF R n KEITH CONRAD

ISOMETRIES OF R n KEITH CONRAD 1. Introduction An isometry of R n is a function h: R n R n that preserves the distance between vectors: h(v) h(w) = v w for all v and w in R n, where (x 1,..., x n ) = x

### 3 Random vectors and multivariate normal distribution

3 Random vectors and multivariate normal distribution As we saw in Chapter 1, a natural way to think about repeated measurement data is as a series of random vectors, one vector corresponding to each unit.

### is in plane V. However, it may be more convenient to introduce a plane coordinate system in V.

.4 COORDINATES EXAMPLE Let V be the plane in R with equation x +2x 2 +x 0, a two-dimensional subspace of R. We can describe a vector in this plane by its spatial (D)coordinates; for example, vector x 5

### Inner Product Spaces and Orthogonality

Inner Product Spaces and Orthogonality week 3-4 Fall 2006 Dot product of R n The inner product or dot product of R n is a function, defined by u, v a b + a 2 b 2 + + a n b n for u a, a 2,, a n T, v b,

### Econometrics Simple Linear Regression

Econometrics Simple Linear Regression Burcu Eke UC3M Linear equations with one variable Recall what a linear equation is: y = b 0 + b 1 x is a linear equation with one variable, or equivalently, a straight

### 2.1: MATRIX OPERATIONS

.: MATRIX OPERATIONS What are diagonal entries and the main diagonal of a matrix? What is a diagonal matrix? When are matrices equal? Scalar Multiplication 45 Matrix Addition Theorem (pg 0) Let A, B, and

### 1 VECTOR SPACES AND SUBSPACES

1 VECTOR SPACES AND SUBSPACES What is a vector? Many are familiar with the concept of a vector as: Something which has magnitude and direction. an ordered pair or triple. a description for quantities such

### Linear Algebraic Equations, SVD, and the Pseudo-Inverse

Linear Algebraic Equations, SVD, and the Pseudo-Inverse Philip N. Sabes October, 21 1 A Little Background 1.1 Singular values and matrix inversion For non-smmetric matrices, the eigenvalues and singular

### The Bivariate Normal Distribution

The Bivariate Normal Distribution This is Section 4.7 of the st edition (2002) of the book Introduction to Probability, by D. P. Bertsekas and J. N. Tsitsiklis. The material in this section was not included

### Section 1.4. Lines, Planes, and Hyperplanes. The Calculus of Functions of Several Variables

The Calculus of Functions of Several Variables Section 1.4 Lines, Planes, Hyperplanes In this section we will add to our basic geometric understing of R n by studying lines planes. If we do this carefully,

### Section 6.1 - Inner Products and Norms

Section 6.1 - Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,

### Notes for STA 437/1005 Methods for Multivariate Data

Notes for STA 437/1005 Methods for Multivariate Data Radford M. Neal, 26 November 2010 Random Vectors Notation: Let X be a random vector with p elements, so that X = [X 1,..., X p ], where denotes transpose.

### Vector Spaces; the Space R n

Vector Spaces; the Space R n Vector Spaces A vector space (over the real numbers) is a set V of mathematical entities, called vectors, U, V, W, etc, in which an addition operation + is defined and in which

### Session 7 Bivariate Data and Analysis

Session 7 Bivariate Data and Analysis Key Terms for This Session Previously Introduced mean standard deviation New in This Session association bivariate analysis contingency table co-variation least squares

### 5. Linear Regression

5. Linear Regression Outline.................................................................... 2 Simple linear regression 3 Linear model............................................................. 4

### L10: Probability, statistics, and estimation theory

L10: Probability, statistics, and estimation theory Review of probability theory Bayes theorem Statistics and the Normal distribution Least Squares Error estimation Maximum Likelihood estimation Bayesian

### Part 2: Analysis of Relationship Between Two Variables

Part 2: Analysis of Relationship Between Two Variables Linear Regression Linear correlation Significance Tests Multiple regression Linear Regression Y = a X + b Dependent Variable Independent Variable

### 4.3 Least Squares Approximations

18 Chapter. Orthogonality.3 Least Squares Approximations It often happens that Ax D b has no solution. The usual reason is: too many equations. The matrix has more rows than columns. There are more equations

### 4. Matrix inverses. left and right inverse. linear independence. nonsingular matrices. matrices with linearly independent columns

L. Vandenberghe EE133A (Spring 2016) 4. Matrix inverses left and right inverse linear independence nonsingular matrices matrices with linearly independent columns matrices with linearly independent rows

### 6. Vectors. 1 2009-2016 Scott Surgent (surgent@asu.edu)

6. Vectors For purposes of applications in calculus and physics, a vector has both a direction and a magnitude (length), and is usually represented as an arrow. The start of the arrow is the vector s foot,

### Equations Involving Lines and Planes Standard equations for lines in space

Equations Involving Lines and Planes In this section we will collect various important formulas regarding equations of lines and planes in three dimensional space Reminder regarding notation: any quantity

### Lecture Notes 2: Matrices as Systems of Linear Equations

2: Matrices as Systems of Linear Equations 33A Linear Algebra, Puck Rombach Last updated: April 13, 2016 Systems of Linear Equations Systems of linear equations can represent many things You have probably

### 2x + y = 3. Since the second equation is precisely the same as the first equation, it is enough to find x and y satisfying the system

1. Systems of linear equations We are interested in the solutions to systems of linear equations. A linear equation is of the form 3x 5y + 2z + w = 3. The key thing is that we don t multiply the variables

### Notes on Determinant

ENGG2012B Advanced Engineering Mathematics Notes on Determinant Lecturer: Kenneth Shum Lecture 9-18/02/2013 The determinant of a system of linear equations determines whether the solution is unique, without