THE $25,000,000,000 EIGENVECTOR THE LINEAR ALGEBRA BEHIND GOOGLE
|
|
|
- Tyrone Waters
- 9 years ago
- Views:
Transcription
1 THE $5,,, EIGENVECTOR THE LINEAR ALGEBRA BEHIND GOOGLE KURT BRYAN AND TANYA LEISE Abstract. Google s success derives in large part from its PageRank algorithm, which ranks the importance of webpages according to an eigenvector of a weighted link matrix. Analysis of the PageRank formula provides a wonderful applied topic for a linear algebra course. Instructors may assign this article as a project to more advanced students, or spend one or two lectures presenting the material with assigned homework from the exercises. This material also complements the discussion of Markov chains in matrix algebra. Maple and Mathematica files supporting this material can be found at bryan. Key words. linear algebra, PageRank, eigenvector, stochastic matrix AMS subject classifications. 5-, 5A8, 5A5. Introduction. When Google went online in the late 99 s, one thing that set it apart from other search engines was that its search result listings always seemed deliver the good stuff up front. With other search engines you often had to wade through screen after screen of links to irrelevant web pages that just happened to match the search text. Part of the magic behind Google is its PageRank algorithm, which quantitatively rates the importance of each page on the web, allowing Google to rank the pages and thereby present to the user the more important (and typically most relevant and helpful) pages first. Understanding how to calculate PageRank is essential for anyone designing a web page that they want people to access frequently, since getting listed first in a Google search leads to many people looking at your page. Indeed, due to Google s prominence as a search engine, its ranking system has had a deep influence on the development and structure of the internet, and on what kinds of information and services get accessed most frequently. Our goal in this paper is to explain one of the core ideas behind how Google calculates web page rankings. This turns out to be a delightful application of standard linear algebra. Search engines such as Google have to do three basic things:. Crawl the web and locate all web pages with public access.. Index the data from step, so that it can be searched efficiently for relevant keywords or phrases.. Rate the importance of each page in the database, so that when a user does a search and the subset of pages in the database with the desired information has been found, the more important pages can be presented first. This paper will focus on step. In an interconnected web of pages, how can one meaningfully define and quantify the importance of any given page? The rated importance of web pages is not the only factor in how links are presented, but it is a significant one. There are also successful ranking algorithms other than PageRank. The interested reader will find a wealth of information about ranking algorithms and search engines, and we list just a few references for getting started (see the extensive bibliography in [9], for example, for a more complete list). For a brief overview of how Google handles the entire process see [6], and for an in-depth treatment of PageRank see [] and a companion article [9]. Another article with good concrete examples is [5]. For more background on PageRank and explanations of essential principles of web design to maximize a website s PageRank, go to the websites [4,, 4]. To find out more about search engine principles in general and other ranking algorithms, see [] and [8]. Finally, for an account of some newer approaches to searching the web, see [] and [].. Developing a formula to rank pages. THE APPROXIMATE MARKET VALUE OF GOOGLE WHEN THE COMPANY WENT PUBLIC IN 4. Department of Mathematics, Rose-Hulman Institute of Technology, Terre Haute, IN 478; [email protected]; phone: 8) ; fax: (8) Mathematics and Computer Science Department, Amherst College, Amherst, MA ; [email protected]; phone: (4)54-54; fax: (4)54-55.
2 K. BRYAN AND T. LEISE 4 Fig... An example of a web with only four pages. An arrow from page A to page B indicates a link from page A to page B... The basic idea. In what follows we will use the phrase importance score or just score for any quantitative rating of a web page s importance. The importance score for any web page will always be a non-negative real number. A core idea in assigning a score to any given web page is that the page s score is derived from the links made to that page from other web pages. The links to a given page are called the backlinks for that page. The web thus becomes a democracy where pages vote for the importance of other pages by linking to them. Suppose the web of interest contains n pages, each page indexed by an integer k, k n. A typical example is illustrated in Figure., in which an arrow from page A to page B indicates a link from page A to page B. Such a web is an example of a directed graph. We ll use x k to denote the importance score of page k in the web. The x k is non-negative and x j > x k indicates that page j is more important than page k (so x j = indicates page j has the least possible importance score). A very simple approach is to take x k as the number of backlinks for page k. In the example in Figure., we have x =, x =, x =, and x 4 =, so that page would be the most important, pages and 4 tie for second, and page is least important. A link to page k becomes a vote for page k s importance. This approach ignores an important feature one would expect a ranking algorithm to have, namely, that a link to page k from an important page should boost page k s importance score more than a link from an unimportant page. For example, a link to your homepage directly from Yahoo ought to boost your page s score much more than a link from, say, (no relation to the author). In the web of Figure., pages and 4 both have two backlinks: each links to the other, but page s second backlink is from the seemingly important page, while page 4 s second backlink is from the relatively unimportant page. As such, perhaps we should rate page s importance higher than that of page 4. As a first attempt at incorporating this idea let s compute the score of page j as the sum of the scores of all pages linking to page j. For example, consider the web of Figure.. The score of page would be determined by the relation x = x + x 4. Since x and x 4 will depend on x this scheme seems strangely self-referential, but it is the approach we will use, with one more modification. Just as in elections, we don t want a single individual to gain influence merely by casting multiple votes. In the same vein, we seek a scheme in which a web page doesn t gain extra influence simply by linking to lots of other pages. If page j contains n j links, one of which links to page k, then we will boost page k s score by x j /n j, rather than by x j. In this scheme each web page gets a total of one vote, weighted by that web page s score, that is evenly divided up among all of its outgoing links. To quantify this for a web of n pages, let L k {,,..., n} denote the set of pages with a link to page k, that is, L k is the set of page k s backlinks. For each k we require x k = j L k x j n j, (.) where n j is the number of outgoing links from page j (which must be positive since if j L k then A graph consists of a set of vertices (in this context, the web pages) and a set of edges. Each edge joins a pair of vertices. The graph is undirected if the edges have no direction. The graph is directed if each edge (in the web context, the links) has a direction, that is, a starting and ending vertex.
3 THE $5,,, EIGENVECTOR 5 4 Fig... A web of five pages, consisting of two disconnected subwebs W (pages and ) and W (pages, 4, 5). page j links to at least page k!). We will assume that a link from a page to itself will not be counted. In this democracy of the web you don t get to vote for yourself! Let s apply this approach to the four-page web of Figure.. For page we have x = x / + x 4 /, since pages and 4 are backlinks for page and page contains only one link, while page 4 contains two links (splitting its vote in half). Similarly, x = x /, x = x / + x / + x 4 /, and x 4 = x / + x /. These linear equations can be written Ax = x, where x = [x x x x 4 ] T and A =. (.) This transforms the web ranking problem into the standard problem of finding an eigenvector for a square matrix! (Recall that the eigenvalues λ and eigenvectors x of a matrix A satisfy the equation Ax = λx, x by definition.) We thus seek an eigenvector x with eigenvalue for the matrix A. We will refer to A as the link matrix for the given web. It turns out that the link matrix A in equation (.) does indeed have eigenvectors with eigenvalue, namely, all multiples of the vector [ 4 9 6] T (recall that any non-zero multiple of an eigenvector is again an eigenvector). Let s agree to scale these importance score eigenvectors so that the components sum to. In this case we obtain x =.87, x = 4.9, x = 9.9, and x 4 = Note that this ranking differs from that generated by simply counting backlinks. It might seem surprising that page, linked to by all other pages, is not the most important. To understand this, note that page links only to page and so casts its entire vote for page. This, with the vote of page, results in page getting the highest importance score. More generally, the matrix A for any web must have as an eigenvalue if the web in question has no dangling nodes (pages with no outgoing links). To see this, first note that for a general web of n pages formula (.) gives rise to a matrix A with A ij = /n j if page j links to page i, A ij = otherwise. The jth column of A then contains n j non-zero entries, each equal to /n j, and the column thus sums to. This motivates the following definition, used in the study of Markov chains: Definition.. A square matrix is called a column-stochastic matrix if all of its entries are nonnegative and the entries in each column sum to one. The matrix A for a web with no dangling nodes is column-stochastic. We now prove Proposition. Every column-stochastic matrix has as an eigenvalue. Proof. Let A be an n n column-stochastic matrix and let e denote an n dimensional column vector with all entries equal to. Recall that A and its transpose A T have the same eigenvalues. Since A is column-stochastic it is easy to see that A T e = e, so that is an eigenvalue for A T and hence for A. In what follows we use V (A) to denote the eigenspace for eigenvalue of a column-stochastic matrix A... Shortcomings. Several difficulties arise with using formula (.) to rank websites. In this section we discuss two issues: webs with non-unique rankings and webs with dangling nodes.... Non-Unique Rankings. For our rankings it is desirable that the dimension of V (A) equal one, so that there is a unique eigenvector x with i x i = that we can use for importance scores. This is true in the web of Figure. and more generally is always true for the special case of
4 4 K. BRYAN AND T. LEISE a strongly connected web (that is, you can get from any page to any other page in a finite number of steps); see Exercise below. Unfortunately, it is not always true that the link matrix A will yield a unique ranking for all webs. Consider the web in Figure., for which the link matrix is A =. We find here that V (A) is two-dimensional; one possible pair of basis vectors is x = [/, /,,, ] T and y = [,, /, /, ] T. But note that any linear combination of these two vectors yields another vector in V (A), e.g., 4 x + 4 y = [/8, /8, /8, /8, ]T. It is not clear which, if any, of these eigenvectors we should use for the rankings! It is no coincidence that for the web of Figure. we find that dim(v (A)) >. It is a consequence of the fact that if a web W, considered as an undirected graph (ignoring which direction each arrows points), consists of r disconnected subwebs W,..., W r, then dim(v (A)) r, and hence there is no unique importance score vector x V (A) with i x i =. This makes intuitive sense: if a web W consists of r disconnected subwebs W,..., W r then one would expect difficulty in finding a common reference frame for comparing the scores of pages in one subweb with those in another subweb. Indeed, it is not hard to see why a web W consisting of r disconnected subwebs forces dim(v (A)) r. Suppose a web W has n pages and r component subwebs W,..., W r. Let n i denote the number of pages in W i. Index the pages in W with indices through n, the pages in W with indices n + through n + n, the pages in W with n + n + through n + n + n, etc. In general, let N i = i j= n j for i, with N =, so W i contains pages N i + through N i. For example, in the web of Figure we can take N = and N = 5, so W contains pages and, W contains pages, 4, and 5. The web in Figure. is a particular example of the general case, in which the matrix A assumes a block diagonal structure A = A... A.... A r, where A i denotes the link matrix for W i. In fact, W i can be considered as a web in its own right. Each n i n i matrix A i is column-stochastic, and hence possesses some eigenvector v i lr ni with eigenvector. For each i between and r construct a vector w i lr n which has components for all elements corresponding to blocks other than block i. For example, w = v. v, w =,.... Then it is easy to see that the vectors w i, i r, are linearly independent eigenvectors for A
5 THE $5,,, EIGENVECTOR 5 with eigenvalue because. Aw i = A v i = w i.. Thus V (A) has dimension at least r.... Dangling Nodes. Another difficulty may arise when using the matrix A to generate rankings. A web with dangling nodes produces a matrix A which contains one or more columns of all zeros. In this case A is column-substochastic, that is, the column sums of A are all less than or equal to one. Such a matrix must have all eigenvalues less than or equal to in magnitude, but need not actually be an eigenvalue for A. Nevertheless, the pages in a web with dangling nodes can still be ranked use a similar technique. The corresponding substochastic matrix must have a positive eigenvalue λ and a corresponding eigenvector x with non-negative entries (called the Perron eigenvector) that can be used to rank the web pages. See Exercise 4 below. We will not further consider the problem of dangling nodes here, however. Exercise. Suppose the people who own page in the web of Figure are infuriated by the fact that its importance score, computed using formula (.), is lower than the score of page. In an attempt to boost page s score, they create a page 5 that links to page ; page also links to page 5. Does this boost page s score above that of page? Exercise. Construct a web consisting of three or more subwebs and verify that dim(v (A)) equals (or exceeds) the number of the components in the web. Exercise. Add a link from page 5 to page in the web of Figure. The resulting web, considered as an undirected graph, is connected. What is the dimension of V (A)? Exercise 4. In the web of Figure., remove the link from page to page. In the resulting web page is now a dangling node. Set up the corresponding substochastic matrix and find its largest positive (Perron) eigenvalue. Find a non-negative Perron eigenvector for this eigenvalue, and scale the vector so that components sum to one. Does the resulting ranking seem reasonable? Exercise 5. Prove that in any web the importance score of a page with no backlinks is zero. Exercise 6. Implicit in our analysis up to this point is the assertion that the manner in which the pages of a web W are indexed has no effect on the importance score assigned to any given page. Prove this, as follows: Let W contains n pages, each page assigned an index through n, and let A be the resulting link matrix. Suppose we then transpose the indices of pages i and j (so page i is now page j and vice-versa). Let à be the link matrix for the relabelled web. Argue that à = PAP, where P is the elementary matrix obtained by transposing rows i and j of the n n identity matrix. Note that the operation A PA has the effect of swapping rows i and j of A, while A AP swaps columns i and j. Also, P = I, the identity matrix. Suppose that x is an eigenvector for A, so Ax = λx for some λ. Show that y = Px is an eigenvector for à with eigenvalue λ. Explain why this shows that transposing the indices of any two pages leaves the importance scores unchanged, and use this result to argue that any permutation of the page indices leaves the importance scores unchanged.. A remedy for dim(v (A)) >. An enormous amount of computing resources are needed to determine an eigenvector for the link matrix corresponding to a web containing billions of pages. It is thus important to know that our algorithm will yield a unique set of sensible web rankings. The analysis above shows that our first attempt to rank web pages leads to difficulties if the web isn t connected. And the worldwide web, treated as an undirected graph, contains many disjoint components; see [9] for some interesting statistics concerning the structure of the web.
6 6 K. BRYAN AND T. LEISE Below we present and analyze a modification of the above method that is guaranteed to overcome this shortcoming. The analysis that follows is basically a special case of the Perron-Frobenius theorem, and we only prove what we need for this application. For a full statement and proof of the Perron-Frobenius theorem, see chapter 8 in []... A modification to the link matrix A. For an n page web with no dangling nodes we can generate unambiguous importance scores as follows, including cases of web with multiple subwebs. Let S denote an n n matrix with all entries /n. The matrix S is column-stochastic, and it is easy to check that V (S) is one-dimensional. We will replace the matrix A with the matrix M = ( m)a + ms, (.) where m. M is a weighted average of A and S. The value of m originally used by Google is reportedly.5 [9, ]. For any m [, ] the matrix M is column-stochastic and we show below that V (M) is always one-dimensional if m (, ]. Thus M can be used to compute unambiguous importance scores. In the case when m = we have the original problem, for then M = A. At the other extreme is m =, yielding M = S. This is the ultimately egalitarian case: the only normalized eigenvector x with eigenvalue has x i = /n for all i and all web pages are rated equally important. Using M in place of A gives a web page with no backlinks (a dangling node) the importance score of m/n (Exercise 9), and the matrix M is substochastic for any m < since the matrix A is substochastic. Therefore the modified formula yields nonzero importance scores for dangling links (if m > ) but does not resolve the issue of dangling nodes. In the remainder of this article, we only consider webs with no dangling nodes. The equation x = Mx can also be cast as x = ( m)ax + ms, (.) where s is a column vector with all entries /n. Note that Sx = s if i x i =. We will prove below that V (M) is always one-dimensional, but first let s look at a couple of examples. Example : For the web of four pages in Figure. with matrix A given by (.), the new formula gives (with m =.5) M = , and yields importance scores x.68, x.4, x.88, and x 4.. This yields the same ranking of pages as the earlier computation, but the scores are slightly different. Example shows more explicitly the advantages of using M in place of A. Example : As a second example, for the web of Figure. with m =.5 we obtain the matrix M = (.)..... The space V (M) is indeed one-dimensional, with normalized eigenvector components of x =., x =., x =.85, x 4 =.85, and x 5 =.. The modification, using M instead of A, allows us to compare pages in different subwebs. Each entry M ij of M defined by equation (.) is strictly positive, which motivates the following definition. Definition.. A matrix M is positive if M ij > for all i and j. This is the key property that guarantees dim(v (M)) =, which we prove in the next section.
7 THE $5,,, EIGENVECTOR 7.. Analysis of the matrix M. Note that Proposition shows that V (M) is nonempty since M is stochastic. The goal of this section is to show that V (M) is in fact one-dimensional. This is a consequence of the following two propositions. Proposition. If M is positive and column-stochastic, then any eigenvector in V (M) has all positive or all negative components. Proof. We use proof by contradiction. First note that in the standard triangle inequality i y i i y i (with all y i real) the inequality is strict when the y i are of mixed sign. Suppose x V (M) contains elements of mixed sign. From x = Mx we have x i = n j= M ijx j and the summands M ij x j are of mixed sign (since M ij > ). As a result we have a strict inequality x i = M ij x j < j= M ij x j. (.4) Sum both sides of inequality (.4) from i = to i = n, and swap the i and j summations. Then use the fact that M is column-stochastic ( i M ij = for all j) to find x i < i= i= j= M ij x j = j= ( n ) M ij x j = j= i= x j, a contradiction. Hence x cannot contain both positive and negative elements. If x i for all i (and not all x i are zero) then x i > follows immediately from x i = n j= M ijx j and M ij >. Similarly x i for all i implies that each x i <. The following proposition will also be useful for analyzing dim(v (M)): Proposition. Let v and w be linearly independent vectors in lr m, m. Then, for some values of s and t that are not both zero, the vector x = sv + tw has both positive and negative components. Proof. Linear independence implies neither v nor w is zero. Let d = i v i. If d = then v must contain P components of mixed sign, and taking s = and t = yields the conclusion. i If d set s = wi d, t =, and x = sv + tw. Since v and w are independent x. However, i x i =. We conclude that x has both positive and negative components. We can now prove that using M in place of A yields an unambiguous ranking for any web with no dangling nodes. Lemma.. If M is positive and column-stochastic then V (M) has dimension. Proof. We again use proof by contradiction. Suppose there are two linearly independent eigenvectors v and w in the subspace V (M). For any real numbers s and t that are not both zero, the nonzero vector x = sv + tw must be in V (M), and so have components that are all negative or all positive. But by Proposition, for some choice of s and t the vector x must contain components of mixed sign, a contradiction. We conclude that V (M) cannot contain two linearly independent vectors, and so has dimension one. Lemma. provides the punchline for our analysis of the ranking algorithm using the matrix M (for < m < ). The space V (M) is one-dimensional, and moreover, the relevant eigenvectors have entirely positive or negative components. We are thus guaranteed the existence of a unique eigenvector x V (M) with positive components such that i x i =. Exercise 7. Prove that if A is an n n column-stochastic matrix and m, then M = ( m)a + ms is also a column-stochastic matrix. Exercise 8. Show that the product of two column-stochastic matrices is also column-stochastic. Exercise 9. Show that a page with no backlinks is given importance score m n by formula (.). Exercise. Suppose that A is the link matrix for a strongly connected web of n pages (any page can be reached from any other page by following a finite number of links). Show that dim(v (A)) = as follows. Let (A k ) ij denote the (i, j)-entry of A k. Note that page i can be reached from page j in one step if and only A ij > (since A ij > means there s a link from j to i!) Show that (A ) ij > if and only if page i can be reached from page j in exactly two steps. Hint: (A ) ij = k A ika kj ; all A ij are non-negative, so (A ) ij > implies that for some k both A ik and A kj are positive. j=
8 8 K. BRYAN AND T. LEISE Show more generally that (A p ) ij > if and only if page i can be reached from page j in EXACTLY p steps. Argue that (I + A + A + + A p ) ij > if and only if page i can be reached from page j in p or fewer steps (note p = is a legitimate choice any page can be reached from itself in zero steps!) Explain why I + A + A + + A n is a positive matrix if the web is strongly connected. Use the last part (and Exercise 8) so show that B = n (I + A + A + + A n ) is positive and column-stochastic (and hence by Lemma., dim(v (B)) = ). Show that if x V (A) then x V (B). Why does this imply that dim(v (A)) =? Exercise. Consider again the web in Figure., with the addition of a page 5 that links to page, where page also links to page 5. Calculate the new ranking by finding the eigenvector of M (corresponding to λ = ) that has positive components summing to one. Use m =.5. Exercise. Add a sixth page that links to every page of the web in the previous exercise, but to which no other page links. Rank the pages using A, then using M with m =.5, and compare the results. Exercise. Construct a web consisting of two or more subwebs and determine the ranking given by formula (.). At present the web contains at least eight billion pages how does one compute an eigenvector for an eight billion by eight billion matrix? One reasonable approach is an iterative procedure called the power method (along with modifications) that we will now examine for the special case at hand. It is worth noting that there is much additional analysis one can do, and many improved methods for the computation of PageRank. The reference [7] provides a typical example and additional references. 4. Computing the Importance Score Eigenvector. The rough idea behind the power method for computing an eigenvector of a matrix M is this: One starts with a typical vector x, then generates the sequence x k = Mx k (so x k = M k x ) and lets k approaches infinity. The vector x k is, to good approximation, an eigenvector for the dominant (largest magnitude) eigenvalue of M. However, depending on the magnitude of this eigenvalue, the vector x k may also grow without bound or decay to the zero vector. One thus typically rescales at each iteration, say by Mx k Mx k computing x k =, where can be any vector norm. The method generally requires that the corresponding eigenspace be one-dimensional, a condition that is satisfied in the case when M is defined by equation (.). To use the power method on the matrices M that arise from the web ranking problem we would generally need to know that any other eigenvalues λ of M satisfy λ <. This assures that the power method will converge to the eigenvector we want. Actually, the following proposition provides what we need, with no reference to any other eigenvalues of M! Definition 4.. The -norm of a vector v is v = i v i. Proposition 4. Let M be a positive column-stochastic n n matrix and let V denote the subspace of lr n consisting of vectors v such that j v j =. Then Mv V for any v V, and Mv c v for any v V, where c = max j n min i n M ij <. straightforward: Let w = Mv, so that w i = n j= M ijv j and w i = i= i= j= M ij v j = ( n ) v j M ij = Hence w = Mv V. To prove the bound in the proposition note that w = e i w i = e i M ij v j, i= j= i= i= j= Proof. To see that Mv V is v j =. See [5] for a general introduction to the power method and the use of spectral decomposition to find the rate of convergence of the vectors x k = M k x. j=
9 THE $5,,, EIGENVECTOR 9 where e i = sgn(w i ). Note that the e i are not all of one sign, since i w i = (unless w in which case the bound clearly holds). Reverse the double sum to obtain w = ( n ) v j e i M ij = j= i= a j v j, (4.) where a j = n i= e im ij. Since the e i are of mixed sign and i M ij = with < M ij <, it is easy to see that We can thus bound j= < + min i n M ij a j min i n M ij <. a j min i n M ij <. Let c = max j n min i n M ij. Observe that c < and a j c for all j. From equation (4.) we have w = a j v j = a j v j a j v j c v j = c v, j= j= which proves the proposition. Proposition 4 sets the stage for the following proposition. Proposition 5. Every positive column-stochastic matrix M has a unique vector q with positive components such that Mq = q with q =. The vector q can be computed as q = lim k M k x for any initial guess x with positive components such that x =. Proof. From Proposition the matrix M has as an eigenvalue and by Lemma. the subspace V (M) is one-dimensional. Also, all non-zero vectors in V (M) have entirely positive or negative components. It is clear that there is a unique vector q V (M) with positive components such that i q i =. Let x be any vector in lr n with positive components such that x =. We can write x = q + v where v V (V as in Proposition 4). We find that M k x = M k q + M k v = q + M k v. As a result j= j= M k x q = M k v. (4.) A straightforward induction and Proposition 4 shows that M k v c k v for c < (c as in Proposition 4) and so lim k M k v =. From equation (4.) we conclude that lim k M k x = q. Example: Let M be the matrix defined by equation (.) for the web of Figure.. We take x = [.4,.,.8,.8,.9] T as an initial guess; recall that we had q = [.,.,.85,.85,.] T. The table below shows the value of M k x q for several values of k, as well as the ratio M k x q / M k x q. Compare this ratio to c from Proposition 4, which in this case is.94. k M k x q M k x q M k x q It is clear that the bound M k x q c k x q is rather pessimistic (note.85 is the value m, and.85 turns out to be the second largest eigenvalue for M). One can show that in general the power method will converge asymptotically according to Mx k q λ x q, where λ
10 K. BRYAN AND T. LEISE is the second largest eigenvalue of M. Moreover, for M of the form M = ( m)a + ms with A column-stochastic and all S ij = /n it can be shown that λ m (see, e.g., [], Theorem 5.). As a result, the power method will converge much more rapidly than indicated by c k x q. Nonetheless, the value of c in Proposition 4 provides a very simple bound on the convergence of the power method here. It is easy to see that since all entries of M are at least m/n, we will always have c m/n in Proposition 4. As a practical matter, note that the n n positive matrix M has no non-zero elements, so the multiplication Mv for v lr n will typically take O(n ) multiplications and additions, a formidable computation if n = 8,,,. But equation (.) shows that if x is positive with x = then the multiplication Mx is equivalent to ( m)ax + ms. This is a far more efficient computation, since A can be expected to contain mostly zeros (most web pages link to only a few other pages). We ve now proved our main theorem: Theorem 4.. The matrix M defined by (.) for a web with no dangling nodes will always be a positive column-stochastic matrix and so have a unique q with positive components such that Mq = q and i q i =. The vector q may be computed as the limit of iterations x k = ( m)ax k + ms, where x is any initial vector with positive components and x =. The eigenvector x defined by equation (.) also has a probabilistic interpretation. Consider a web-surfer on a web of n pages with no dangling nodes. The surfer begins at some web page (it doesn t matter where) and randomly moves from web page to web page according to the following procedure: If the surfer is currently at a page with r outgoing links, he either randomly chooses any one of these links with uniform probability m OR he jumps to any randomly selected page on the r r +n m n web, each with probability m m n (note that r =, so this accounts for everything he can do). The surfer repeats this page-hopping procedure ad infinitum. The component x j of the normalized vector x in equation (.) is the fraction of time that the surfer spends, in the long run, on page j of the web. More important pages tend to be linked to by many other pages and so the surfer hits those most often. Exercise 4. For the web in Exercise, compute the values of M k x q and Mk x q M k x q for k =, 5,, 5, using an initial guess x not too close to the actual eigenvector q (so that you can watch the convergence). Determine c = max j n min i n M ij and the absolute value of the second largest eigenvalue of M. M k x q M k x q, Exercise 5. To see why the second largest eigenvalue plays a role in bounding consider an n n positive column-stochastic matrix M that is diagonalizable. Let x be any vector with non-negative components that sum to one. Since M is diagonalizable, we can create a basis of eigenvectors {q, v,..., v n }, where q is the steady state vector, and then write x = aq + n k= b kv k. Determine M k x, and then show that a = and the sum of the components of each v k must equal. Next apply Proposition 4 to prove that, except for the non-repeated eigenvalue λ =, the other eigenvalues are all strictly less than one in absolute value. Use this to evaluate lim k M k x q M k x q. Exercise 6. Consider the link matrix A = Show that M = ( m)a + ms (all S ij = /) is not diagonalizable for m <. Exercise 7. How should the value of m be chosen? How does this choice affect the rankings and the computation time?. REFERENCES [] A. Berman and R. Plemmons, Nonnegative Matrices in the Mathematical Sciences, Academic Press, New York, 979. [] M. W. Berry and M. Browne, Understanding Search Engines: Mathematical Modeling and Text Retrieval, Second Edition, SIAM, Philadelphia, 5. [] M. Bianchini, M. Gori, and F. Scarselli, Inside PageRank, ACM Trans. Internet Tech., 5 (5), pp. 9 8.
11 THE $5,,, EIGENVECTOR [4] S. Brin and L. Page, The anatomy of a large-scale hypertextual web search engine, http : //www db.stanford.edu/ backrub/google.html (accessed August, 5). [5] A. Farahat, T. Lofaro, J. C. Miller, G. Rae, and L.A. Ward, Authority Rankings from HITS, PageRank, and SALSA: Existence, Uniqueness, and Effect of Initialization, SIAM J. Sci. Comput., 7 (6), pp. 8-. [6] A. Hill, Google Inside Out, Maximum PC, April 4, pp [7] S. Kamvar, T. Haveliwala, and G. Golub, Adaptive methods for the computation of PageRank, Linear Algebra Appl., 86 (4), pp [8] A. N. Langville and C. D. Meyer, A survey of eigenvector methods of web information retrieval, SIAM Review, 47 (5), pp [9] A. N. Langville and C. D. Meyer, Deeper inside PageRank, Internet Math., (5), pp [] C. D. Meyer, Matrix Analysis and Applied Linear Algebra, SIAM, Philadelphia,. [] Cleve Moler, The world s largest matrix computation, http : // notes /clevescorner/oct cleve.html (accessed August, 5). [] Mostafa, J., Seeking better web searches, Sci. Amer., 9 (5), pp [] Sara Robinson, The Ongoing search for efficient web search algorithms, SIAM News, 7 (Nov 4). [4] Ian Rogers, The Google Pagerank algorithm and how it works, http : // (accessed August, 5). [5] W. J. Stewart, An Introduction to the Numerical Solution of Markov Chains, Princeton University Press, Princeton, 994.
Similarity and Diagonalization. Similar Matrices
MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that
Section 6.1 - Inner Products and Norms
Section 6.1 - Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,
DATA ANALYSIS II. Matrix Algorithms
DATA ANALYSIS II Matrix Algorithms Similarity Matrix Given a dataset D = {x i }, i=1,..,n consisting of n points in R d, let A denote the n n symmetric similarity matrix between the points, given as where
Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013
Notes on Orthogonal and Symmetric Matrices MENU, Winter 201 These notes summarize the main properties and uses of orthogonal and symmetric matrices. We covered quite a bit of material regarding these topics,
Continued Fractions and the Euclidean Algorithm
Continued Fractions and the Euclidean Algorithm Lecture notes prepared for MATH 326, Spring 997 Department of Mathematics and Statistics University at Albany William F Hammond Table of Contents Introduction
MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors. Jordan canonical form (continued).
MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors Jordan canonical form (continued) Jordan canonical form A Jordan block is a square matrix of the form λ 1 0 0 0 0 λ 1 0 0 0 0 λ 0 0 J = 0
NOTES ON LINEAR TRANSFORMATIONS
NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all
Solution to Homework 2
Solution to Homework 2 Olena Bormashenko September 23, 2011 Section 1.4: 1(a)(b)(i)(k), 4, 5, 14; Section 1.5: 1(a)(b)(c)(d)(e)(n), 2(a)(c), 13, 16, 17, 18, 27 Section 1.4 1. Compute the following, if
Continuity of the Perron Root
Linear and Multilinear Algebra http://dx.doi.org/10.1080/03081087.2014.934233 ArXiv: 1407.7564 (http://arxiv.org/abs/1407.7564) Continuity of the Perron Root Carl D. Meyer Department of Mathematics, North
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
UNCOUPLING THE PERRON EIGENVECTOR PROBLEM
UNCOUPLING THE PERRON EIGENVECTOR PROBLEM Carl D Meyer INTRODUCTION Foranonnegative irreducible matrix m m with spectral radius ρ,afundamental problem concerns the determination of the unique normalized
Notes on Determinant
ENGG2012B Advanced Engineering Mathematics Notes on Determinant Lecturer: Kenneth Shum Lecture 9-18/02/2013 The determinant of a system of linear equations determines whether the solution is unique, without
Practical Guide to the Simplex Method of Linear Programming
Practical Guide to the Simplex Method of Linear Programming Marcel Oliver Revised: April, 0 The basic steps of the simplex algorithm Step : Write the linear programming problem in standard form Linear
160 CHAPTER 4. VECTOR SPACES
160 CHAPTER 4. VECTOR SPACES 4. Rank and Nullity In this section, we look at relationships between the row space, column space, null space of a matrix and its transpose. We will derive fundamental results
Au = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.
Chapter 7 Eigenvalues and Eigenvectors In this last chapter of our exploration of Linear Algebra we will revisit eigenvalues and eigenvectors of matrices, concepts that were already introduced in Geometry
The PageRank Citation Ranking: Bring Order to the Web
The PageRank Citation Ranking: Bring Order to the Web presented by: Xiaoxi Pang 25.Nov 2010 1 / 20 Outline Introduction A ranking for every page on the Web Implementation Convergence Properties Personalized
Vector and Matrix Norms
Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a non-empty
October 3rd, 2012. Linear Algebra & Properties of the Covariance Matrix
Linear Algebra & Properties of the Covariance Matrix October 3rd, 2012 Estimation of r and C Let rn 1, rn, t..., rn T be the historical return rates on the n th asset. rn 1 rṇ 2 r n =. r T n n = 1, 2,...,
Numerical Analysis Lecture Notes
Numerical Analysis Lecture Notes Peter J. Olver 6. Eigenvalues and Singular Values In this section, we collect together the basic facts about eigenvalues and eigenvectors. From a geometrical viewpoint,
MATH 551 - APPLIED MATRIX THEORY
MATH 55 - APPLIED MATRIX THEORY FINAL TEST: SAMPLE with SOLUTIONS (25 points NAME: PROBLEM (3 points A web of 5 pages is described by a directed graph whose matrix is given by A Do the following ( points
Lecture 3: Finding integer solutions to systems of linear equations
Lecture 3: Finding integer solutions to systems of linear equations Algorithmic Number Theory (Fall 2014) Rutgers University Swastik Kopparty Scribe: Abhishek Bhrushundi 1 Overview The goal of this lecture
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +
Inner products on R n, and more
Inner products on R n, and more Peyam Ryan Tabrizian Friday, April 12th, 2013 1 Introduction You might be wondering: Are there inner products on R n that are not the usual dot product x y = x 1 y 1 + +
Orthogonal Diagonalization of Symmetric Matrices
MATH10212 Linear Algebra Brief lecture notes 57 Gram Schmidt Process enables us to find an orthogonal basis of a subspace. Let u 1,..., u k be a basis of a subspace V of R n. We begin the process of finding
Math 4310 Handout - Quotient Vector Spaces
Math 4310 Handout - Quotient Vector Spaces Dan Collins The textbook defines a subspace of a vector space in Chapter 4, but it avoids ever discussing the notion of a quotient space. This is understandable
Chapter 3. Cartesian Products and Relations. 3.1 Cartesian Products
Chapter 3 Cartesian Products and Relations The material in this chapter is the first real encounter with abstraction. Relations are very general thing they are a special type of subset. After introducing
A linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form
Section 1.3 Matrix Products A linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form (scalar #1)(quantity #1) + (scalar #2)(quantity #2) +...
What is Linear Programming?
Chapter 1 What is Linear Programming? An optimization problem usually has three essential ingredients: a variable vector x consisting of a set of unknowns to be determined, an objective function of x to
Linear Programming. March 14, 2014
Linear Programming March 1, 01 Parts of this introduction to linear programming were adapted from Chapter 9 of Introduction to Algorithms, Second Edition, by Cormen, Leiserson, Rivest and Stein [1]. 1
The Mathematics Behind Google s PageRank
The Mathematics Behind Google s PageRank Ilse Ipsen Department of Mathematics North Carolina State University Raleigh, USA Joint work with Rebecca Wills Man p.1 Two Factors Determine where Google displays
MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.
MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix. Nullspace Let A = (a ij ) be an m n matrix. Definition. The nullspace of the matrix A, denoted N(A), is the set of all n-dimensional column
by the matrix A results in a vector which is a reflection of the given
Eigenvalues & Eigenvectors Example Suppose Then So, geometrically, multiplying a vector in by the matrix A results in a vector which is a reflection of the given vector about the y-axis We observe that
A note on companion matrices
Linear Algebra and its Applications 372 (2003) 325 33 www.elsevier.com/locate/laa A note on companion matrices Miroslav Fiedler Academy of Sciences of the Czech Republic Institute of Computer Science Pod
Linear Maps. Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (February 5, 2007)
MAT067 University of California, Davis Winter 2007 Linear Maps Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (February 5, 2007) As we have discussed in the lecture on What is Linear Algebra? one of
Linear Algebra Review. Vectors
Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka [email protected] http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa Cogsci 8F Linear Algebra review UCSD Vectors The length
Duality of linear conic problems
Duality of linear conic problems Alexander Shapiro and Arkadi Nemirovski Abstract It is well known that the optimal values of a linear programming problem and its dual are equal to each other if at least
3. Mathematical Induction
3. MATHEMATICAL INDUCTION 83 3. Mathematical Induction 3.1. First Principle of Mathematical Induction. Let P (n) be a predicate with domain of discourse (over) the natural numbers N = {0, 1,,...}. If (1)
Suk-Geun Hwang and Jin-Woo Park
Bull. Korean Math. Soc. 43 (2006), No. 3, pp. 471 478 A NOTE ON PARTIAL SIGN-SOLVABILITY Suk-Geun Hwang and Jin-Woo Park Abstract. In this paper we prove that if Ax = b is a partial signsolvable linear
1 Solving LPs: The Simplex Algorithm of George Dantzig
Solving LPs: The Simplex Algorithm of George Dantzig. Simplex Pivoting: Dictionary Format We illustrate a general solution procedure, called the simplex algorithm, by implementing it on a very simple example.
University of Lille I PC first year list of exercises n 7. Review
University of Lille I PC first year list of exercises n 7 Review Exercise Solve the following systems in 4 different ways (by substitution, by the Gauss method, by inverting the matrix of coefficients
Chapter 17. Orthogonal Matrices and Symmetries of Space
Chapter 17. Orthogonal Matrices and Symmetries of Space Take a random matrix, say 1 3 A = 4 5 6, 7 8 9 and compare the lengths of e 1 and Ae 1. The vector e 1 has length 1, while Ae 1 = (1, 4, 7) has length
Notes on Factoring. MA 206 Kurt Bryan
The General Approach Notes on Factoring MA 26 Kurt Bryan Suppose I hand you n, a 2 digit integer and tell you that n is composite, with smallest prime factor around 5 digits. Finding a nontrivial factor
Math 55: Discrete Mathematics
Math 55: Discrete Mathematics UC Berkeley, Fall 2011 Homework # 5, due Wednesday, February 22 5.1.4 Let P (n) be the statement that 1 3 + 2 3 + + n 3 = (n(n + 1)/2) 2 for the positive integer n. a) What
General Framework for an Iterative Solution of Ax b. Jacobi s Method
2.6 Iterative Solutions of Linear Systems 143 2.6 Iterative Solutions of Linear Systems Consistent linear systems in real life are solved in one of two ways: by direct calculation (using a matrix factorization,
MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.
MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column
SHARP BOUNDS FOR THE SUM OF THE SQUARES OF THE DEGREES OF A GRAPH
31 Kragujevac J. Math. 25 (2003) 31 49. SHARP BOUNDS FOR THE SUM OF THE SQUARES OF THE DEGREES OF A GRAPH Kinkar Ch. Das Department of Mathematics, Indian Institute of Technology, Kharagpur 721302, W.B.,
Chapter 6. Orthogonality
6.3 Orthogonal Matrices 1 Chapter 6. Orthogonality 6.3 Orthogonal Matrices Definition 6.4. An n n matrix A is orthogonal if A T A = I. Note. We will see that the columns of an orthogonal matrix must be
1. Let P be the space of all polynomials (of one real variable and with real coefficients) with the norm
Uppsala Universitet Matematiska Institutionen Andreas Strömbergsson Prov i matematik Funktionalanalys Kurs: F3B, F4Sy, NVP 005-06-15 Skrivtid: 9 14 Tillåtna hjälpmedel: Manuella skrivdon, Kreyszigs bok
DETERMINANTS IN THE KRONECKER PRODUCT OF MATRICES: THE INCIDENCE MATRIX OF A COMPLETE GRAPH
DETERMINANTS IN THE KRONECKER PRODUCT OF MATRICES: THE INCIDENCE MATRIX OF A COMPLETE GRAPH CHRISTOPHER RH HANUSA AND THOMAS ZASLAVSKY Abstract We investigate the least common multiple of all subdeterminants,
Mathematics Course 111: Algebra I Part IV: Vector Spaces
Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 1996-7 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are
On the representability of the bi-uniform matroid
On the representability of the bi-uniform matroid Simeon Ball, Carles Padró, Zsuzsa Weiner and Chaoping Xing August 3, 2012 Abstract Every bi-uniform matroid is representable over all sufficiently large
Lecture L3 - Vectors, Matrices and Coordinate Transformations
S. Widnall 16.07 Dynamics Fall 2009 Lecture notes based on J. Peraire Version 2.0 Lecture L3 - Vectors, Matrices and Coordinate Transformations By using vectors and defining appropriate operations between
7 Gaussian Elimination and LU Factorization
7 Gaussian Elimination and LU Factorization In this final section on matrix factorization methods for solving Ax = b we want to take a closer look at Gaussian elimination (probably the best known method
THE DIMENSION OF A VECTOR SPACE
THE DIMENSION OF A VECTOR SPACE KEITH CONRAD This handout is a supplementary discussion leading up to the definition of dimension and some of its basic properties. Let V be a vector space over a field
Mathematical Induction
Mathematical Induction (Handout March 8, 01) The Principle of Mathematical Induction provides a means to prove infinitely many statements all at once The principle is logical rather than strictly mathematical,
INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS
INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS STEVEN P. LALLEY AND ANDREW NOBEL Abstract. It is shown that there are no consistent decision rules for the hypothesis testing problem
[1] Diagonal factorization
8.03 LA.6: Diagonalization and Orthogonal Matrices [ Diagonal factorization [2 Solving systems of first order differential equations [3 Symmetric and Orthonormal Matrices [ Diagonal factorization Recall:
1 VECTOR SPACES AND SUBSPACES
1 VECTOR SPACES AND SUBSPACES What is a vector? Many are familiar with the concept of a vector as: Something which has magnitude and direction. an ordered pair or triple. a description for quantities such
4.5 Linear Dependence and Linear Independence
4.5 Linear Dependence and Linear Independence 267 32. {v 1, v 2 }, where v 1, v 2 are collinear vectors in R 3. 33. Prove that if S and S are subsets of a vector space V such that S is a subset of S, then
Lecture 1: Schur s Unitary Triangularization Theorem
Lecture 1: Schur s Unitary Triangularization Theorem This lecture introduces the notion of unitary equivalence and presents Schur s theorem and some of its consequences It roughly corresponds to Sections
Classification of Cartan matrices
Chapter 7 Classification of Cartan matrices In this chapter we describe a classification of generalised Cartan matrices This classification can be compared as the rough classification of varieties in terms
4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION
4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION STEVEN HEILMAN Contents 1. Review 1 2. Diagonal Matrices 1 3. Eigenvectors and Eigenvalues 2 4. Characteristic Polynomial 4 5. Diagonalizability 6 6. Appendix:
1 Sets and Set Notation.
LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES Sets and Set Notation. Definition (Naive Definition of a Set). A set is any collection of objects, called the elements of that set. We will most
Sensitivity Analysis 3.1 AN EXAMPLE FOR ANALYSIS
Sensitivity Analysis 3 We have already been introduced to sensitivity analysis in Chapter via the geometry of a simple example. We saw that the values of the decision variables and those of the slack and
Big Data Technology Motivating NoSQL Databases: Computing Page Importance Metrics at Crawl Time
Big Data Technology Motivating NoSQL Databases: Computing Page Importance Metrics at Crawl Time Edward Bortnikov & Ronny Lempel Yahoo! Labs, Haifa Class Outline Link-based page importance measures Why
COMBINATORIAL PROPERTIES OF THE HIGMAN-SIMS GRAPH. 1. Introduction
COMBINATORIAL PROPERTIES OF THE HIGMAN-SIMS GRAPH ZACHARY ABEL 1. Introduction In this survey we discuss properties of the Higman-Sims graph, which has 100 vertices, 1100 edges, and is 22 regular. In fact
LINEAR ALGEBRA W W L CHEN
LINEAR ALGEBRA W W L CHEN c W W L Chen, 1997, 2008 This chapter is available free to all individuals, on understanding that it is not to be used for financial gain, and may be downloaded and/or photocopied,
Numerical Analysis Lecture Notes
Numerical Analysis Lecture Notes Peter J. Olver 5. Inner Products and Norms The norm of a vector is a measure of its size. Besides the familiar Euclidean norm based on the dot product, there are a number
Linear Algebra I. Ronald van Luijk, 2012
Linear Algebra I Ronald van Luijk, 2012 With many parts from Linear Algebra I by Michael Stoll, 2007 Contents 1. Vector spaces 3 1.1. Examples 3 1.2. Fields 4 1.3. The field of complex numbers. 6 1.4.
HOMEWORK 5 SOLUTIONS. n!f n (1) lim. ln x n! + xn x. 1 = G n 1 (x). (2) k + 1 n. (n 1)!
Math 7 Fall 205 HOMEWORK 5 SOLUTIONS Problem. 2008 B2 Let F 0 x = ln x. For n 0 and x > 0, let F n+ x = 0 F ntdt. Evaluate n!f n lim n ln n. By directly computing F n x for small n s, we obtain the following
1 Norms and Vector Spaces
008.10.07.01 1 Norms and Vector Spaces Suppose we have a complex vector space V. A norm is a function f : V R which satisfies (i) f(x) 0 for all x V (ii) f(x + y) f(x) + f(y) for all x,y V (iii) f(λx)
10.2 ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS. The Jacobi Method
578 CHAPTER 1 NUMERICAL METHODS 1. ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS As a numerical technique, Gaussian elimination is rather unusual because it is direct. That is, a solution is obtained after
BANACH AND HILBERT SPACE REVIEW
BANACH AND HILBET SPACE EVIEW CHISTOPHE HEIL These notes will briefly review some basic concepts related to the theory of Banach and Hilbert spaces. We are not trying to give a complete development, but
Linear Algebra Notes for Marsden and Tromba Vector Calculus
Linear Algebra Notes for Marsden and Tromba Vector Calculus n-dimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of
a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.
Chapter 1 LINEAR EQUATIONS 1.1 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,..., a n, b are given
T ( a i x i ) = a i T (x i ).
Chapter 2 Defn 1. (p. 65) Let V and W be vector spaces (over F ). We call a function T : V W a linear transformation form V to W if, for all x, y V and c F, we have (a) T (x + y) = T (x) + T (y) and (b)
15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
Matrices 2. Solving Square Systems of Linear Equations; Inverse Matrices
Matrices 2. Solving Square Systems of Linear Equations; Inverse Matrices Solving square systems of linear equations; inverse matrices. Linear algebra is essentially about solving systems of linear equations,
Inner Product Spaces
Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and
A characterization of trace zero symmetric nonnegative 5x5 matrices
A characterization of trace zero symmetric nonnegative 5x5 matrices Oren Spector June 1, 009 Abstract The problem of determining necessary and sufficient conditions for a set of real numbers to be the
1 if 1 x 0 1 if 0 x 1
Chapter 3 Continuity In this chapter we begin by defining the fundamental notion of continuity for real valued functions of a single real variable. When trying to decide whether a given function is or
SOLUTIONS TO EXERCISES FOR. MATHEMATICS 205A Part 3. Spaces with special properties
SOLUTIONS TO EXERCISES FOR MATHEMATICS 205A Part 3 Fall 2008 III. Spaces with special properties III.1 : Compact spaces I Problems from Munkres, 26, pp. 170 172 3. Show that a finite union of compact subspaces
Systems of Linear Equations
Systems of Linear Equations Beifang Chen Systems of linear equations Linear systems A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where a, a,, a n and
Matrix Representations of Linear Transformations and Changes of Coordinates
Matrix Representations of Linear Transformations and Changes of Coordinates 01 Subspaces and Bases 011 Definitions A subspace V of R n is a subset of R n that contains the zero element and is closed under
Math 312 Homework 1 Solutions
Math 31 Homework 1 Solutions Last modified: July 15, 01 This homework is due on Thursday, July 1th, 01 at 1:10pm Please turn it in during class, or in my mailbox in the main math office (next to 4W1) Please
MATH 289 PROBLEM SET 4: NUMBER THEORY
MATH 289 PROBLEM SET 4: NUMBER THEORY 1. The greatest common divisor If d and n are integers, then we say that d divides n if and only if there exists an integer q such that n = qd. Notice that if d divides
LEARNING OBJECTIVES FOR THIS CHAPTER
CHAPTER 2 American mathematician Paul Halmos (1916 2006), who in 1942 published the first modern linear algebra book. The title of Halmos s book was the same as the title of this chapter. Finite-Dimensional
No: 10 04. Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics
No: 10 04 Bilkent University Monotonic Extension Farhad Husseinov Discussion Papers Department of Economics The Discussion Papers of the Department of Economics are intended to make the initial results
Real Roots of Univariate Polynomials with Real Coefficients
Real Roots of Univariate Polynomials with Real Coefficients mostly written by Christina Hewitt March 22, 2012 1 Introduction Polynomial equations are used throughout mathematics. When solving polynomials
Solutions to Math 51 First Exam January 29, 2015
Solutions to Math 5 First Exam January 29, 25. ( points) (a) Complete the following sentence: A set of vectors {v,..., v k } is defined to be linearly dependent if (2 points) there exist c,... c k R, not
Lecture 3. Linear Programming. 3B1B Optimization Michaelmas 2015 A. Zisserman. Extreme solutions. Simplex method. Interior point method
Lecture 3 3B1B Optimization Michaelmas 2015 A. Zisserman Linear Programming Extreme solutions Simplex method Interior point method Integer programming and relaxation The Optimization Tree Linear Programming
3 Some Integer Functions
3 Some Integer Functions A Pair of Fundamental Integer Functions The integer function that is the heart of this section is the modulo function. However, before getting to it, let us look at some very simple
SCORE SETS IN ORIENTED GRAPHS
Applicable Analysis and Discrete Mathematics, 2 (2008), 107 113. Available electronically at http://pefmath.etf.bg.ac.yu SCORE SETS IN ORIENTED GRAPHS S. Pirzada, T. A. Naikoo The score of a vertex v in
AN ALGORITHM FOR DETERMINING WHETHER A GIVEN BINARY MATROID IS GRAPHIC
AN ALGORITHM FOR DETERMINING WHETHER A GIVEN BINARY MATROID IS GRAPHIC W. T. TUTTE. Introduction. In a recent series of papers [l-4] on graphs and matroids I used definitions equivalent to the following.
Inner Product Spaces and Orthogonality
Inner Product Spaces and Orthogonality week 3-4 Fall 2006 Dot product of R n The inner product or dot product of R n is a function, defined by u, v a b + a 2 b 2 + + a n b n for u a, a 2,, a n T, v b,
Numerical Methods I Eigenvalue Problems
Numerical Methods I Eigenvalue Problems Aleksandar Donev Courant Institute, NYU 1 [email protected] 1 Course G63.2010.001 / G22.2420-001, Fall 2010 September 30th, 2010 A. Donev (Courant Institute)
CS 103X: Discrete Structures Homework Assignment 3 Solutions
CS 103X: Discrete Structures Homework Assignment 3 s Exercise 1 (20 points). On well-ordering and induction: (a) Prove the induction principle from the well-ordering principle. (b) Prove the well-ordering
Section 4.4 Inner Product Spaces
Section 4.4 Inner Product Spaces In our discussion of vector spaces the specific nature of F as a field, other than the fact that it is a field, has played virtually no role. In this section we no longer
The Determinant: a Means to Calculate Volume
The Determinant: a Means to Calculate Volume Bo Peng August 20, 2007 Abstract This paper gives a definition of the determinant and lists many of its well-known properties Volumes of parallelepipeds are
8 Divisibility and prime numbers
8 Divisibility and prime numbers 8.1 Divisibility In this short section we extend the concept of a multiple from the natural numbers to the integers. We also summarize several other terms that express
