Moving Least Squares Approximation
|
|
|
- Clara Morton
- 9 years ago
- Views:
Transcription
1 Chapter 7 Moving Least Squares Approimation An alternative to radial basis function interpolation and approimation is the so-called moving least squares method. As we will see below, in this method the approimation Pf to f is obtained by solving many (small) linear systems, instead of via solution of a single but large linear system as we did in the previous chapters. To make a connection with the previous chapters we start with the Backus-Gilbert formulation of the moving least squares method since this corresponds to a linearly constrained quadratic minimization problem. 7.1 Moving Least Squares Approimation: The Backus- Gilbert Approach The connection between the standard moving least squares formulation (to be eplained in the net section) and Backus-Gilbert theory was pointed out by Bos and Šalkauskas in [67]. Mathematically, in the Backus-Gilbert approach one considers a quasi-interpolant of the form Pf() = f( i )Ψ i (), (7.1) where f = [f( 1 ),..., f( N )] T represents the given data. From Theorem we know that the quasi-interpolant that minimizes the point-wise error is given if the generating functions Ψ i are cardinal functions, i.e., Ψ i ( j ) = δ ij, i, j = 1,..., N. In the moving least squares method one does not attempt to minimize the pointwise error, but instead seeks to find the values of the generating functions Ψ i () = Ψ(, i ) by minimizing 1 2 Ψ 2 1 i () W (, i ) (7.2) subject to the polynomial reproduction constraints p( i )Ψ i () = p(), for all p Π s d, (7.3) 74
2 where Π s d is the space of s-variate polynomials of total degree at most d which has dimension m = ( ) s+d d. Remarks: 1. In the above formulation there is no eplicit emphasis on nearness of fit as this is implicitly obtained by the quasi-interpolation ansatz and the closeness of the generating functions to the pointwise optimal delta functions. This is achieved by the above problem formulation if the W (, i ) are weight functions that decrease with distance from the origin. Many of the radial functions used earlier are candidates for the weight functions. However, strict positive definiteness is not required, so that, e.g., (radial or tensor product) B-splines can also be used. The polynomial reproduction constraint is added so that the quasi-interpolant will achieve a desired approimation order. This will become clear in Section 7.6 below. 2. The smoothness functional (7.2) used here is also motivated by practical applications. In the Backus-Gilbert theory which was developed in the contet of geophysics (see [17]) it is desired that the generating functions Ψ i are as close as possible to the ideal cardinal functions (i.e., delta functions). Therefore, one needs to minimize their spread. The polynomial reproduction constraints correspond to discrete moment conditions for the function Ψ = Ψ(, ). If we think of as a fied (evaluation) point, then we have another constrained quadratic minimization problem of the form discussed in previous chapters. The unknowns are collected in the coefficient vector Ψ() = [Ψ(, 1 ),..., Ψ(, N )] T. The smoothness functional (7.2) 1 2 Ψ()T Q()Ψ() is given via the diagonal matri ( Q() = diag 1 W (, 1 ),..., 1 W (, N ) ), (7.4) where W (, i ) are positive weight functions (and thus for any the matri Q() is positive definite). The linear polynomial reproduction constraint (7.3) can be written in matri form as AΨ() = p(), where A is the m N matri with entries A ji = p j ( i ), i = 1,..., N, j = 1,..., m, and p = [p 1,..., p m ] T is a vector that contains a basis for the space Π s d of polynomials of degree d. According to our earlier work we use Lagrange multipliers and then know that (cf. (6.4) and (6.5)) λ() = ( AQ 1 ()A T ) 1 p() (7.5) Ψ() = Q 1 ()A T λ(). (7.6) 75
3 Equation (7.5) implies that the Lagrange multipliers are obtained as the solution of a Gram system G()λ() = p(), where the entries of G are the weighted l 2 inner products G jk () = p j, p k W () = p j ( i )p k ( i )W (, i ), j, k = 1,..., m. (7.7) The special feature here is that the weight varies with the evaluation point. Two short comments are called for. First, the Gram matri is symmetric and positive definite since the polynomial basis is linearly independent and the weights are positive. Second, in practice, the polynomials will be represented in shifted form, i.e., centered at the point of evaluation, so that only p 1 () 1 0. Equation (7.6) can be written componentwise, i.e., the generating functions in (7.1) are given by Ψ i () = W (, i ) λ j ()p j ( i ), i = 1,..., N. Therefore, once the values of the Lagrange multipliers λ j (), j = 1,..., N, have been determined we have eplicit formulas for the values of the generating functions. In general, however, finding the Lagrange multipliers involves solving a (small) linear system that changes as soon as changes. 7.2 Standard Interpretation of MLS Approimation We now consider the following approimation problem. Assume we are given data values f( i ), i = 1,..., N, on some set X = { 1,..., N } IR s of distinct data sites, where f is some (smooth) function, as well as an approimation space U = span{u 1,..., u m } (with m < N), along with the same weighted l 2 inner product f, g W () = f( i )g( i )W i (), IR s fied, (7.8) as introduced above in (7.7). Again, the positive weights W i, i = 1,..., N, depend on the evaluation point. We will interpret the weight functions in a way similar to radial basis functions, i.e., W i () = W (, i ), with the points i coming from the set X. We now wish to find the best approimation from U to f at the point with respect to the norm induced by (7.8). This means we will obtain the approimation (at the point ) as Pf() = c j ()u j (), (7.9) where the coefficients c j () are such that [Pf( i ) f( i )] 2 W i () (7.10) 76
4 is minimized. Due to the definition of the inner product (7.8) whose weight function moves with the evaluation point, the coefficients c j in (7.9) depend also on. As a consequence one has to solve the normal equations c j () u j, u k W () = f, u k W (), k = 1,..., m, (7.11) anew each time the evaluation point is changed. In matri notation (7.11) becomes G()c() = f u (), (7.12) with the positive definite Gram matri G() = ( ) m u j, u k W (), coefficient vector j,k=1 c() and right-hand side vector f u () as in (7.11) all depending on. In the moving least squares method one usually takes U to be a space of (multivariate) polynomials, i.e., Pf() = c j ()p j (), IR s, (7.13) where the {p 1,..., p m } is a basis for the space Π s d of s-variate polynomials of degree d. The Gram system (7.12) now becomes where the matri G() has entries G()c() = f p (), (7.14) G jk () = p j, p k W () = p j ( i )p k ( i )W (, i ), j, k = 1,..., m, (7.15) and the right-hand side vector consists of the projections of the data f onto the basis functions, i.e., f p () = [ f, p 1 W (),..., f, p m W () ] T. Remarks: 1. The fact that the coefficients depend on the evaluation point, and thus for every evaluation of Pf a Gram system (with different matri G()) needs to be solved, initially scared people away from the moving least squares approach. However, one can either choose compactly supported weight functions so that only a few terms are active in the sum in (7.15), or even completely avoid the solution of linear systems (see, e.g., [202]). 2. We point out that since we are working with a polynomial basis, the matri G can also be interpreted as a moment matri for the weight W. This interpretation is used in the engineering literature (see, e.g., [381]), and also plays an essential role when connecting moving least squares approimation to the more efficient concept of approimate approimation [434]. For a discussion of approimate moving least squares approimation see [203, 204, 205, 206]. 77
5 The connection to the constrained quadratic minimization problems discussed earlier can be seen as follows. We are now minimizing (for fied ) 1 2 ct ()G()c() µ T () [ G()c() AQ 1 ()f ], (7.16) where G() is the Gram matri (7.7), Q() the diagonal matri of weight functions (7.4) and A the matri of polynomials used earlier. The term AQ 1 ()f corresponds to the right-hand side vector f p () of (7.14). The solution of the linear system resulting from the minimization problem (7.16) gives us µ() = ( G()G 1 ()G T () ) 1 AQ 1 ()f = G T ()AQ 1 ()f c() = G 1 ()G T ()µ() = µ() so that as in the case of radial basis function interpolation by solving only the Gram system G()c() = f p () we automatically minimize the functional c T ()G()c() = = c j ()c k ()G jk () k=1 k=1 c j ()c k () p j, p k W () which should be interpreted as the native space norm of the approimant Pf() = c j ()p j (). 7.3 A Dual Representation for the Standard Approach We now know that on the one hand (from the Backus-Gilbert formulation) G()λ() = p() λ() = G 1 ()p(). (7.17) By taking multiple right-hand sides p() with X we get where the m N matri Λ has entries Λ ji = λ j ( i ). The standard formulation, on the other hand, gives us with Λ = G 1 ()A, (7.18) G()c() = f p () c() = G 1 ()f p () = G 1 ()AQ 1 ()f (7.19) f p () = [ f, p 1 W (),..., f, p m W () ] T = AQ 1 ()f as above. By combining (7.18) with (7.19) we get c() = G 1 ()AQ 1 ()f = ΛQ 1 ()f = f λ (), 78
6 where f λ () is defined analogously to f p (). Componentwise this gives us and therefore, c j () = f, λ j W (), j = 1,..., m, Pf() = f, λ j W () p j (). (7.20) It is also possible to formulate the moving least squares method by using the Lagrange multipliers of the Backus-Gilbert approach as basis functions for the approimation space U. Then, using the same argumentation as above, we end up with Pf() = d j ()λ j () (7.21) with d j () = f, p j W (), j = 1,..., m. Remarks: 1. The Lagrange multipliers form a basis that is dual to the polynomials. In particular one can show that for any X λ j, p k W () = δ jk, j, k = 1,..., m. This shows that we have two sets of basis functions that are bi-orthogonal on the set X. 2. Note that the epansions (7.20) and (7.21) are generalizations of (finite) eigenfunction or Fourier series epansions. 7.4 Equivalence of Our Approaches to Moving Least Squares Approimation We now show that the two main approaches to the moving least squares method described above are equivalent, i.e., we show that Pf() computed via (7.1) and (7.13) are the same. The approimant (7.1) in the Backus-Gilbert ansatz is of the form Pf() = f( i )Ψ i () = Ψ T ()f, where as before Ψ() = [Ψ(, 1 ),..., Ψ(, N )] T and f = [f( 1 ),..., f( N )] T. The standard moving least squares formulation (7.13), on the other hand, establishes Pf() in the form Pf() = c j ()p j () = p T ()c(), 79
7 where p = [p 1,..., p m ] T and c() = [c 1 (),..., c m ()] T. In (7.16) we wrote the normal equations for the standard approach as which implies Thus, using the standard approach, we get G()c() = AQ 1 ()f c() = G 1 ()AQ 1 ()f. Pf() = p T ()c() = p T ()G 1 ()AQ 1 ()f. (7.22) For the Backus-Gilbert approach we derived (see (7.5) and (7.6)) λ() = G 1 ()p() Ψ() = Q 1 ()A T λ(), where G() = AQ 1 ()A T (see (7.7) or (7.15)). Therefore, we now obtain Pf() = Ψ T ()f = [ Q 1 ()A T G 1 ()p() ] T f which, by the symmetry of Q() and G(), is the same as (7.22). Remarks: 1. The equivalence of the two approaches shows that the moving least squares approimant has all of the following properties: It reproduces any polynomial of degree at most d in s variables eactly. It produces the best locally weighted least squares fit. Viewed as a quasi-interpolant, the generating functions Ψ i are as close as possible to the optimal cardinal basis functions in the sense that (7.2) is minimized. Since polynomials are infinitely smooth, either of the representations of Pf shows that its smoothness is determined by the smoothness of the weight function(s) W i () = W (, i ). 2. In particular, the standard moving least squares method will reproduce the polynomial basis functions p 1,..., p m even though this is not eplicitly enforced by the minimization (solution of the normal equations). Moreover, the more general ansatz with approimation space U allows us to build moving least squares approimations that also reproduce any other function that is included in U. This can be very beneficial for the solution of partial differential equations with known singularities (see, e.g., the papers [16] by Babuška and Melenk, and [49] by Belytschko and co-authors). By also considering the dual epansion (7.21) we have three alternative representations for the moving least squares quasi-interpolant. This is summarized in the following theorem. 80
8 Theorem Let f : Ω IR be some function whose values on the set of points X = { i } N IRs are given as data. Let p 1,..., p m be a basis for Π s d, let {W (, i)} N be a set of positive weight functions centered at the points of X, and let λ j, j = 1,..., m, be the Lagrange multipliers defined by (7.5). Furthermore, consider the generating functions Ψ i () = W (, i ) λ j ()p j ( i ), i = 1,..., N. The best local least squares approimation to f on X in the sense of (7.10) is given by Pf() = = = f, λ j W () p j () f, p j W () λ j () f( i )Ψ i (). 7.5 Eamples Shepard s Method The moving least squares method in the case m = 1 with p 1 () 1 is known to yield Shepard s method [578]. In the statistics literature Shepard s method is known as a kernel method (see, e.g., the papers from the 1950s and 60s [534, 501, 476, 623]). Using our notation we have Pf() = c 1 (). The Gram matri consists of only one element G() = p 1, p 1 W () = W (, i ) so that implies The dual basis is defined by c 1 () = G()c() = f p () f( i )W (, i ) W (, i ) G()λ() = p(). 81
9 so that 1 λ 1 () =, W (, i ) and Pf() = d 1 ()λ 1 () (7.23) with d 1 () = f, p 1 W () = The generating functions are defined as f( i )W (, i ). Ψ i () = W (, i )λ 1 ()p 1 ( i ) = W (, i). W (, i ) This gives rise to the well-known quasi-interpolation formula for Shepard s method Pf() = = f( i )Ψ i () f( i ) W (, i ). W (, k ) k=1 Of course this is the same as the basis epansion c 1 () and the dual epansion (7.23). We should now have bi-orthogonality of the basis and dual basis, i.e., λ 1, p 1 W () = 1. Indeed λ 1, p 1 W () = = λ 1 ( i )W (, i ) W (, i ), W ( i, k ) k=1 and this equals 1 if we restrict to be an element of the set X Plots of Basis-Dual Basis Pairs We also illustrate the moving least squares basis functions, dual basis functions and generating functions for a one-dimensional eample with X being the set of 13 equally spaced points in [ 5, 5]. We take m = 2, i.e., we consider the case that ensures 82
10 Figure 7.1: Plot of Gaussian weight function centered at 7 = 0. reproduction of quadratic polynomials. The weight function is taken to be a standard Gaussian as depicted in Figure 7.1. The three basis polynomials p 1 () = 1, p 2 () =, and p 3 () = 2 are shown in Figure 7.2, whereas the dual basis functions λ 1, λ 2, and λ 3 are displayed in Figure 7.3. The figure shows that, ecept for the boundary effects caused by the finite interval, these functions resemble a quadratic, linear and constant polynomial Figure 7.2: Plot of three polynomial basis functions for moving least squares approimation. In Figure 7.4 we plot one of the generating functions (centered at 7 = 0) along with an approimate moving least squares generating function of the form Ψ(, y) = 1 ( ) 3 y 2 e y 2 σ σπ 2 σ with scale parameter σ as derived in [205]. 83
11 Figure 7.3: Plot of three dual basis functions for moving least squares approimation Figure 7.4: Plot of moving least squares generating function (left) and approimate generating function (right) centered at 7 = Approimation Order of Moving Least Squares Since the moving least squares approimants can be written as quasi-interpolants, we can use standard techniques to derive their point-wise error estimates. The standard argument proceeds as follows. Let f be a given (smooth) function that generates the data, i.e., f 1 = f( 1 ),..., f N = f( N ), and let p be an arbitrary polynomial. Moreover, assume that the moving least squares approimant is given in the form Pf() = f( i )Ψ i () with the generating functions Ψ i satisfying the polynomial reproduction property p( i )Ψ i () = p(), for all p Π s d, 84
12 as described at the beginning of this chapter. Then, due to the polynomial reproduction property of the generating functions, f() Pf() f() p() + p() = f() p() + f() p() + f p [1 + f( i )Ψ i () p( i )Ψ i () f( i )Ψ i () p( i ) f( i ) Ψ i () ] Ψ i (). (7.24) We see that in order to refine the error estimate we now have to answer two questions: How well do polynomials approimate f? This will be done with standard Taylor epansions. Are the generating functions bounded? The epression Ψ i () is known as the Lebesgue function, and finding a bound for the Lebesgue function is the main task that remains. By taking the polynomial p above to be the Taylor polynomial for f at of total degree d, the remainder term immediately yields an estimate of the form f p C 1 h d+1 ma Ω Dα f(), α = d + 1, where we have used the abbreviation = C 1 h d+1 f d+1, (7.25) f d+1 = ma Ω Dα f(), α = d + 1. Thus, if we can establish a uniform bound for the Lebesgue function, then (7.24) and (7.25) will result in f() Pf() Ch d+1 f d+1 which shows that moving least squares approimation with polynomial reproduction of degree d has approimation order O(h d+1 ). For Shepard s method, i.e., moving least squares approimation with constant reproduction (i.e., m = 1 or d = 0), we saw above that the generating functions are of the form Ψ i () = W (, i) W (, j ) 85
13 and therefore the Lebesgue function admits the uniform bound Ψ i () = 1, This shows that Shepard s method has approimation order O(h). Bounding the Lebesgue function in the general case is more involved and is the subject of the papers [378] by Levin and [632] by Wendland. This results in approimation order O(h d+1 ) for a moving least squares method that reproduces polynomials of degree d. In both papers the authors assumed that the weight function is compactly supported, and that the support size is scaled proportional to the fill distance. However, similar estimates should be possible if the weight function only decays fast enough (see, e.g., the survey by de Boor [60]). Aside from this consideration, the choice of weight function W does not play a role in determining the approimation order of the moving least squares method. As noted earlier, it only determines the smoothness of Pf. For eample, in the paper [146] from the statistics literature on local regression the authors state that often the choice [of weight function] is not too critical, and the use of the so-called tri-cube W (, i ) = (1 i 3 ) 3 +, IR s, is suggested. Of course, many other weight functions such as (radial) B-splines or any of the (bell-shaped) radial basis functions studied earlier can also be used. If the weight function is compactly supported, then the generating functions Ψ i will be so, too. This leads to computationally efficient methods since the Gram matri G() will be sparse. An interesting question is also the size of the support of the different local weight functions. Obviously, a fied support size for all weight functions is possible. However, this will cause serious problems as soon as the data are not uniformly distributed. Therefore, in the arguments in [378] and [632] the assumption is made that the data are at least quasi-uniformly distributed. Another choice for the support size of the individual weight functions is based on the number of nearest neighbors, i.e., the support size is chosen so that each of the local weight functions contains the same number of centers in its support. A third possibility is suggested by Schaback [556]. He proposes to use another moving least squares approimation based on (equally spaced) auiliary points to determine a smooth function δ so that at each evaluation point the radius of the support of the weight function is given by δ(). However, convergence estimates for these latter two choices do not eist. Sobolev error estimates are provided for moving least squares approimation with compactly supported weight functions in [7]. The rates obtained in that paper are not in terms of the fill distance but instead in terms of the support size R of the weight function. Moreover, it is assumed that for general s and m = ( ) s+d d the local Lagrange functions are bounded. As mentioned above, this is the hard part, and such bounds are only provided in the case s = 2 with d = 1 and d = 2 in [7]. However, if combined with the general bounds for the Lebesgue function provided by Wendland the paper [7] yields the following estimates: f() Pf() CR d+1 f d+1 86
14 but also In the weaker L 2 -norm we have (f Pf)() CR d f d+1. and f Pf L2 (B j Ω) CR d+1 f W d+1 2 (B j Ω) (f Pf) L2 (B j Ω) CR d f W d+1 2 (B j Ω), where the balls B j provide a finite cover of the domain Ω, i.e., Ω j B j, and the number of overlapping balls is bounded. Remarks: 1. In the statistics literature the moving least squares idea is known as local (polynomial) regression. There is a book by Fan and Gijbels [186] and a review article by Cleveland and Loader [146] according to which the basic ideas of local regression can be traced back at least to work of Gram [267], Woolhouse [648], and De Forest [148, 149] from the 1870s and 1880s. 2. In particular, in the statistics literature one learns that the use of least squares approimation is justified when the data f 1,..., f N are normally distributed, whereas, if the noise in the data is not Gaussian, then other criteria should be used. See, e.g., the survey article [146] for more details. 3. The general moving least squares method first appeared in the approimation theory literature in a paper by Lancaster and Šalkauskas [358] who also pointed out the connection to earlier (more specialized) work by Shepard [578] and McLain [436]. 4. Early error estimates for some special cases were provided by Farwig in [188, 189]. 87
Inner Product Spaces
Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and
MATH 590: Meshfree Methods
MATH 590: Meshfree Methods Chapter 7: Conditionally Positive Definite Functions Greg Fasshauer Department of Applied Mathematics Illinois Institute of Technology Fall 2010 [email protected] MATH 590 Chapter
EECS 556 Image Processing W 09. Interpolation. Interpolation techniques B splines
EECS 556 Image Processing W 09 Interpolation Interpolation techniques B splines What is image processing? Image processing is the application of 2D signal processing methods to images Image representation
Mathematics 31 Pre-calculus and Limits
Mathematics 31 Pre-calculus and Limits Overview After completing this section, students will be epected to have acquired reliability and fluency in the algebraic skills of factoring, operations with radicals
Metric Spaces. Chapter 7. 7.1. Metrics
Chapter 7 Metric Spaces A metric space is a set X that has a notion of the distance d(x, y) between every pair of points x, y X. The purpose of this chapter is to introduce metric spaces and give some
Section 1.3 P 1 = 1 2. = 1 4 2 8. P n = 1 P 3 = Continuing in this fashion, it should seem reasonable that, for any n = 1, 2, 3,..., = 1 2 4.
Difference Equations to Differential Equations Section. The Sum of a Sequence This section considers the problem of adding together the terms of a sequence. Of course, this is a problem only if more than
BANACH AND HILBERT SPACE REVIEW
BANACH AND HILBET SPACE EVIEW CHISTOPHE HEIL These notes will briefly review some basic concepts related to the theory of Banach and Hilbert spaces. We are not trying to give a complete development, but
Increasing for all. Convex for all. ( ) Increasing for all (remember that the log function is only defined for ). ( ) Concave for all.
1. Differentiation The first derivative of a function measures by how much changes in reaction to an infinitesimal shift in its argument. The largest the derivative (in absolute value), the faster is evolving.
1 Review of Least Squares Solutions to Overdetermined Systems
cs4: introduction to numerical analysis /9/0 Lecture 7: Rectangular Systems and Numerical Integration Instructor: Professor Amos Ron Scribes: Mark Cowlishaw, Nathanael Fillmore Review of Least Squares
Elasticity Theory Basics
G22.3033-002: Topics in Computer Graphics: Lecture #7 Geometric Modeling New York University Elasticity Theory Basics Lecture #7: 20 October 2003 Lecturer: Denis Zorin Scribe: Adrian Secord, Yotam Gingold
Introduction to General and Generalized Linear Models
Introduction to General and Generalized Linear Models General Linear Models - part I Henrik Madsen Poul Thyregod Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby
4.3 Lagrange Approximation
206 CHAP. 4 INTERPOLATION AND POLYNOMIAL APPROXIMATION Lagrange Polynomial Approximation 4.3 Lagrange Approximation Interpolation means to estimate a missing function value by taking a weighted average
The Heat Equation. Lectures INF2320 p. 1/88
The Heat Equation Lectures INF232 p. 1/88 Lectures INF232 p. 2/88 The Heat Equation We study the heat equation: u t = u xx for x (,1), t >, (1) u(,t) = u(1,t) = for t >, (2) u(x,) = f(x) for x (,1), (3)
15. Symmetric polynomials
15. Symmetric polynomials 15.1 The theorem 15.2 First examples 15.3 A variant: discriminants 1. The theorem Let S n be the group of permutations of {1,, n}, also called the symmetric group on n things.
Similarity and Diagonalization. Similar Matrices
MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that
MATH 551 - APPLIED MATRIX THEORY
MATH 55 - APPLIED MATRIX THEORY FINAL TEST: SAMPLE with SOLUTIONS (25 points NAME: PROBLEM (3 points A web of 5 pages is described by a directed graph whose matrix is given by A Do the following ( points
This unit will lay the groundwork for later units where the students will extend this knowledge to quadratic and exponential functions.
Algebra I Overview View unit yearlong overview here Many of the concepts presented in Algebra I are progressions of concepts that were introduced in grades 6 through 8. The content presented in this course
FINITE DIFFERENCE METHODS
FINITE DIFFERENCE METHODS LONG CHEN Te best known metods, finite difference, consists of replacing eac derivative by a difference quotient in te classic formulation. It is simple to code and economic to
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem
Continued Fractions and the Euclidean Algorithm
Continued Fractions and the Euclidean Algorithm Lecture notes prepared for MATH 326, Spring 997 Department of Mathematics and Statistics University at Albany William F Hammond Table of Contents Introduction
Random variables, probability distributions, binomial random variable
Week 4 lecture notes. WEEK 4 page 1 Random variables, probability distributions, binomial random variable Eample 1 : Consider the eperiment of flipping a fair coin three times. The number of tails that
LINEAR INEQUALITIES. less than, < 2x + 5 x 3 less than or equal to, greater than, > 3x 2 x 6 greater than or equal to,
LINEAR INEQUALITIES When we use the equal sign in an equation we are stating that both sides of the equation are equal to each other. In an inequality, we are stating that both sides of the equation are
Introduction to Matrix Algebra
Psychology 7291: Multivariate Statistics (Carey) 8/27/98 Matrix Algebra - 1 Introduction to Matrix Algebra Definitions: A matrix is a collection of numbers ordered by rows and columns. It is customary
POLYNOMIAL HISTOPOLATION, SUPERCONVERGENT DEGREES OF FREEDOM, AND PSEUDOSPECTRAL DISCRETE HODGE OPERATORS
POLYNOMIAL HISTOPOLATION, SUPERCONVERGENT DEGREES OF FREEDOM, AND PSEUDOSPECTRAL DISCRETE HODGE OPERATORS N. ROBIDOUX Abstract. We show that, given a histogram with n bins possibly non-contiguous or consisting
Inner product. Definition of inner product
Math 20F Linear Algebra Lecture 25 1 Inner product Review: Definition of inner product. Slide 1 Norm and distance. Orthogonal vectors. Orthogonal complement. Orthogonal basis. Definition of inner product
Applications to Data Smoothing and Image Processing I
Applications to Data Smoothing and Image Processing I MA 348 Kurt Bryan Signals and Images Let t denote time and consider a signal a(t) on some time interval, say t. We ll assume that the signal a(t) is
15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
The sample space for a pair of die rolls is the set. The sample space for a random number between 0 and 1 is the interval [0, 1].
Probability Theory Probability Spaces and Events Consider a random experiment with several possible outcomes. For example, we might roll a pair of dice, flip a coin three times, or choose a random real
MATHEMATICAL METHODS OF STATISTICS
MATHEMATICAL METHODS OF STATISTICS By HARALD CRAMER TROFESSOK IN THE UNIVERSITY OF STOCKHOLM Princeton PRINCETON UNIVERSITY PRESS 1946 TABLE OF CONTENTS. First Part. MATHEMATICAL INTRODUCTION. CHAPTERS
POLYNOMIAL FUNCTIONS
POLYNOMIAL FUNCTIONS Polynomial Division.. 314 The Rational Zero Test.....317 Descarte s Rule of Signs... 319 The Remainder Theorem.....31 Finding all Zeros of a Polynomial Function.......33 Writing a
Lecture 7: Finding Lyapunov Functions 1
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.243j (Fall 2003): DYNAMICS OF NONLINEAR SYSTEMS by A. Megretski Lecture 7: Finding Lyapunov Functions 1
5 Numerical Differentiation
D. Levy 5 Numerical Differentiation 5. Basic Concepts This chapter deals with numerical approximations of derivatives. The first questions that comes up to mind is: why do we need to approximate derivatives
Nonlinear Iterative Partial Least Squares Method
Numerical Methods for Determining Principal Component Analysis Abstract Factors Béchu, S., Richard-Plouet, M., Fernandez, V., Walton, J., and Fairley, N. (2016) Developments in numerical treatments for
Fitting Subject-specific Curves to Grouped Longitudinal Data
Fitting Subject-specific Curves to Grouped Longitudinal Data Djeundje, Viani Heriot-Watt University, Department of Actuarial Mathematics & Statistics Edinburgh, EH14 4AS, UK E-mail: [email protected] Currie,
Nonlinear Systems of Ordinary Differential Equations
Differential Equations Massoud Malek Nonlinear Systems of Ordinary Differential Equations Dynamical System. A dynamical system has a state determined by a collection of real numbers, or more generally
Høgskolen i Narvik Sivilingeniørutdanningen STE6237 ELEMENTMETODER. Oppgaver
Høgskolen i Narvik Sivilingeniørutdanningen STE637 ELEMENTMETODER Oppgaver Klasse: 4.ID, 4.IT Ekstern Professor: Gregory A. Chechkin e-mail: [email protected] Narvik 6 PART I Task. Consider two-point
Section 6.1 - Inner Products and Norms
Section 6.1 - Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,
Practical Guide to the Simplex Method of Linear Programming
Practical Guide to the Simplex Method of Linear Programming Marcel Oliver Revised: April, 0 The basic steps of the simplex algorithm Step : Write the linear programming problem in standard form Linear
1.5 SOLUTION SETS OF LINEAR SYSTEMS
1-2 CHAPTER 1 Linear Equations in Linear Algebra 1.5 SOLUTION SETS OF LINEAR SYSTEMS Many of the concepts and computations in linear algebra involve sets of vectors which are visualized geometrically as
THE ELECTROMAGNETIC FIELD DUE TO THE ELECTRONS.
THE ELECTROMAGNETIC FIELD DUE TO THE ELECTRONS 367 Proceedings of the London Mathematical Society Vol 1 1904 p 367-37 (Retyped for readability with same page breaks) ON AN EXPRESSION OF THE ELECTROMAGNETIC
Least-Squares Intersection of Lines
Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a
Regression III: Advanced Methods
Lecture 16: Generalized Additive Models Regression III: Advanced Methods Bill Jacoby Michigan State University http://polisci.msu.edu/jacoby/icpsr/regress3 Goals of the Lecture Introduce Additive Models
Two Topics in Parametric Integration Applied to Stochastic Simulation in Industrial Engineering
Two Topics in Parametric Integration Applied to Stochastic Simulation in Industrial Engineering Department of Industrial Engineering and Management Sciences Northwestern University September 15th, 2014
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written
G. GRAPHING FUNCTIONS
G. GRAPHING FUNCTIONS To get a quick insight int o how the graph of a function looks, it is very helpful to know how certain simple operations on the graph are related to the way the function epression
1 Introduction to Matrices
1 Introduction to Matrices In this section, important definitions and results from matrix algebra that are useful in regression analysis are introduced. While all statements below regarding the columns
An Introduction to Partial Differential Equations in the Undergraduate Curriculum
An Introduction to Partial Differential Equations in the Undergraduate Curriculum J. Tolosa & M. Vajiac LECTURE 11 Laplace s Equation in a Disk 11.1. Outline of Lecture The Laplacian in Polar Coordinates
7 Gaussian Elimination and LU Factorization
7 Gaussian Elimination and LU Factorization In this final section on matrix factorization methods for solving Ax = b we want to take a closer look at Gaussian elimination (probably the best known method
Parabolic Equations. Chapter 5. Contents. 5.1.2 Well-Posed Initial-Boundary Value Problem. 5.1.3 Time Irreversibility of the Heat Equation
7 5.1 Definitions Properties Chapter 5 Parabolic Equations Note that we require the solution u(, t bounded in R n for all t. In particular we assume that the boundedness of the smooth function u at infinity
Homework 2 Solutions
Homework Solutions 1. (a) Find the area of a regular heagon inscribed in a circle of radius 1. Then, find the area of a regular heagon circumscribed about a circle of radius 1. Use these calculations to
Bindel, Spring 2012 Intro to Scientific Computing (CS 3220) Week 3: Wednesday, Feb 8
Spaces and bases Week 3: Wednesday, Feb 8 I have two favorite vector spaces 1 : R n and the space P d of polynomials of degree at most d. For R n, we have a canonical basis: R n = span{e 1, e 2,..., e
Linear Algebra I. Ronald van Luijk, 2012
Linear Algebra I Ronald van Luijk, 2012 With many parts from Linear Algebra I by Michael Stoll, 2007 Contents 1. Vector spaces 3 1.1. Examples 3 1.2. Fields 4 1.3. The field of complex numbers. 6 1.4.
Polynomial and Synthetic Division. Long Division of Polynomials. Example 1. 6x 2 7x 2 x 2) 19x 2 16x 4 6x3 12x 2 7x 2 16x 7x 2 14x. 2x 4.
_.qd /7/5 9: AM Page 5 Section.. Polynomial and Synthetic Division 5 Polynomial and Synthetic Division What you should learn Use long division to divide polynomials by other polynomials. Use synthetic
Support Vector Machines Explained
March 1, 2009 Support Vector Machines Explained Tristan Fletcher www.cs.ucl.ac.uk/staff/t.fletcher/ Introduction This document has been written in an attempt to make the Support Vector Machines (SVM),
Chapter 4. Polynomial and Rational Functions. 4.1 Polynomial Functions and Their Graphs
Chapter 4. Polynomial and Rational Functions 4.1 Polynomial Functions and Their Graphs A polynomial function of degree n is a function of the form P = a n n + a n 1 n 1 + + a 2 2 + a 1 + a 0 Where a s
Computational Geometry Lab: FEM BASIS FUNCTIONS FOR A TETRAHEDRON
Computational Geometry Lab: FEM BASIS FUNCTIONS FOR A TETRAHEDRON John Burkardt Information Technology Department Virginia Tech http://people.sc.fsu.edu/ jburkardt/presentations/cg lab fem basis tetrahedron.pdf
Numerical Analysis Lecture Notes
Numerical Analysis Lecture Notes Peter J. Olver 6. Eigenvalues and Singular Values In this section, we collect together the basic facts about eigenvalues and eigenvectors. From a geometrical viewpoint,
Support Vector Machines with Clustering for Training with Very Large Datasets
Support Vector Machines with Clustering for Training with Very Large Datasets Theodoros Evgeniou Technology Management INSEAD Bd de Constance, Fontainebleau 77300, France [email protected] Massimiliano
Optimal shift scheduling with a global service level constraint
Optimal shift scheduling with a global service level constraint Ger Koole & Erik van der Sluis Vrije Universiteit Division of Mathematics and Computer Science De Boelelaan 1081a, 1081 HV Amsterdam The
AN INTRODUCTION TO NUMERICAL METHODS AND ANALYSIS
AN INTRODUCTION TO NUMERICAL METHODS AND ANALYSIS Revised Edition James Epperson Mathematical Reviews BICENTENNIAL 0, 1 8 0 7 z ewiley wu 2007 r71 BICENTENNIAL WILEY-INTERSCIENCE A John Wiley & Sons, Inc.,
Section 4.4 Inner Product Spaces
Section 4.4 Inner Product Spaces In our discussion of vector spaces the specific nature of F as a field, other than the fact that it is a field, has played virtually no role. In this section we no longer
arxiv:hep-th/0507236v1 25 Jul 2005
Non perturbative series for the calculation of one loop integrals at finite temperature Paolo Amore arxiv:hep-th/050736v 5 Jul 005 Facultad de Ciencias, Universidad de Colima, Bernal Diaz del Castillo
4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION
4: EIGENVALUES, EIGENVECTORS, DIAGONALIZATION STEVEN HEILMAN Contents 1. Review 1 2. Diagonal Matrices 1 3. Eigenvectors and Eigenvalues 2 4. Characteristic Polynomial 4 5. Diagonalizability 6 6. Appendix:
ALMOST COMMON PRIORS 1. INTRODUCTION
ALMOST COMMON PRIORS ZIV HELLMAN ABSTRACT. What happens when priors are not common? We introduce a measure for how far a type space is from having a common prior, which we term prior distance. If a type
MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.
MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column
About the Gamma Function
About the Gamma Function Notes for Honors Calculus II, Originally Prepared in Spring 995 Basic Facts about the Gamma Function The Gamma function is defined by the improper integral Γ) = The integral is
Point Lattices in Computer Graphics and Visualization how signal processing may help computer graphics
Point Lattices in Computer Graphics and Visualization how signal processing may help computer graphics Dimitri Van De Ville Ecole Polytechnique Fédérale de Lausanne Biomedical Imaging Group [email protected]
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
MATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets.
MATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets. Norm The notion of norm generalizes the notion of length of a vector in R n. Definition. Let V be a vector space. A function α
FUNCTIONAL ANALYSIS LECTURE NOTES: QUOTIENT SPACES
FUNCTIONAL ANALYSIS LECTURE NOTES: QUOTIENT SPACES CHRISTOPHER HEIL 1. Cosets and the Quotient Space Any vector space is an abelian group under the operation of vector addition. So, if you are have studied
Introduction to the Finite Element Method
Introduction to the Finite Element Method 09.06.2009 Outline Motivation Partial Differential Equations (PDEs) Finite Difference Method (FDM) Finite Element Method (FEM) References Motivation Figure: cross
MATH 4330/5330, Fourier Analysis Section 11, The Discrete Fourier Transform
MATH 433/533, Fourier Analysis Section 11, The Discrete Fourier Transform Now, instead of considering functions defined on a continuous domain, like the interval [, 1) or the whole real line R, we wish
Section 3-3 Approximating Real Zeros of Polynomials
- Approimating Real Zeros of Polynomials 9 Section - Approimating Real Zeros of Polynomials Locating Real Zeros The Bisection Method Approimating Multiple Zeros Application The methods for finding zeros
Support Vector Machine (SVM)
Support Vector Machine (SVM) CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
Sensitivity Analysis 3.1 AN EXAMPLE FOR ANALYSIS
Sensitivity Analysis 3 We have already been introduced to sensitivity analysis in Chapter via the geometry of a simple example. We saw that the values of the decision variables and those of the slack and
Work as the Area Under a Graph of Force vs. Displacement
Work as the Area Under a Graph of vs. Displacement Situation A. Consider a situation where an object of mass, m, is lifted at constant velocity in a uniform gravitational field, g. The applied force is
Example 4.1 (nonlinear pendulum dynamics with friction) Figure 4.1: Pendulum. asin. k, a, and b. We study stability of the origin x
Lecture 4. LaSalle s Invariance Principle We begin with a motivating eample. Eample 4.1 (nonlinear pendulum dynamics with friction) Figure 4.1: Pendulum Dynamics of a pendulum with friction can be written
Constrained optimization.
ams/econ 11b supplementary notes ucsc Constrained optimization. c 2010, Yonatan Katznelson 1. Constraints In many of the optimization problems that arise in economics, there are restrictions on the values
Statistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
Derivative Approximation by Finite Differences
Derivative Approximation by Finite Differences David Eberly Geometric Tools, LLC http://wwwgeometrictoolscom/ Copyright c 998-26 All Rights Reserved Created: May 3, 2 Last Modified: April 25, 25 Contents
The Matrix Elements of a 3 3 Orthogonal Matrix Revisited
Physics 116A Winter 2011 The Matrix Elements of a 3 3 Orthogonal Matrix Revisited 1. Introduction In a class handout entitled, Three-Dimensional Proper and Improper Rotation Matrices, I provided a derivation
4.5 Linear Dependence and Linear Independence
4.5 Linear Dependence and Linear Independence 267 32. {v 1, v 2 }, where v 1, v 2 are collinear vectors in R 3. 33. Prove that if S and S are subsets of a vector space V such that S is a subset of S, then
Survey and remarks on Viro s definition of Khovanov homology
RIMS Kôkyûroku Bessatsu B39 (203), 009 09 Survey and remarks on Viro s definition of Khovanov homology By Noboru Ito Abstract This paper reviews and offers remarks upon Viro s definition of the Khovanov
Spatial Statistics Chapter 3 Basics of areal data and areal data modeling
Spatial Statistics Chapter 3 Basics of areal data and areal data modeling Recall areal data also known as lattice data are data Y (s), s D where D is a discrete index set. This usually corresponds to data
Linear Algebra Notes for Marsden and Tromba Vector Calculus
Linear Algebra Notes for Marsden and Tromba Vector Calculus n-dimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of
Correlation and Convolution Class Notes for CMSC 426, Fall 2005 David Jacobs
Correlation and Convolution Class otes for CMSC 46, Fall 5 David Jacobs Introduction Correlation and Convolution are basic operations that we will perform to extract information from images. They are in
Math 4310 Handout - Quotient Vector Spaces
Math 4310 Handout - Quotient Vector Spaces Dan Collins The textbook defines a subspace of a vector space in Chapter 4, but it avoids ever discussing the notion of a quotient space. This is understandable
1 Prior Probability and Posterior Probability
Math 541: Statistical Theory II Bayesian Approach to Parameter Estimation Lecturer: Songfeng Zheng 1 Prior Probability and Posterior Probability Consider now a problem of statistical inference in which
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES Contents 1. Random variables and measurable functions 2. Cumulative distribution functions 3. Discrete
Introduction to Support Vector Machines. Colin Campbell, Bristol University
Introduction to Support Vector Machines Colin Campbell, Bristol University 1 Outline of talk. Part 1. An Introduction to SVMs 1.1. SVMs for binary classification. 1.2. Soft margins and multi-class classification.
Derivative Free Optimization
Department of Mathematics Derivative Free Optimization M.J.D. Powell LiTH-MAT-R--2014/02--SE Department of Mathematics Linköping University S-581 83 Linköping, Sweden. Three lectures 1 on Derivative Free
Algebra Unpacked Content For the new Common Core standards that will be effective in all North Carolina schools in the 2012-13 school year.
This document is designed to help North Carolina educators teach the Common Core (Standard Course of Study). NCDPI staff are continually updating and improving these tools to better serve teachers. Algebra
Vector and Matrix Norms
Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a non-empty
Piecewise Cubic Splines
280 CHAP. 5 CURVE FITTING Piecewise Cubic Splines The fitting of a polynomial curve to a set of data points has applications in CAD (computer-assisted design), CAM (computer-assisted manufacturing), and
Multi-variable Calculus and Optimization
Multi-variable Calculus and Optimization Dudley Cooke Trinity College Dublin Dudley Cooke (Trinity College Dublin) Multi-variable Calculus and Optimization 1 / 51 EC2040 Topic 3 - Multi-variable Calculus
INTEREST RATES AND FX MODELS
INTEREST RATES AND FX MODELS 8. Portfolio greeks Andrew Lesniewski Courant Institute of Mathematical Sciences New York University New York March 27, 2013 2 Interest Rates & FX Models Contents 1 Introduction
Partial Fractions Decomposition
Partial Fractions Decomposition Dr. Philippe B. Laval Kennesaw State University August 6, 008 Abstract This handout describes partial fractions decomposition and how it can be used when integrating rational
