# The Method of Lagrange Multipliers

 To view this video please enable JavaScript, and consider upgrading to a web browser that supports HTML5 video
Save this PDF as:

Size: px
Start display at page:

## Transcription

1 The Method of Lagrange Multipliers S. Sawyer October 25, Lagrange s Theorem. Suppose that we want to maximize (or imize a function of n variables f(x = f(x 1, x 2,..., x n for x = (x 1, x 2,..., x n (1.1a p constraints g 1 (x = c 1, g 2 (x = c 2,..., and g p (x = c p (1.1b As an example for p = 1, find { n x 2 i : or x 1,...,x 5 x 1,...,x n 5 x 2 i } x i = 1 A first guess for (1.1 (with f(x = n solutions of the n equations { x1 + 2x 2 + x 3 = 1 and x 3 2x 4 + x 5 = 6 x2 i (1.2a (1.2b in (1.2 might be to look for x i f(x = 0, 1 i n (1.3 However, this leads to x i = 0 in (1.2, which does not satisfy the constraint. Lagrange s solution is to introduce p new parameters (called Lagrange Multipliers and then solve a more complicated problem: Theorem (Lagrange Assug appropriate smoothness conditions, imum or maximum of f(x the constraints (1.1b that is not on the boundary of the region where f(x and g j (x are defined can be found by introducing p new parameters λ 1, λ 2,..., λ p and solving the system p f(x + λ j g j (x = 0, 1 i n (1.4a x i j=1 g j (x = c j, 1 j p (1.4b This amounts to solving n+p equations for the n+p real variables in x and λ. In contrast, (1.3 has n equations for the n unknowns in x. Fortunately, the system (1.4 is often easy to solve, and is usually much easier than using the constraints to substitute for some of the x i.

2 The Method of Lagrange Multipliers Examples. (1 There are p = 1 constraints in (1.2a, so that (1.4a becomes x 2 k + λ x k = 2x i + λ = 0, 1 i n x i Thus x i = λ/2 for 1 i n. The constraint n x i = 1 in (1.4b implies n x i = nλ/2 = 1 or λ = 2/n, from which x i = 1/n for 1 i n. For x i = 1/n, f(x = n/n 2 = 1/n. One can check that this is a imum as opposed to a maximum or saddle point by noting that f(x = 1 if x 1 = 1, x i = 0 for 2 i n. (2 A System with Two Constraints: There are p = 2 constraints in (1.2b, which is to find 5 x 2 i { x1 + 2x 2 + x 3 = 1 and x 3 2x 4 + x 5 = 6 (2.1 The method of Lagrange multipliers says to look for solutions of ( 5 x 2 k + λ(x 1 + 2x 2 + x 3 + µ(x 3 2x 4 + x 5 = 0 (2.2 x i where we write λ, µ for the two Lagrange multipliers λ 1, λ 2. The equations (2.2 imply 2x 1 + λ = 0, 2x 2 + 2λ = 0, 2x 3 + λ + µ = 0, 2x 4 2µ = 0, and 2x 5 + µ = 0. Combining the first three equations with the first constraint in (2.1 implies 2 + 6λ + µ = 0. Combining the last three equations in (2.2 with the second constraint in (2.1 implies 12+λ+6µ = 0. Thus 6λ + µ = 2 λ + 6µ = 12 Adding these two equations implies 7(λ + µ = 14 or λ + µ = 2. Subtracting the equations implies 5(λ µ = 10 or λ µ = 2. Thus (λ + µ + (λ µ = 2λ = 0 and λ = 0, µ = 2. This implies x 1 = x 2 = 0, x 3 = x 5 = 1, and x 4 = 2. The imum value in (2.1 is 6. (3 A BLUE problem: Let X 1,..., X n be independent random variables with E(X i = µ and Var(X i = σ 2 i. Find the coefficients a i that imize Var a i X i E a i X i = µ (2.3

3 The Method of Lagrange Multipliers This asks us to find the Best Linear Unbiased Estimator n a ix i (abbreviated BLUE for µ for given values of σi 2. Since Var(aX = a 2 Var(X and Var(X + Y = Var(X + Var(Y for independent random variables X and Y, we have Var( n a ix i = n a2 i Var(X i = n a2 i σ2 i. Thus (2.3 is equivalent to finding a 2 i σi 2 a i = 1 Using one Lagrange multiplier λ for the constraint leads to the equations 2a i σi 2 + λ = 0 or a i = λ/(2σi 2. The constraint n a i = 1 then implies that the BLUE for µ is / n a i X i where a i = c/σi 2 for c = 1 (1/σk 2 (2.4 If σi 2 = σ2 for all i, then a i = 1/n and n a ix i = (1/n n X i = X is the BLUE for µ. Conversely, if Var(X i = σi 2 is variable, then the BLUE n a ix i for µ puts relatively less weight on the noisier (higher-variance observations (that is, the weight a i is smaller, but still uses the information in the noiser observations. Formulas like (2.4 are often used in survey sampling. 3. A Short Proof of Lagrange s Theorem. The extremal condition (1.3 (without any constraints can be written in vector form as f(x = Now by Taylor s Theorem ( x 1 f(x, x 2 f(x,..., f(x x n = 0 (3.1 f(x + hy = f(x + hy f(x + O(h 2 (3.2 where h is a scalar, O(h 2 denotes terms that are bounded by h 2, and x y is the dot product. Thus (3.1 gives the vector direction in which f(x changes the most per unit change in x, where unit change in measured in terms of the length of the vector x. In particular, if y = f(x 0, then f(x hy < f(x < f(x + hy

4 The Method of Lagrange Multipliers for sufficiently small values of h, and the only way that x can be a local imum or maximum would be if x were on the boundary of the set of points where f(x is defined. This implies that f(x = 0 at non-boundary imum and maximum values of f(x. Now consider the problem of finding max f(x g(x = c (3.3 for one constraint. If x = x 1 (t is a path in the surface defined by g(x = c, then by the chain rule d dt g( x 1 (0 = d dt x 1(0 g ( x 1 (0 = 0 (3.4 This implies that g ( x 1 (0 is orthogonal to the tangent vector (d/dtx 1 (0 for any path x 1 (t in the surface defined by g(x = c. Conversely, if x is any point in the surface g(x = c and y is any vector such that y g(x = 0, then it follows from the Implicit Function Theorem there exists a path x 1 (t in the surface g(x = c such that x 1 (0 = x and (d/dtx 1 (0 = y. This result and (3.4 imply that the gradient vector g(x is always orthogonal to the surface defined by g(x = c at x. Now let x be a solution of (3.3. I claim that f(x = λ g(x for some scalar λ. First, we can always write f(x = c g(x+y where y g(x = 0. If x(t is a path in the surface with x(0 = x and (d/dtx(0 f(x = 0, it follows from (3.2 with y = (d/dtx(0 that there are values for f(x for x = x(t in the surface that both larger and smaller than f(x. Thus, if x is a maximum of imum of f(x in the surface and f(x = c g(x + y for y g(x = 0, then y f(x = y g(x + y y = y y = 0 and y = 0. This means that f(x = c g(x, which completes the proof of Lagrange s Theorem for one constraint (p = 1. Next, suppose that we want to solve max f(x g 1 (x = c 1,..., g p (x = c p (3.5 for p constraints. Let x be a solution of (3.5. Recall that the each vector g j (x is orthogonal to the surface g j (x = c j at x. Let L be the linear space L = span{ g j (x : 1 j p } I claim that f(x L. This would imply f(x = p λ j g j (x j=1

5 The Method of Lagrange Multipliers for some choice of scalar values λ j, which would prove Lagrange s Theorem. To prove that f(x L, first note that, in general, we can write f(x = w + y where w L and y is perpendicular to L, which means that y z = 0 for any z L. In particular, y g j (x = 0 for 1 j p. Now find a path x 1 (t through x in the intersection of the surfaces g j (x = c j such that x 1 (0 = x and (d/dtx 1 (0 = y. (The existence of such a path for sufficiently small t follows from a stronger form of the Implicit Function Theorem. It then follows from (3.2 and (3.5 that y f(x = 0. Since f(x = w + y where y w = 0, it follows that y f(x = y w + y y = y y = 0 and y = 0, This implies that f(x = w L, which completes the proof of Lagrange s Theorem. 4. Warnings. The same warnings apply here as for most methods for finding a maximum or imum: The system (1.4 does not look for a maximum (or imum of f(x constraints g j (x = c j, but only a point x on the set of values detered by g j (x = c j whose first-order changes in x are zero. This is satisfied by a value x = x 0 that provides a imum or maximum typical for f(x in a neighborhood of x 0, but may only be a local imum or maximum. There may be several local ima or maxima, each yielding a solution of (1.4. The criterion (1.4 also holds for saddle points of f(x that are local maxima in some directions or coordinates and local ima in others. In these cases, the different values f(x at the solutions of (1.4 have to be evaluated individually to find the global maximum. A particular situation to avoid is to look for a maximum value of f(x by solving (1.4 or (1.3 when f(x takes arbitrarily large values when any of the components of x are large (as is the case for f(x in (1.2 and (1.4 has a unique solution x 0. In that case, x 0 is probably the global imum of f(x the constraints, and not a maximum. In that case, rather than find the best possible value of f(x, one may end up with the worst possible value. After solving (1.3 or (1.4, one often has to look at the problem more carefully to see if it is a global maximum, a global imum, or neither. Another situation to avoid is when the maximum or imum is on the boundary of the values for which f(x is defined. In that case, the maximum or imum is not an interior value, and the first-order changes in f(x (that is, the partial derivatives of f(x may not be zero at that point. An example is f(x = x on the unit interval 0 x 1. The imum value of f(x = x on the interval is x = 0 and the maximum is x = 1, but neither are solutions of f (x = 0.

### HOMEWORK 4 SOLUTIONS. All questions are from Vector Calculus, by Marsden and Tromba

HOMEWORK SOLUTIONS All questions are from Vector Calculus, by Marsden and Tromba Question :..6 Let w = f(x, y) be a function of two variables, and let x = u + v, y = u v. Show that Solution. By the chain

### PROBLEM SET. Practice Problems for Exam #2. Math 2350, Fall Nov. 7, 2004 Corrected Nov. 10 ANSWERS

PROBLEM SET Practice Problems for Exam #2 Math 2350, Fall 2004 Nov. 7, 2004 Corrected Nov. 10 ANSWERS i Problem 1. Consider the function f(x, y) = xy 2 sin(x 2 y). Find the partial derivatives f x, f y,

### (a) We have x = 3 + 2t, y = 2 t, z = 6 so solving for t we get the symmetric equations. x 3 2. = 2 y, z = 6. t 2 2t + 1 = 0,

Name: Solutions to Practice Final. Consider the line r(t) = 3 + t, t, 6. (a) Find symmetric equations for this line. (b) Find the point where the first line r(t) intersects the surface z = x + y. (a) We

### Problem 1 (10 pts) Find the radius of convergence and interval of convergence of the series

1 Problem 1 (10 pts) Find the radius of convergence and interval of convergence of the series a n n=1 n(x + 2) n 5 n 1. n(x + 2)n Solution: Do the ratio test for the absolute convergence. Let a n =. Then,

### The variable λ is a dummy variable called a Lagrange multiplier ; we only really care about the values of x, y, and z.

Math a Lagrange Multipliers Spring, 009 The method of Lagrange multipliers allows us to maximize or minimize functions with the constraint that we only consider points on a certain surface To find critical

### SOLUTIONS. f x = 6x 2 6xy 24x, f y = 3x 2 6y. To find the critical points, we solve

SOLUTIONS Problem. Find the critical points of the function f(x, y = 2x 3 3x 2 y 2x 2 3y 2 and determine their type i.e. local min/local max/saddle point. Are there any global min/max? Partial derivatives

### 1 3 4 = 8i + 20j 13k. x + w. y + w

) Find the point of intersection of the lines x = t +, y = 3t + 4, z = 4t + 5, and x = 6s + 3, y = 5s +, z = 4s + 9, and then find the plane containing these two lines. Solution. Solve the system of equations

### Multivariable Calculus Practice Midterm 2 Solutions Prof. Fedorchuk

Multivariable Calculus Practice Midterm Solutions Prof. Fedorchuk. ( points) Let f(x, y, z) xz + e y x. a. (4 pts) Compute the gradient f. b. ( pts) Find the directional derivative D,, f(,, ). c. ( pts)

### SECOND DERIVATIVE TEST FOR CONSTRAINED EXTREMA

SECOND DERIVATIVE TEST FOR CONSTRAINED EXTREMA This handout presents the second derivative test for a local extrema of a Lagrange multiplier problem. The Section 1 presents a geometric motivation for the

### Recall that two vectors in are perpendicular or orthogonal provided that their dot

Orthogonal Complements and Projections Recall that two vectors in are perpendicular or orthogonal provided that their dot product vanishes That is, if and only if Example 1 The vectors in are orthogonal

### ELEC-E8104 Stochastics models and estimation, Lecture 3b: Linear Estimation in Static Systems

Stochastics models and estimation, Lecture 3b: Linear Estimation in Static Systems Minimum Mean Square Error (MMSE) MMSE estimation of Gaussian random vectors Linear MMSE estimator for arbitrarily distributed

### Vectors, Gradient, Divergence and Curl.

Vectors, Gradient, Divergence and Curl. 1 Introduction A vector is determined by its length and direction. They are usually denoted with letters with arrows on the top a or in bold letter a. We will use

### ME128 Computer-Aided Mechanical Design Course Notes Introduction to Design Optimization

ME128 Computer-ided Mechanical Design Course Notes Introduction to Design Optimization 2. OPTIMIZTION Design optimization is rooted as a basic problem for design engineers. It is, of course, a rare situation

### A QUICK GUIDE TO THE FORMULAS OF MULTIVARIABLE CALCULUS

A QUIK GUIDE TO THE FOMULAS OF MULTIVAIABLE ALULUS ontents 1. Analytic Geometry 2 1.1. Definition of a Vector 2 1.2. Scalar Product 2 1.3. Properties of the Scalar Product 2 1.4. Length and Unit Vectors

### Lecture 14: Section 3.3

Lecture 14: Section 3.3 Shuanglin Shao October 23, 2013 Definition. Two nonzero vectors u and v in R n are said to be orthogonal (or perpendicular) if u v = 0. We will also agree that the zero vector in

### Taylor Polynomials and Taylor Series Math 126

Taylor Polynomials and Taylor Series Math 26 In many problems in science and engineering we have a function f(x) which is too complicated to answer the questions we d like to ask. In this chapter, we will

### Math 21a Review Session for Exam 2 Solutions to Selected Problems

Math 1a Review Session for Exam Solutions to Selected Problems John Hall April 5, 9 Note: Problems which do not have solutions were done in the review session. 1. Suppose that the temperature distribution

### Increasing for all. Convex for all. ( ) Increasing for all (remember that the log function is only defined for ). ( ) Concave for all.

1. Differentiation The first derivative of a function measures by how much changes in reaction to an infinitesimal shift in its argument. The largest the derivative (in absolute value), the faster is evolving.

### Solution based on matrix technique Rewrite. ) = 8x 2 1 4x 1x 2 + 5x x1 2x 2 2x 1 + 5x 2

8.2 Quadratic Forms Example 1 Consider the function q(x 1, x 2 ) = 8x 2 1 4x 1x 2 + 5x 2 2 Determine whether q(0, 0) is the global minimum. Solution based on matrix technique Rewrite q( x1 x 2 = x1 ) =

### OLS is not only unbiased it is also the most precise (efficient) unbiased estimation technique - ie the estimator has the smallest variance

Lecture 5: Hypothesis Testing What we know now: OLS is not only unbiased it is also the most precise (efficient) unbiased estimation technique - ie the estimator has the smallest variance (if the Gauss-Markov

### The Gradient and Level Sets

The Gradient and Level Sets. Let f(x, y) = x + y. (a) Find the gradient f. Solution. f(x, y) = x, y. (b) Pick your favorite positive number k, and let C be the curve f(x, y) = k. Draw the curve on the

### SOLUTIONS TO HOMEWORK ASSIGNMENT #5, Math 253

SOLUTIONS TO HOMEWORK ASSIGNMENT #5, Math 53. For what values of the constant k does the function f(x, y) =kx 3 + x +y 4x 4y have (a) no critical points; (b) exactly one critical point; (c) exactly two

### Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written

### 1 Eigenvalues and Eigenvectors

Math 20 Chapter 5 Eigenvalues and Eigenvectors Eigenvalues and Eigenvectors. Definition: A scalar λ is called an eigenvalue of the n n matrix A is there is a nontrivial solution x of Ax = λx. Such an x

### SECOND-ORDER LINEAR HOMOGENEOUS DIFFERENTIAL EQUATIONS

L SECOND-ORDER LINEAR HOOGENEOUS DIFFERENTIAL EQUATIONS SECOND-ORDER LINEAR HOOGENEOUS DIFFERENTIAL EQUATIONS WITH CONSTANT COEFFICIENTS A second-order linear differential equation is one of the form d

### Rolle s Theorem. q( x) = 1

Lecture 1 :The Mean Value Theorem We know that constant functions have derivative zero. Is it possible for a more complicated function to have derivative zero? In this section we will answer this question

### Math 53 Worksheet Solutions- Minmax and Lagrange

Math 5 Worksheet Solutions- Minmax and Lagrange. Find the local maximum and minimum values as well as the saddle point(s) of the function f(x, y) = e y (y x ). Solution. First we calculate the partial

### Math 150, Fall 2009 Solutions to Practice Final Exam [1] The equation of the tangent line to the curve. cosh y = x + sin y + cos y

Math 150, Fall 2009 Solutions to Practice Final Exam [1] The equation of the tangent line to the curve at the point (0, 0) is cosh y = x + sin y + cos y Answer : y = x Justification: The equation of the

### LINE INTEGRALS OF VECTOR FUNCTIONS: GREEN S THEOREM. Contents. 2. Green s Theorem 3

LINE INTEGRALS OF VETOR FUNTIONS: GREEN S THEOREM ontents 1. A differential criterion for conservative vector fields 1 2. Green s Theorem 3 1. A differential criterion for conservative vector fields We

### Definition The covariance of X and Y, denoted by cov(x, Y ) is defined by. cov(x, Y ) = E(X µ 1 )(Y µ 2 ).

Correlation Regression Bivariate Normal Suppose that X and Y are r.v. s with joint density f(x y) and suppose that the means of X and Y are respectively µ 1 µ 2 and the variances are 1 2. Definition The

### Section 12.6: Directional Derivatives and the Gradient Vector

Section 26: Directional Derivatives and the Gradient Vector Recall that if f is a differentiable function of x and y and z = f(x, y), then the partial derivatives f x (x, y) and f y (x, y) give the rate

### 1 Inner Products and Norms on Real Vector Spaces

Math 373: Principles Techniques of Applied Mathematics Spring 29 The 2 Inner Product 1 Inner Products Norms on Real Vector Spaces Recall that an inner product on a real vector space V is a function from

### 4.3 Lagrange Approximation

206 CHAP. 4 INTERPOLATION AND POLYNOMIAL APPROXIMATION Lagrange Polynomial Approximation 4.3 Lagrange Approximation Interpolation means to estimate a missing function value by taking a weighted average

### Duality in General Programs. Ryan Tibshirani Convex Optimization 10-725/36-725

Duality in General Programs Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: duality in linear programs Given c R n, A R m n, b R m, G R r n, h R r : min x R n c T x max u R m, v R r b T

### 9 Multiplication of Vectors: The Scalar or Dot Product

Arkansas Tech University MATH 934: Calculus III Dr. Marcel B Finan 9 Multiplication of Vectors: The Scalar or Dot Product Up to this point we have defined what vectors are and discussed basic notation

### Solving simultaneous equations. Jackie Nicholas

Mathematics Learning Centre Solving simultaneous equations Jackie Nicholas c 2005 University of Sydney Mathematics Learning Centre, University of Sydney 1 1 Simultaneous linear equations We will introduce

### DERIVATIVES AS MATRICES; CHAIN RULE

DERIVATIVES AS MATRICES; CHAIN RULE 1. Derivatives of Real-valued Functions Let s first consider functions f : R 2 R. Recall that if the partial derivatives of f exist at the point (x 0, y 0 ), then we

### Tangent and normal lines to conics

4.B. Tangent and normal lines to conics Apollonius work on conics includes a study of tangent and normal lines to these curves. The purpose of this document is to relate his approaches to the modern viewpoints

### OLS in Matrix Form. Let y be an n 1 vector of observations on the dependent variable.

OLS in Matrix Form 1 The True Model Let X be an n k matrix where we have observations on k independent variables for n observations Since our model will usually contain a constant term, one of the columns

### Solutions for Review Problems

olutions for Review Problems 1. Let be the triangle with vertices A (,, ), B (4,, 1) and C (,, 1). (a) Find the cosine of the angle BAC at vertex A. (b) Find the area of the triangle ABC. (c) Find a vector

### 2 Integrating Both Sides

2 Integrating Both Sides So far, the only general method we have for solving differential equations involves equations of the form y = f(x), where f(x) is any function of x. The solution to such an equation

### Continuous random variables

Continuous random variables So far we have been concentrating on discrete random variables, whose distributions are not continuous. Now we deal with the so-called continuous random variables. A random

### ( 1)2 + 2 2 + 2 2 = 9 = 3 We would like to make the length 6. The only vectors in the same direction as v are those

1.(6pts) Which of the following vectors has the same direction as v 1,, but has length 6? (a), 4, 4 (b),, (c) 4,, 4 (d), 4, 4 (e) 0, 6, 0 The length of v is given by ( 1) + + 9 3 We would like to make

### Math 407A: Linear Optimization

Math 407A: Linear Optimization Lecture 4: LP Standard Form 1 1 Author: James Burke, University of Washington LPs in Standard Form Minimization maximization Linear equations to linear inequalities Lower

### Continuous Random Variables

Continuous Random Variables COMP 245 STATISTICS Dr N A Heard Contents 1 Continuous Random Variables 2 11 Introduction 2 12 Probability Density Functions 3 13 Transformations 5 2 Mean, Variance and Quantiles

### Convex Optimization SVM s and Kernel Machines

Convex Optimization SVM s and Kernel Machines S.V.N. Vishy Vishwanathan vishy@axiom.anu.edu.au National ICT of Australia and Australian National University Thanks to Alex Smola and Stéphane Canu S.V.N.

### Polynomials can be added or subtracted simply by adding or subtracting the corresponding terms, e.g., if

1. Polynomials 1.1. Definitions A polynomial in x is an expression obtained by taking powers of x, multiplying them by constants, and adding them. It can be written in the form c 0 x n + c 1 x n 1 + c

### 3.7 Non-autonomous linear systems of ODE. General theory

3.7 Non-autonomous linear systems of ODE. General theory Now I will study the ODE in the form ẋ = A(t)x + g(t), x(t) R k, A, g C(I), (3.1) where now the matrix A is time dependent and continuous on some

### Date: April 12, 2001. Contents

2 Lagrange Multipliers Date: April 12, 2001 Contents 2.1. Introduction to Lagrange Multipliers......... p. 2 2.2. Enhanced Fritz John Optimality Conditions...... p. 12 2.3. Informative Lagrange Multipliers...........

### FIXED POINT ITERATION

FIXED POINT ITERATION We begin with a computational example. solving the two equations Consider E1: x =1+.5sinx E2: x =3+2sinx Graphs of these two equations are shown on accompanying graphs, with the solutions

### max cx s.t. Ax c where the matrix A, cost vector c and right hand side b are given and x is a vector of variables. For this example we have x

Linear Programming Linear programming refers to problems stated as maximization or minimization of a linear function subject to constraints that are linear equalities and inequalities. Although the study

### PHP 2510 Central limit theorem, confidence intervals. PHP 2510 October 20,

PHP 2510 Central limit theorem, confidence intervals PHP 2510 October 20, 2008 1 Distribution of the sample mean Case 1: Population distribution is normal For an individual in the population, X i N(µ,

### Definition: Suppose that two random variables, either continuous or discrete, X and Y have joint density

HW MATH 461/561 Lecture Notes 15 1 Definition: Suppose that two random variables, either continuous or discrete, X and Y have joint density and marginal densities f(x, y), (x, y) Λ X,Y f X (x), x Λ X,

### Conditional expectation

A Conditional expectation A.1 Review of conditional densities, expectations We start with the continuous case. This is sections 6.6 and 6.8 in the book. Let X, Y be continuous random variables. We defined

### The Method of Partial Fractions Math 121 Calculus II Spring 2015

Rational functions. as The Method of Partial Fractions Math 11 Calculus II Spring 015 Recall that a rational function is a quotient of two polynomials such f(x) g(x) = 3x5 + x 3 + 16x x 60. The method

### Practice Final Math 122 Spring 12 Instructor: Jeff Lang

Practice Final Math Spring Instructor: Jeff Lang. Find the limit of the sequence a n = ln (n 5) ln (3n + 8). A) ln ( ) 3 B) ln C) ln ( ) 3 D) does not exist. Find the limit of the sequence a n = (ln n)6

### Homework One Solutions. Keith Fratus

Homework One Solutions Keith Fratus June 8, 011 1 Problem One 1.1 Part a In this problem, we ll assume the fact that the sum of two complex numbers is another complex number, and also that the product

### Chapter 15 Introduction to Linear Programming

Chapter 15 Introduction to Linear Programming An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Brief History of Linear Programming The goal of linear programming is to determine the values of

### Math 241, Exam 1 Information.

Math 241, Exam 1 Information. 9/24/12, LC 310, 11:15-12:05. Exam 1 will be based on: Sections 12.1-12.5, 14.1-14.3. The corresponding assigned homework problems (see http://www.math.sc.edu/ boylan/sccourses/241fa12/241.html)

### Multiplicity. Chapter 6

Chapter 6 Multiplicity The fundamental theorem of algebra says that any polynomial of degree n 0 has exactly n roots in the complex numbers if we count with multiplicity. The zeros of a polynomial are

### In this chapter we turn to surfaces in general. We discuss the following topics. Describing surfaces with equations and parametric descriptions.

Chapter 4 Surfaces In this chapter we turn to surfaces in general. We discuss the following topics. Describing surfaces with equations and parametric descriptions. Some constructions of surfaces: surfaces

### Matrix Norms. Tom Lyche. September 28, Centre of Mathematics for Applications, Department of Informatics, University of Oslo

Matrix Norms Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo September 28, 2009 Matrix Norms We consider matrix norms on (C m,n, C). All results holds for

### UNIT 2 MATRICES - I 2.0 INTRODUCTION. Structure

UNIT 2 MATRICES - I Matrices - I Structure 2.0 Introduction 2.1 Objectives 2.2 Matrices 2.3 Operation on Matrices 2.4 Invertible Matrices 2.5 Systems of Linear Equations 2.6 Answers to Check Your Progress

### Solving Systems of Linear Equations

LECTURE 5 Solving Systems of Linear Equations Recall that we introduced the notion of matrices as a way of standardizing the expression of systems of linear equations In today s lecture I shall show how

### MULTIVARIATE PROBABILITY DISTRIBUTIONS

MULTIVARIATE PROBABILITY DISTRIBUTIONS. PRELIMINARIES.. Example. Consider an experiment that consists of tossing a die and a coin at the same time. We can consider a number of random variables defined

### 4.5 Chebyshev Polynomials

230 CHAP. 4 INTERPOLATION AND POLYNOMIAL APPROXIMATION 4.5 Chebyshev Polynomials We now turn our attention to polynomial interpolation for f (x) over [ 1, 1] based on the nodes 1 x 0 < x 1 < < x N 1. Both

### Fourier series. Jan Philip Solovej. English summary of notes for Analysis 1. May 8, 2012

Fourier series Jan Philip Solovej English summary of notes for Analysis 1 May 8, 2012 1 JPS, Fourier series 2 Contents 1 Introduction 2 2 Fourier series 3 2.1 Periodic functions, trigonometric polynomials

### By W.E. Diewert. July, Linear programming problems are important for a number of reasons:

APPLIED ECONOMICS By W.E. Diewert. July, 3. Chapter : Linear Programming. Introduction The theory of linear programming provides a good introduction to the study of constrained maximization (and minimization)

### R is a function which is of class C 1. We have already thought about its level sets, sets of the form

Manifolds 1 Chapter 5 Manifolds We are now going to begin our study of calculus on curved spaces. Everything we have done up to this point has been concerned with what one might call the flat Euclidean

### is in plane V. However, it may be more convenient to introduce a plane coordinate system in V.

.4 COORDINATES EXAMPLE Let V be the plane in R with equation x +2x 2 +x 0, a two-dimensional subspace of R. We can describe a vector in this plane by its spatial (D)coordinates; for example, vector x 5

### 16.5: CURL AND DIVERGENCE

16.5: URL AN IVERGENE KIAM HEONG KWA 1. url Let F = P i + Qj + Rk be a vector field on a solid region R 3. If all first-order partial derivatives of P, Q, and R exist, then the curl of F on is the vector

### Section 5.4 (Systems of Linear Differential Equation); Eigenvalues and Eigenvectors

Section 5.4 (Systems of Linear Differential Equation); Eigenvalues and Eigenvectors July 1, 2009 Today s Session Today s Session A Summary of This Session: Today s Session A Summary of This Session: (1)

### Problems for Quiz 14

Problems for Quiz 14 Math 3. Spring, 7. 1. Consider the initial value problem (IVP defined by the partial differential equation (PDE u t = u xx u x + u, < x < 1, t > (1 with boundary conditions and initial

### Fourier Series. Chapter Some Properties of Functions Goal Preliminary Remarks

Chapter 3 Fourier Series 3.1 Some Properties of Functions 3.1.1 Goal We review some results about functions which play an important role in the development of the theory of Fourier series. These results

### 88 CHAPTER 2. VECTOR FUNCTIONS. . First, we need to compute T (s). a By definition, r (s) T (s) = 1 a sin s a. sin s a, cos s a

88 CHAPTER. VECTOR FUNCTIONS.4 Curvature.4.1 Definitions and Examples The notion of curvature measures how sharply a curve bends. We would expect the curvature to be 0 for a straight line, to be very small

### 1 Error in Euler s Method

1 Error in Euler s Method Experience with Euler s 1 method raises some interesting questions about numerical approximations for the solutions of differential equations. 1. What determines the amount of

### Linear Systems. Singular and Nonsingular Matrices. Find x 1, x 2, x 3 such that the following three equations hold:

Linear Systems Example: Find x, x, x such that the following three equations hold: x + x + x = 4x + x + x = x + x + x = 6 We can write this using matrix-vector notation as 4 {{ A x x x {{ x = 6 {{ b General

### MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column

### To give it a definition, an implicit function of x and y is simply any relationship that takes the form:

2 Implicit function theorems and applications 21 Implicit functions The implicit function theorem is one of the most useful single tools you ll meet this year After a while, it will be second nature to

### Inner product. Definition of inner product

Math 20F Linear Algebra Lecture 25 1 Inner product Review: Definition of inner product. Slide 1 Norm and distance. Orthogonal vectors. Orthogonal complement. Orthogonal basis. Definition of inner product

### THE DIFFUSION EQUATION

THE DIFFUSION EQUATION R. E. SHOWALTER 1. Heat Conduction in an Interval We shall describe the diffusion of heat energy through a long thin rod G with uniform cross section S. As before, we identify G

### Solutions to Math 51 First Exam January 29, 2015

Solutions to Math 5 First Exam January 29, 25. ( points) (a) Complete the following sentence: A set of vectors {v,..., v k } is defined to be linearly dependent if (2 points) there exist c,... c k R, not

### 5 Numerical Differentiation

D. Levy 5 Numerical Differentiation 5. Basic Concepts This chapter deals with numerical approximations of derivatives. The first questions that comes up to mind is: why do we need to approximate derivatives

### Basic Terminology for Systems of Equations in a Nutshell. E. L. Lady. 3x 1 7x 2 +4x 3 =0 5x 1 +8x 2 12x 3 =0.

Basic Terminology for Systems of Equations in a Nutshell E L Lady A system of linear equations is something like the following: x 7x +4x =0 5x +8x x = Note that the number of equations is not required

### 3.5 Spline interpolation

3.5 Spline interpolation Given a tabulated function f k = f(x k ), k = 0,... N, a spline is a polynomial between each pair of tabulated points, but one whose coefficients are determined slightly non-locally.

Facts About Eigenvalues By Dr David Butler Definitions Suppose A is an n n matrix An eigenvalue of A is a number λ such that Av = λv for some nonzero vector v An eigenvector of A is a nonzero vector v

### Equations Involving Lines and Planes Standard equations for lines in space

Equations Involving Lines and Planes In this section we will collect various important formulas regarding equations of lines and planes in three dimensional space Reminder regarding notation: any quantity

### Summary of week 8 (Lectures 22, 23 and 24)

WEEK 8 Summary of week 8 (Lectures 22, 23 and 24) This week we completed our discussion of Chapter 5 of [VST] Recall that if V and W are inner product spaces then a linear map T : V W is called an isometry

### The Delta Method and Applications

Chapter 5 The Delta Method and Applications 5.1 Linear approximations of functions In the simplest form of the central limit theorem, Theorem 4.18, we consider a sequence X 1, X,... of independent and

### BX in ( u, v) basis in two ways. On the one hand, AN = u+

1. Let f(x) = 1 x +1. Find f (6) () (the value of the sixth derivative of the function f(x) at zero). Answer: 7. We expand the given function into a Taylor series at the point x = : f(x) = 1 x + x 4 x

### CITY UNIVERSITY LONDON. BEng Degree in Computer Systems Engineering Part II BSc Degree in Computer Systems Engineering Part III PART 2 EXAMINATION

No: CITY UNIVERSITY LONDON BEng Degree in Computer Systems Engineering Part II BSc Degree in Computer Systems Engineering Part III PART 2 EXAMINATION ENGINEERING MATHEMATICS 2 (resit) EX2005 Date: August

### Definition 12 An alternating bilinear form on a vector space V is a map B : V V F such that

4 Exterior algebra 4.1 Lines and 2-vectors The time has come now to develop some new linear algebra in order to handle the space of lines in a projective space P (V ). In the projective plane we have seen

### PROVING STATEMENTS IN LINEAR ALGEBRA

Mathematics V2010y Linear Algebra Spring 2007 PROVING STATEMENTS IN LINEAR ALGEBRA Linear algebra is different from calculus: you cannot understand it properly without some simple proofs. Knowing statements

### Recall that the gradient of a differentiable scalar field ϕ on an open set D in R n is given by the formula:

Chapter 7 Div, grad, and curl 7.1 The operator and the gradient: Recall that the gradient of a differentiable scalar field ϕ on an open set D in R n is given by the formula: ( ϕ ϕ =, ϕ,..., ϕ. (7.1 x 1

### This makes sense. t 2 1 + 1/t 2 dt = 1. t t 2 + 1dt = 2 du = 1 3 u3/2 u=5

1. (Line integrals Using parametrization. Two types and the flux integral) Formulas: ds = x (t) dt, d x = x (t)dt and d x = T ds since T = x (t)/ x (t). Another one is Nds = T ds ẑ = (dx, dy) ẑ = (dy,

### 24. The Branch and Bound Method

24. The Branch and Bound Method It has serious practical consequences if it is known that a combinatorial problem is NP-complete. Then one can conclude according to the present state of science that no

### Introduction to Algebraic Geometry. Bézout s Theorem and Inflection Points

Introduction to Algebraic Geometry Bézout s Theorem and Inflection Points 1. The resultant. Let K be a field. Then the polynomial ring K[x] is a unique factorisation domain (UFD). Another example of a

### Teaching Mathematics to Engineering Students: To Use or Not To Use TI-Nspire CAS

TIME-2008 Symposium Technology and its Integration in Mathematics Education 10 th ACDCA Summer Academy 8 th Int l Derive & TI-NspireConference Hosted by the Tshwane University of Technology 22-26 September