Convex analysis and profit/cost/support functions
|
|
|
- Matilda Sanders
- 10 years ago
- Views:
Transcription
1 CALIFORNIA INSTITUTE OF TECHNOLOGY Division of the Humanities and Social Sciences Convex analysis and profit/cost/support functions KC Border October 2004 Revised January 2009 Let A be a subset of R m Convex analysts may give one of two definitions for the support function of A as either an infimum or a supremum Recall that the supremum of a set of real numbers is its least upper bound and the infimum is its greatest lower bound If A has no upper bound, then by convention sup A = and if A has no lower bound, then inf A = For the empty set, sup A = and inf A = ; otherwise inf A sup A This makes a kind of sense: Every real number is an upper bound for the empty set, since there is no member of the empty set that is greater than Thus the least upper bound must be Similarly, every real number is also a lower bound, so the infimum is Thus support functions as infima or suprema may assume the values and By convention, 0 = 0; if > 0 is a real number, then = and = ; and if < 0 is a real number, then = and = These conventions are used to simplify statements involving positive homogeneity Rather than choose one definition, I shall give the two definitions different names based on their economic interpretation Profit maximization The profit function π A of A is defined by π A p = sup p y y A Cost minimization The cost function c A of A is defined by c A p = inf y A p y Clearly, π A p = c A p Clearly c A p = π A p π A is convex, lower semicontinuous, and positively homogeneous of degree 1 Positive homogeneity of π A is obvious given the conventions on multiplication of infinities To see that it is convex, let g x be the linear hence convex function defined by g x p = x p Then π A p = sup x A g x p Since the pointwise supremum of a family of convex functions is convex, π A is convex Also each g x is continuous, hence lower semicontinuous, and the supremum of a family of lower semicontinuous functions is lower semicontinuous See my notes on maximization c A is concave, upper semicontinuous, and positively homogeneous of degree 1 Positive homogeneity of c A is obvious given the conventions on multiplication of infinities To see that it is concave, let g x be the linear hence concave function defined by g x p = x p Then c A p = inf x A g x p Since the pointwise infimum of a family of concave functions is concave, c A is concave Also each g x is continuous, hence upper semicontinuous, and the infimum of a family of upper semicontinuous functions is upper semicontinuous See my notes on maximization 1
2 KC Border Convex analysis and profit/cost/support functions 2 The set The set {p R m : π A p < } is a closed convex cone, called the effective domain of π A, and denoted dom π A The effective domain will always include the point 0 provided A is nonempty By convention π p = for all p, and we say that π A is improper If A = R m, then 0 is the only point in the effective domain of π A It is easy to see that the effective domain dom π A of π A is a cone, that is, if p dom π A, then p dom π A for every 0 Note that {0} is a degenerate cone It is also straightforward to show that dom π A is convex For if π A p < and π A q <, for 0 1, by convexity of π A, we π A x + 1 y πa p + 1 π A q < The closedness of dom π A is more difficult {p R m : c A p > } is a closed convex cone, called the effective domain of c A, and denoted dom c A The effective domain will always include the point 0 provided A is nonempty By convention c p = for all p, and we say that c is improper If A = R m, then 0 is the only point in the effective domain of c A It is easy to see that the effective domain dom c A of c A is a cone, that is, if p dom c A, then p dom c A for every 0 Note that {0} is a degenerate cone It is also straightforward to show that dom c A is convex For if c A p > and c A q >, for 0 1, by concavity of c A, we c A x + 1 y ca p + 1 c A q > The closedness of dom c A is more difficult Recoverability Separating Hyperplane Theorem If A is a closed convex set, and x does not belong to A, then there is a nonzero p satisfying p x > π A p For a proof see my notes Note that given our conventions, A may be empty From this theorem we easily get the next proposition The closed convex hull co A of co A = { y R m : p R m [ p y π A p ]} Separating Hyperplane Theorem If A is a closed convex set, and x does not belong to A, then there is a nonzero p satisfying p x < c A p For a proof see my notes Note that given our conventions, A may be empty From this theorem we easily get the next proposition The closed convex hull co A of co A = { y R m : p R m [ p y c A p ]} Now let f be a continuous real-valued function defined on a closed convex cone D We can extend f to all of R m by setting it to outside of D if f is convex or if f is concave If f is convex and positively homogeneous of degree 1, define A = { y R m : p R m [ p y fp ]} Then A is closed and convex and f = π A If f is concave and positively homogeneous of degree 1, define A = { y R m : p R m [ p y fp ]} Then A is closed and convex and f = c A
3 KC Border Convex analysis and profit/cost/support functions 3 Extremizers are subgradients If ỹp maximizes p over A, that is, if ỹp belongs to A and p ỹp p y for all y A, then ỹp is a subgradient of π A at p That is, If ŷp minimizes p over A, that is, if ŷp belongs to A and p ŷp p y for all y A, then ŷp is a supergradient of c A at p That is, π A p + ỹp q p π A q c A p + ŷp q p c A q for all q R m for all q R m To see this, note that for any q R m, by definition we q ỹp π A q Now add π A p p ỹp = 0 to the left hand side to get the subgradient inequality Note that π A p may be finite for a closed convex set A, and yet there may be no maximizer For instance, let A = {x, y R 2 : x < 0, y < 0, xy 1} Then for p = 1, 0, we π A p = 0 as 1, 0 1/n, n = 1/n, but 1, 0 x, y = x < 0 for each x, y A Thus there is no maximizer in A It turns out that if there is no maximizer of p, then π A has no subgradient at p In fact, the following is true, but I won t present the proof, which relies on the Separating Hyperplane Theorem See my notes for a proof Theorem If A is closed and convex, then x is a subgradient of π A at p if and only if x A and x maximizes p over A To see this, note that for any q R m, by definition we q ŷp c A q Now add c A p p ŷp = 0 to the left hand side to get the supergradient inequality Note that x A p may be finite for a closed convex set A, and yet there may be no minimizer For instance, let A = {x, y R 2 : x > 0, y > 0, xy 1} Then for p = 1, 0, we π A p = 0 as 1, 0 1/n, n = 1/n, but 1, 0 x, y = x > 0 for each x, y A Thus there is no minimizer in A It turns out that if there is no minimizer of p, then c A has no supergradient at p In fact, the following is true, but I won t present the proof, which relies on the Separating Hyperplane Theorem See my notes for a proof Theorem If A is closed and convex, then x is a supergradient of c A at p if and only if x A and x minimizes p over A Comparative statics Consequently, if A is closed and convex, and ỹp is the unique maximizer of p over A, then π A is differentiable at p and Consequently, if A is closed and convex, and ŷp is the unique minimizer of p over A, then c A is differentiable at p and ỹp = π Ap ŷp = c Ap
4 KC Border Convex analysis and profit/cost/support functions 4 To see that differentiability of π A implies the profit maximizer is unique, consider q of the form p ± e i, where e i is the i th unit coordinate vector, and > 0 The subgradient inequality for q = p + e i is ỹp e i π A p + e i π A p and for q = p e i is ỹp e i π A p e i π A p Dividing these by and respectively yields To see that differentiability of c A implies the cost minimizer is unique, consider q of the form p ± e i, where e i is the i th unit coordinate vector, and > 0 The supergradient inequality for q = p + e i is ŷp e i c A p + e i c A p and for q = p e i is ŷp e i c A p e i c A p Dividing these by and respectively yields so yi p π Ap + e i π A p yi p π Ap e i π A p so yi p c Ap + e i c A p yi p c Ap e i c A p π A p e i π A p y i p π Ap+e i π A p Letting 0 yields ỹ i p = D i π A p Thus if π A is twice differentiable at p, that is, if the maximizer ỹp is differentiable with respect to p, then the i th component D j ỹ i p = D ij π A p Consequently, the matrix [ Dj ỹ i p ] c A p+e i c A p y i p c Ap e i c A p Letting 0 yields ŷ i p = D i c A p Thus if c A is twice differentiable at p, that is, if the minimizer ŷp is differentiable with respect to p, then the i th component D j ŷ i p = D ij c A p Consequently, the matrix [ Dj ŷ i p ] is positive semidefinite is negative semidefinite In particular, D i ỹ i 0 In particular, D i ŷ i 0 Even without twice differentiability, from the subgradient inequality, we π A p + ỹp q p π A q π A q + ỹq p q π A p so adding the two inequalities, we get ỹp ỹq p q 0 Even without twice differentiability, from the supergradient inequality, we c A p + ŷp q p c A q c A q + ŷq p q c A p so adding the two inequalities, we get ŷp ŷq p q 0
5 KC Border Convex analysis and profit/cost/support functions 5 Thus if q differs from p only in its i th component, say q i = p i + p i, then we ỹ i p i 0 Dividing by the positive quantity p i 2 does not change this inequality, so ỹ i p i 0 Thus if q differs from p only in its i th component, say q i = p i + p i, then we ŷ i p i 0 Dividing by the positive quantity p i 2 does not change this inequality, so ŷ i p i 0 Cyclical monotonicity and empirical restrictions A real function g : X R R is increasing if x y = gx gy That is, gx gy and x y the same sign An equivalent way to say this is gx gy x y 0 for all x, y, which can be rewritten as gxy x + gyx y 0 for all x, y We can generalize this to a function g from R m into R m like this: Definition A function g : X R m R m is monotone increasing if gx y x + gy x y 0 for all x, y X We already seen that the subgradient of π A is monotone increasing More is true Definition A mapping g : X R m R m is cyclically monotone increasing if for every cycle x 0, x 1,, x n, x n+1 = x 0 in X, we gx 0 x 1 x gx n x n+1 x n 0 If f : X R m R is convex, and g : X R m R m is a selection from the subdifferential of f, that is, if gx is a subgradient of f at x for every x, then g is cyclically monotone increasing A real function g : X R R is decreasing if x y = gx gy That is, gx gy and x y the opposite sign An equivalent way to say this is gx gy x y 0 for all x, y, which can be rewritten as gxy x + gyx y 0 for all x, y We can generalize this to a function g from R m into R m like this: Definition A function g : X R m R m is monotone decreasing if gx y x + gy x y 0 for all x, y X We already seen that the supergradient of c A is monotone decreasing More is true Definition A mapping g : X R m R m is cyclically monotone decreasing if for every cycle x 0, x 1,, x n, x n+1 = x 0 in X, we gx 0 x 1 x gx n x n+1 x n 0 If f : X R m R is concave, and g : X R m R m is a selection from the superdifferential of f, that is, if gx is a supergradient of f at x for every x, then g is cyclically monotone decreasing
6 KC Border Convex analysis and profit/cost/support functions 6 Proof : Let x 0, x 1,, x n, x n+1 = x 0 be a cycle in X From the subgradient inequality at x i, we or fx i + gx i x i+1 x i fx i+1 gx i x i+1 x i fx i+1 fx i for each i = 0,, n Summing gives n gx i x i+1 x i 0, i=0 where the right-hand side takes into account fx n+1 = fx 0 Corollary The profit maximizing points correspondence ŷ is cyclically monotonic increasing Remarkably the converse is true Theorem Rockafellar Let X be a convex set in R m and let φ: X R m be a cyclically monotone increasing correspondence that is, if every selection from φ is a cyclically monotone increasing function Then there is a lower semicontinuous convex function f : X R such that φx fx for every x Proof : Let x 0, x 1,, x n, x n+1 = x 0 be a cycle in X From the supergradient inequality at x i, we or fx i + gx i x i+1 x i fx i+1 gx i x i+1 x i fx i+1 fx i for each i = 0,, n Summing gives n gx i x i+1 x i 0, i=0 where the right-hand side takes into account fx n+1 = fx 0 Corollary The cost minimizing points correspondence ŷ is cyclically monotonic decreasing Remarkably the converse is true Theorem Rockafellar Let X be a convex set in R m and let φ: X R m be a cyclically monotone decreasing correspondence that is, if every selection from φ is a cyclically monotone decreasing function Then there is an upper semicontinuous concave function f : X R such that φx fx for every x
Notes V General Equilibrium: Positive Theory. 1 Walrasian Equilibrium and Excess Demand
Notes V General Equilibrium: Positive Theory In this lecture we go on considering a general equilibrium model of a private ownership economy. In contrast to the Notes IV, we focus on positive issues such
No: 10 04. Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics
No: 10 04 Bilkent University Monotonic Extension Farhad Husseinov Discussion Papers Department of Economics The Discussion Papers of the Department of Economics are intended to make the initial results
2.3 Convex Constrained Optimization Problems
42 CHAPTER 2. FUNDAMENTAL CONCEPTS IN CONVEX OPTIMIZATION Theorem 15 Let f : R n R and h : R R. Consider g(x) = h(f(x)) for all x R n. The function g is convex if either of the following two conditions
1 if 1 x 0 1 if 0 x 1
Chapter 3 Continuity In this chapter we begin by defining the fundamental notion of continuity for real valued functions of a single real variable. When trying to decide whether a given function is or
Duality of linear conic problems
Duality of linear conic problems Alexander Shapiro and Arkadi Nemirovski Abstract It is well known that the optimal values of a linear programming problem and its dual are equal to each other if at least
Separation Properties for Locally Convex Cones
Journal of Convex Analysis Volume 9 (2002), No. 1, 301 307 Separation Properties for Locally Convex Cones Walter Roth Department of Mathematics, Universiti Brunei Darussalam, Gadong BE1410, Brunei Darussalam
FIRST YEAR CALCULUS. Chapter 7 CONTINUITY. It is a parabola, and we can draw this parabola without lifting our pencil from the paper.
FIRST YEAR CALCULUS WWLCHENW L c WWWL W L Chen, 1982, 2008. 2006. This chapter originates from material used by the author at Imperial College, University of London, between 1981 and 1990. It It is is
A NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION
1 A NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION Dimitri Bertsekas M.I.T. FEBRUARY 2003 2 OUTLINE Convexity issues in optimization Historical remarks Our treatment of the subject Three unifying lines of
Undergraduate Notes in Mathematics. Arkansas Tech University Department of Mathematics
Undergraduate Notes in Mathematics Arkansas Tech University Department of Mathematics An Introductory Single Variable Real Analysis: A Learning Approach through Problem Solving Marcel B. Finan c All Rights
Mathematical Methods of Engineering Analysis
Mathematical Methods of Engineering Analysis Erhan Çinlar Robert J. Vanderbei February 2, 2000 Contents Sets and Functions 1 1 Sets................................... 1 Subsets.............................
Mathematics Course 111: Algebra I Part IV: Vector Spaces
Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 1996-7 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are
Research Article Stability Analysis for Higher-Order Adjacent Derivative in Parametrized Vector Optimization
Hindawi Publishing Corporation Journal of Inequalities and Applications Volume 2010, Article ID 510838, 15 pages doi:10.1155/2010/510838 Research Article Stability Analysis for Higher-Order Adjacent Derivative
Mathematics for Econometrics, Fourth Edition
Mathematics for Econometrics, Fourth Edition Phoebus J. Dhrymes 1 July 2012 1 c Phoebus J. Dhrymes, 2012. Preliminary material; not to be cited or disseminated without the author s permission. 2 Contents
These axioms must hold for all vectors ū, v, and w in V and all scalars c and d.
DEFINITION: A vector space is a nonempty set V of objects, called vectors, on which are defined two operations, called addition and multiplication by scalars (real numbers), subject to the following axioms
MINIMIZATION OF ENTROPY FUNCTIONALS UNDER MOMENT CONSTRAINTS. denote the family of probability density functions g on X satisfying
MINIMIZATION OF ENTROPY FUNCTIONALS UNDER MOMENT CONSTRAINTS I. Csiszár (Budapest) Given a σ-finite measure space (X, X, µ) and a d-tuple ϕ = (ϕ 1,..., ϕ d ) of measurable functions on X, for a = (a 1,...,
MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.
MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix. Nullspace Let A = (a ij ) be an m n matrix. Definition. The nullspace of the matrix A, denoted N(A), is the set of all n-dimensional column
A FIRST COURSE IN OPTIMIZATION THEORY
A FIRST COURSE IN OPTIMIZATION THEORY RANGARAJAN K. SUNDARAM New York University CAMBRIDGE UNIVERSITY PRESS Contents Preface Acknowledgements page xiii xvii 1 Mathematical Preliminaries 1 1.1 Notation
1 Norms and Vector Spaces
008.10.07.01 1 Norms and Vector Spaces Suppose we have a complex vector space V. A norm is a function f : V R which satisfies (i) f(x) 0 for all x V (ii) f(x + y) f(x) + f(y) for all x,y V (iii) f(λx)
Cost Minimization and the Cost Function
Cost Minimization and the Cost Function Juan Manuel Puerta October 5, 2009 So far we focused on profit maximization, we could look at a different problem, that is the cost minimization problem. This is
Linear Algebra I. Ronald van Luijk, 2012
Linear Algebra I Ronald van Luijk, 2012 With many parts from Linear Algebra I by Michael Stoll, 2007 Contents 1. Vector spaces 3 1.1. Examples 3 1.2. Fields 4 1.3. The field of complex numbers. 6 1.4.
Convex Programming Tools for Disjunctive Programs
Convex Programming Tools for Disjunctive Programs João Soares, Departamento de Matemática, Universidade de Coimbra, Portugal Abstract A Disjunctive Program (DP) is a mathematical program whose feasible
Walrasian Demand. u(x) where B(p, w) = {x R n + : p x w}.
Walrasian Demand Econ 2100 Fall 2015 Lecture 5, September 16 Outline 1 Walrasian Demand 2 Properties of Walrasian Demand 3 An Optimization Recipe 4 First and Second Order Conditions Definition Walrasian
Metric Spaces. Chapter 7. 7.1. Metrics
Chapter 7 Metric Spaces A metric space is a set X that has a notion of the distance d(x, y) between every pair of points x, y X. The purpose of this chapter is to introduce metric spaces and give some
10. Proximal point method
L. Vandenberghe EE236C Spring 2013-14) 10. Proximal point method proximal point method augmented Lagrangian method Moreau-Yosida smoothing 10-1 Proximal point method a conceptual algorithm for minimizing
What is Linear Programming?
Chapter 1 What is Linear Programming? An optimization problem usually has three essential ingredients: a variable vector x consisting of a set of unknowns to be determined, an objective function of x to
x a x 2 (1 + x 2 ) n.
Limits and continuity Suppose that we have a function f : R R. Let a R. We say that f(x) tends to the limit l as x tends to a; lim f(x) = l ; x a if, given any real number ɛ > 0, there exists a real number
Name. Final Exam, Economics 210A, December 2011 Here are some remarks to help you with answering the questions.
Name Final Exam, Economics 210A, December 2011 Here are some remarks to help you with answering the questions. Question 1. A firm has a production function F (x 1, x 2 ) = ( x 1 + x 2 ) 2. It is a price
TOPIC 4: DERIVATIVES
TOPIC 4: DERIVATIVES 1. The derivative of a function. Differentiation rules 1.1. The slope of a curve. The slope of a curve at a point P is a measure of the steepness of the curve. If Q is a point on the
Date: April 12, 2001. Contents
2 Lagrange Multipliers Date: April 12, 2001 Contents 2.1. Introduction to Lagrange Multipliers......... p. 2 2.2. Enhanced Fritz John Optimality Conditions...... p. 12 2.3. Informative Lagrange Multipliers...........
ALMOST COMMON PRIORS 1. INTRODUCTION
ALMOST COMMON PRIORS ZIV HELLMAN ABSTRACT. What happens when priors are not common? We introduce a measure for how far a type space is from having a common prior, which we term prior distance. If a type
3. INNER PRODUCT SPACES
. INNER PRODUCT SPACES.. Definition So far we have studied abstract vector spaces. These are a generalisation of the geometric spaces R and R. But these have more structure than just that of a vector space.
MATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets.
MATH 304 Linear Algebra Lecture 20: Inner product spaces. Orthogonal sets. Norm The notion of norm generalizes the notion of length of a vector in R n. Definition. Let V be a vector space. A function α
3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions
3. Convex functions Convex Optimization Boyd & Vandenberghe basic properties and examples operations that preserve convexity the conjugate function quasiconvex functions log-concave and log-convex functions
A domain of spacetime intervals in general relativity
A domain of spacetime intervals in general relativity Keye Martin Department of Mathematics Tulane University New Orleans, LA 70118 United States of America [email protected] Prakash Panangaden School
MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.
MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set. Vector space A vector space is a set V equipped with two operations, addition V V (x,y) x + y V and scalar
I. GROUPS: BASIC DEFINITIONS AND EXAMPLES
I GROUPS: BASIC DEFINITIONS AND EXAMPLES Definition 1: An operation on a set G is a function : G G G Definition 2: A group is a set G which is equipped with an operation and a special element e G, called
24. The Branch and Bound Method
24. The Branch and Bound Method It has serious practical consequences if it is known that a combinatorial problem is NP-complete. Then one can conclude according to the present state of science that no
Lecture 5 Principal Minors and the Hessian
Lecture 5 Principal Minors and the Hessian Eivind Eriksen BI Norwegian School of Management Department of Economics October 01, 2010 Eivind Eriksen (BI Dept of Economics) Lecture 5 Principal Minors and
Notes on metric spaces
Notes on metric spaces 1 Introduction The purpose of these notes is to quickly review some of the basic concepts from Real Analysis, Metric Spaces and some related results that will be used in this course.
God created the integers and the rest is the work of man. (Leopold Kronecker, in an after-dinner speech at a conference, Berlin, 1886)
Chapter 2 Numbers God created the integers and the rest is the work of man. (Leopold Kronecker, in an after-dinner speech at a conference, Berlin, 1886) God created the integers and the rest is the work
GI01/M055 Supervised Learning Proximal Methods
GI01/M055 Supervised Learning Proximal Methods Massimiliano Pontil (based on notes by Luca Baldassarre) (UCL) Proximal Methods 1 / 20 Today s Plan Problem setting Convex analysis concepts Proximal operators
Fuzzy Probability Distributions in Bayesian Analysis
Fuzzy Probability Distributions in Bayesian Analysis Reinhard Viertl and Owat Sunanta Department of Statistics and Probability Theory Vienna University of Technology, Vienna, Austria Corresponding author:
Further Study on Strong Lagrangian Duality Property for Invex Programs via Penalty Functions 1
Further Study on Strong Lagrangian Duality Property for Invex Programs via Penalty Functions 1 J. Zhang Institute of Applied Mathematics, Chongqing University of Posts and Telecommunications, Chongqing
4.6 Linear Programming duality
4.6 Linear Programming duality To any minimization (maximization) LP we can associate a closely related maximization (minimization) LP. Different spaces and objective functions but in general same optimal
Adaptive Online Gradient Descent
Adaptive Online Gradient Descent Peter L Bartlett Division of Computer Science Department of Statistics UC Berkeley Berkeley, CA 94709 bartlett@csberkeleyedu Elad Hazan IBM Almaden Research Center 650
Baltic Way 1995. Västerås (Sweden), November 12, 1995. Problems and solutions
Baltic Way 995 Västerås (Sweden), November, 995 Problems and solutions. Find all triples (x, y, z) of positive integers satisfying the system of equations { x = (y + z) x 6 = y 6 + z 6 + 3(y + z ). Solution.
Inner Product Spaces
Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and
(Basic definitions and properties; Separation theorems; Characterizations) 1.1 Definition, examples, inner description, algebraic properties
Lecture 1 Convex Sets (Basic definitions and properties; Separation theorems; Characterizations) 1.1 Definition, examples, inner description, algebraic properties 1.1.1 A convex set In the school geometry
Introduction to Finite Fields (cont.)
Chapter 6 Introduction to Finite Fields (cont.) 6.1 Recall Theorem. Z m is a field m is a prime number. Theorem (Subfield Isomorphic to Z p ). Every finite field has the order of a power of a prime number
Cross product and determinants (Sect. 12.4) Two main ways to introduce the cross product
Cross product and determinants (Sect. 12.4) Two main ways to introduce the cross product Geometrical definition Properties Expression in components. Definition in components Properties Geometrical expression.
Properties of BMO functions whose reciprocals are also BMO
Properties of BMO functions whose reciprocals are also BMO R. L. Johnson and C. J. Neugebauer The main result says that a non-negative BMO-function w, whose reciprocal is also in BMO, belongs to p> A p,and
How To Prove The Dirichlet Unit Theorem
Chapter 6 The Dirichlet Unit Theorem As usual, we will be working in the ring B of algebraic integers of a number field L. Two factorizations of an element of B are regarded as essentially the same if
On a comparison result for Markov processes
On a comparison result for Markov processes Ludger Rüschendorf University of Freiburg Abstract A comparison theorem is stated for Markov processes in polish state spaces. We consider a general class of
Chapter 13: Basic ring theory
Chapter 3: Basic ring theory Matthew Macauley Department of Mathematical Sciences Clemson University http://www.math.clemson.edu/~macaule/ Math 42, Spring 24 M. Macauley (Clemson) Chapter 3: Basic ring
So let us begin our quest to find the holy grail of real analysis.
1 Section 5.2 The Complete Ordered Field: Purpose of Section We present an axiomatic description of the real numbers as a complete ordered field. The axioms which describe the arithmetic of the real numbers
Chapter 5. Banach Spaces
9 Chapter 5 Banach Spaces Many linear equations may be formulated in terms of a suitable linear operator acting on a Banach space. In this chapter, we study Banach spaces and linear operators acting on
E3: PROBABILITY AND STATISTICS lecture notes
E3: PROBABILITY AND STATISTICS lecture notes 2 Contents 1 PROBABILITY THEORY 7 1.1 Experiments and random events............................ 7 1.2 Certain event. Impossible event............................
4.5 Linear Dependence and Linear Independence
4.5 Linear Dependence and Linear Independence 267 32. {v 1, v 2 }, where v 1, v 2 are collinear vectors in R 3. 33. Prove that if S and S are subsets of a vector space V such that S is a subset of S, then
Several Views of Support Vector Machines
Several Views of Support Vector Machines Ryan M. Rifkin Honda Research Institute USA, Inc. Human Intention Understanding Group 2007 Tikhonov Regularization We are considering algorithms of the form min
CHAPTER II THE LIMIT OF A SEQUENCE OF NUMBERS DEFINITION OF THE NUMBER e.
CHAPTER II THE LIMIT OF A SEQUENCE OF NUMBERS DEFINITION OF THE NUMBER e. This chapter contains the beginnings of the most important, and probably the most subtle, notion in mathematical analysis, i.e.,
1 Introduction. Linear Programming. Questions. A general optimization problem is of the form: choose x to. max f(x) subject to x S. where.
Introduction Linear Programming Neil Laws TT 00 A general optimization problem is of the form: choose x to maximise f(x) subject to x S where x = (x,..., x n ) T, f : R n R is the objective function, S
Numerical Verification of Optimality Conditions in Optimal Control Problems
Numerical Verification of Optimality Conditions in Optimal Control Problems Dissertation zur Erlangung des naturwissenschaftlichen Doktorgrades der Julius-Maximilians-Universität Würzburg vorgelegt von
THE FUNDAMENTAL THEOREM OF ALGEBRA VIA PROPER MAPS
THE FUNDAMENTAL THEOREM OF ALGEBRA VIA PROPER MAPS KEITH CONRAD 1. Introduction The Fundamental Theorem of Algebra says every nonconstant polynomial with complex coefficients can be factored into linear
Lecture 7: Finding Lyapunov Functions 1
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.243j (Fall 2003): DYNAMICS OF NONLINEAR SYSTEMS by A. Megretski Lecture 7: Finding Lyapunov Functions 1
Notes on Symmetric Matrices
CPSC 536N: Randomized Algorithms 2011-12 Term 2 Notes on Symmetric Matrices Prof. Nick Harvey University of British Columbia 1 Symmetric Matrices We review some basic results concerning symmetric matrices.
Max-Min Representation of Piecewise Linear Functions
Beiträge zur Algebra und Geometrie Contributions to Algebra and Geometry Volume 43 (2002), No. 1, 297-302. Max-Min Representation of Piecewise Linear Functions Sergei Ovchinnikov Mathematics Department,
CHAPTER 9. Integer Programming
CHAPTER 9 Integer Programming An integer linear program (ILP) is, by definition, a linear program with the additional constraint that all variables take integer values: (9.1) max c T x s t Ax b and x integral
MATH 425, PRACTICE FINAL EXAM SOLUTIONS.
MATH 45, PRACTICE FINAL EXAM SOLUTIONS. Exercise. a Is the operator L defined on smooth functions of x, y by L u := u xx + cosu linear? b Does the answer change if we replace the operator L by the operator
Online Convex Optimization
E0 370 Statistical Learning heory Lecture 19 Oct 22, 2013 Online Convex Optimization Lecturer: Shivani Agarwal Scribe: Aadirupa 1 Introduction In this lecture we shall look at a fairly general setting
3. Linear Programming and Polyhedral Combinatorics
Massachusetts Institute of Technology Handout 6 18.433: Combinatorial Optimization February 20th, 2009 Michel X. Goemans 3. Linear Programming and Polyhedral Combinatorics Summary of what was seen in the
Fuzzy Differential Systems and the New Concept of Stability
Nonlinear Dynamics and Systems Theory, 1(2) (2001) 111 119 Fuzzy Differential Systems and the New Concept of Stability V. Lakshmikantham 1 and S. Leela 2 1 Department of Mathematical Sciences, Florida
4: SINGLE-PERIOD MARKET MODELS
4: SINGLE-PERIOD MARKET MODELS Ben Goldys and Marek Rutkowski School of Mathematics and Statistics University of Sydney Semester 2, 2015 B. Goldys and M. Rutkowski (USydney) Slides 4: Single-Period Market
Vector and Matrix Norms
Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a non-empty
ORIENTATIONS. Contents
ORIENTATIONS Contents 1. Generators for H n R n, R n p 1 1. Generators for H n R n, R n p We ended last time by constructing explicit generators for H n D n, S n 1 by using an explicit n-simplex which
DERIVATIVES AS MATRICES; CHAIN RULE
DERIVATIVES AS MATRICES; CHAIN RULE 1. Derivatives of Real-valued Functions Let s first consider functions f : R 2 R. Recall that if the partial derivatives of f exist at the point (x 0, y 0 ), then we
G = G 0 > G 1 > > G k = {e}
Proposition 49. 1. A group G is nilpotent if and only if G appears as an element of its upper central series. 2. If G is nilpotent, then the upper central series and the lower central series have the same
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES Contents 1. Random variables and measurable functions 2. Cumulative distribution functions 3. Discrete
Fixed Point Theorems For Set-Valued Maps
Fixed Point Theorems For Set-Valued Maps Bachelor s thesis in functional analysis Institute for Analysis and Scientific Computing Vienna University of Technology Andreas Widder, July 2009. Preface In this
9 More on differentiation
Tel Aviv University, 2013 Measure and category 75 9 More on differentiation 9a Finite Taylor expansion............... 75 9b Continuous and nowhere differentiable..... 78 9c Differentiable and nowhere monotone......
it is easy to see that α = a
21. Polynomial rings Let us now turn out attention to determining the prime elements of a polynomial ring, where the coefficient ring is a field. We already know that such a polynomial ring is a UF. Therefore
Introduction to Algebraic Geometry. Bézout s Theorem and Inflection Points
Introduction to Algebraic Geometry Bézout s Theorem and Inflection Points 1. The resultant. Let K be a field. Then the polynomial ring K[x] is a unique factorisation domain (UFD). Another example of a
minimal polyonomial Example
Minimal Polynomials Definition Let α be an element in GF(p e ). We call the monic polynomial of smallest degree which has coefficients in GF(p) and α as a root, the minimal polyonomial of α. Example: We
MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.
MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column
One side James Compactness Theorem
One side James Compactness Theorem 1 1 Department of Mathematics University of Murcia Topological Methods in Analysis and Optimization. On the occasion of the 70th birthday of Prof. Petar Kenderov A birthday
Understanding Basic Calculus
Understanding Basic Calculus S.K. Chung Dedicated to all the people who have helped me in my life. i Preface This book is a revised and expanded version of the lecture notes for Basic Calculus and other
Game Theory: Supermodular Games 1
Game Theory: Supermodular Games 1 Christoph Schottmüller 1 License: CC Attribution ShareAlike 4.0 1 / 22 Outline 1 Introduction 2 Model 3 Revision questions and exercises 2 / 22 Motivation I several solution
Factorization Algorithms for Polynomials over Finite Fields
Degree Project Factorization Algorithms for Polynomials over Finite Fields Sajid Hanif, Muhammad Imran 2011-05-03 Subject: Mathematics Level: Master Course code: 4MA11E Abstract Integer factorization is
INTRODUCTORY SET THEORY
M.Sc. program in mathematics INTRODUCTORY SET THEORY Katalin Károlyi Department of Applied Analysis, Eötvös Loránd University H-1088 Budapest, Múzeum krt. 6-8. CONTENTS 1. SETS Set, equal sets, subset,
Duality in General Programs. Ryan Tibshirani Convex Optimization 10-725/36-725
Duality in General Programs Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: duality in linear programs Given c R n, A R m n, b R m, G R r n, h R r : min x R n c T x max u R m, v R r b T
t := maxγ ν subject to ν {0,1,2,...} and f(x c +γ ν d) f(x c )+cγ ν f (x c ;d).
1. Line Search Methods Let f : R n R be given and suppose that x c is our current best estimate of a solution to P min x R nf(x). A standard method for improving the estimate x c is to choose a direction
The Henstock-Kurzweil-Stieltjes type integral for real functions on a fractal subset of the real line
The Henstock-Kurzweil-Stieltjes type integral for real functions on a fractal subset of the real line D. Bongiorno, G. Corrao Dipartimento di Ingegneria lettrica, lettronica e delle Telecomunicazioni,
GenOpt (R) Generic Optimization Program User Manual Version 3.0.0β1
(R) User Manual Environmental Energy Technologies Division Berkeley, CA 94720 http://simulationresearch.lbl.gov Michael Wetter [email protected] February 20, 2009 Notice: This work was supported by the U.S.
Parallel Total Variation Minimization. Diplomarbeit
Institut für Numerische und Angewandte Mathematik Parallel Total Variation Minimization Diplomarbeit eingereicht von Jahn Philipp Müller betreut von Prof. Dr. Martin Burger Prof. Dr. Sergei Gorlatch Münster
