A NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION
|
|
|
- Laurel Grant
- 9 years ago
- Views:
Transcription
1 1 A NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION Dimitri Bertsekas M.I.T. FEBRUARY 2003
2 2 OUTLINE Convexity issues in optimization Historical remarks Our treatment of the subject Three unifying lines of analysis Common geometrical framework for duality and minimax Unifying framework for existence of solutions and duality gap analysis Unification of Lagrange multiplier theory using an enhanced Fritz John theory and the notion of pseudonormality
3 3 WHY IS CONVEXITY IMPORTANT IN OPTIMIZATION I A convex function has no local minima that are not global A nonconvex function can be convexified while maintaining the optimality of its minima A convex set has nonempty relative interior A convex set has feasible directions at any point f(x) Epigraph f(x) Epigraph Convex set Feasible Directions Nonconvex set x x Convex function Nonconvex function
4 WHY IS CONVEXITY IMPORTANT IN OPTIMIZATION II The existence of minima of convex functions is conveniently characterized using directions of recession A polyhedral convex set is characterized by its extreme points and extreme directions 4 v 1 P 0 Recession Cone R f v 2 Level Sets of Convex Function f v 3
5 5 WHY IS CONVEXITY IMPORTANT IN OPTIMIZATION III A convex function is continuous and has nice differentiability properties Convex functions arise prominently in duality Convex, lower semicontinuous functions are self-dual with respect to conjugacy f(z) (x, f(x)) (-d, 1) z
6 6 SOME HISTORY Late 19th-Early 20th Century: Caratheodory, Minkowski, Steinitz, Farkas 40s-50s: The big turning point Game Theory: von Neumann Duality: Fenchel, Gale, Kuhn, Tucker Optimization-related convexity: Fenchel 60s-70s: Consolidation Rockafellar 80s-90s: Extensions to nonconvex optimization and nonsmooth analysis Clarke, Mordukovich, Rockafellar-Wets
7 7 ABOUT THE BOOK Convex Analysis and Optimization, by D. P. Bertsekas, with A. Nedic and A. Ozdaglar (March 2003) Aims to make the subject accessible through unification and geometric visualization Unification is achieved through several new lines of analysis
8 8 NEW LINES OF ANALYSIS I A unified geometrical approach to convex programming duality and minimax theory Basis: Duality between two elementary geometrical problems II A unified view of theory of existence of solutions and absence of duality gap Basis: Reduction to basic questions on intersections of closed sets IIIA unified view of theory of existence of Lagrange multipliers/constraint qualifications Basis: The notion of constraint pseudonormality, motivated by a new set of enhanced Fritz John conditions
9 9 OUTLINE OF THE BOOK Chapter 1 Basic Convexity Concepts Convex Sets and Functions Semicontinuity and Epigraphs Convex and Affine Hulls Relative Interior Directions of Recession Closed Set Intersections Chapters 2-4 Existence of Optimal Solutions Min Common/ Max Crossing Duality Minimax Saddle Point Theory Polyhedral Convexity Representation Thms Extreme Points Linear Programming Integer Programming Subdifferentiability Calculus of Subgradients Conical Approximations Optimality Conditions Chapters 5-8 Constrained Optimization Duality Geometric Multipliers Strong Duality Results Fenchel Duality Conjugacy Exact Penalty Functions Lagrange Multiplier Theory Pseudonormality Constraint Qualifications Exact Penalty Functions Fritz John Theory Dual Methods Subgradient and Incremental Methods
10 10 OUTLINE OF DUALITY ANALYSIS Closed Set Intersection Results Min Common/ Max Crossing Theory Enhanced Fritz John Conditions Preservation of Closedness Under Partial Minimization p(u) = inf x F(x,u) Minimax Theory Nonlinear Farkas Lemma Absence of Duality Gap Fenchel Duality Theory Existence of Geometric Multipliers Existence of Informative Geometric Multipliers
11 11 I MIN COMMON/MAX CROSSING DUALITY
12 12 GEOMETRICAL VIEW OF DUALITY w Min Common Point w * _ M w Min Common Point w * M M 0 Max Crossing Point q * u Max Crossing Point q * 0 u (a) (b) Min Common Point w * w _ M Max Crossing Point q * S M 0 u (c)
13 13 APPROACH Prove theorems about the geometry of M Conditions on M guarantee that w* = q* Conditions on M that guarantee existence of a max crossing hyperplane Choose M to reduce the constrained optimization problem, the minimax problem, and others, to special cases of the min common/max crossing framework Specialize the min common/max crossing theorems to duality and minimax theorems
14 14 CONVEX PROGRAMMING DUALITY Primal problem: min f(x) subject to x X and g j (x) 0, j=1,,r Dual problem: max q(µ) subject to µ 0 where the dual function is q(µ) = inf x L(x,µ) = inf x {f(x) + µ g(x)} Optimal primal value = inf x sup L(x,µ) µ 0 Optimal dual value = sup µ 0 inf x L(x,µ) Min common/max crossing framework: M = epi(p), p(u) = inf x X, gj(x) uj f(x)
15 15 VISUALIZATION (µ,1) M {(g(x),f(x)) x X} q * f* p(u) q(µ) = inf u {p(u) + µ u} u
16 16 MINIMAX / ZERO SUM GAME THEORY ISSUES Given a function Φ(x,z), where x X and z Z, under what conditions do we have inf x sup z Φ(x,z) = sup z inf x Φ(x,z) Assume convexity/concavity, semicontinuity of Φ Min common/max crossing framework: M = epigraph of the function p(u) = inf x sup z {Φ(x,z) - u z} inf x sup z Φ = Min common value sup z inf x Φ = Max crossing value
17 17 VISUALIZATION w w inf x sup z φ(x,z) (µ,1) M = epi(p) = min common value w * M = epi(p) (µ,1) inf x sup z φ(x,z) 0 q(µ) = min common value w * u sup z inf x φ(x,z) sup z inf x φ(x,z) = max crossing value q * 0 q(µ) u = max crossing value q * (a) (b)
18 18 TWO ISSUES IN CONVEX PROGRAMMING AND MINIMAX When is there no duality gap ( in convex programming), or inf sup = sup inf (in minimax)? When does an optimal dual solution exist (in convex programming), or the sup is attained (in minimax)? Min common/max crossing framework shows that 1st question is a lower semicontinuity issue 2nd question is an issue of existence of a nonvertical support hyperplane (or subgradient) at the origin Further analysis is needed for more specific answers
19 19 II UNIFICATION OF EXISTENCE AND NO DUALITY GAP ISSUES
20 20 INTERSECTIONS OF NESTED FAMILIES OF CLOSED SETS Two basic problems in convex optimization Attainment of a minimum of a function f over a set X Existence of a duality gap The 1st question is a set intersection issue: The set of minima is the intersection of the nonempty level sets {x X f(x) γ} The 2nd question is a lower semicontinuity issue: When is the function p(u) = inf x F(x,u) lower semicontinuous, assuming F(x,u) is convex and lower semicontinuous?
21 21 PRESERVATION OF SEMICONTINUITY UNDER PARTIAL MINIMIZATION 2nd question also involves set intersection Key observation: For p(u) = inf x F(x,u), we have Closure(P(epi(F))) epi(p) P(epi(F)) where P(. ) is projection on the space of u. So if projection preserves closedness, F is l.s.c. C C k+1 x C k y y k+1 y k P(C) Given C, when is P(C) closed? If y k is a sequence in P(C) that converges to y, - we must show that the intersection of the C k is nonempty
22 22 UNIFIED TREATMENT OF EXISTENCE OF SOLUTIONS AND DUALITY GAP ISSUES Results on nonemptiness of intersection of a nested family of closed sets No duality gap results In convex programming inf sup Φ = sup inf Φ Existence of minima of f over X
23 23 THE ROLE OF POLYHEDRAL ASSUMPTIONS: AN EXAMPLE Polyhedral Constraint Set X Level Sets of Convex Function f Nonpolyhedral Constraint Set X Level Sets of Convex Function f x 2 x 2 X X 0 x 1 0 x 1 Minimum Attained f(x) = e x 1 Minimum not Attained The min is attained if X is polyhedral, and f is constant along common directions of recession of f and X
24 24 THE ROLE OF QUADRATIC FUNCTIONS Results on nonemptiness of intersection of sets defined by quadratic inequalities If f is bounded below over X, the min of f over X is attained If the optimal value is finite, there is no duality gap Linear programming Quadratic programming Semidefinite programming
25 25 III LAGRANGE MULTIPLIER THEORY / PSEUDONORMALITY
26 26 LAGRANGE MULTIPLIERS Problem (smooth, nonconvex): Min f(x) subject to x X, h i (x)=0, i = 1,,m Necessary condition for optimality of x* (case X = R n ): Under some constraint qualification, we have f(x*) + Σ I λ i h i (x*) = 0 for some Lagrange multipliers λ i Basic analytical issue: What is the fundamental structure of the constraint set that guarantees the existence of a Lagrange multiplier? Standard constraint qualifications (case X = R n ): The gradients h i (x*) are linearly independent The functions h i are affine
27 27 ENHANCED FRITZ JOHN CONDITIONS If x* is optimal, there exist µ 0 0 and λ i, not all 0, such that µ 0 f(x*) + Σ i λ i h i (x*) = 0, and a sequence {x k } with x k x* and such that f (xk ) < f(x*) for all k, λ i h i (x k ) > 0 for all i with λ i 0 and all k NOTE: If µ 0 > 0, the λ i are Lagrange multipliers with a sensitivity property (we call them informative)
28 28 PSEUDONORMALITY Definition: A feasible point x* is pseudonormal if one cannot find λ i and a sequence {x k } with x k x* such that Σ i λ i h i (x*) = 0, Σ i λ i h i (x k ) > 0 for all k Pseudonormality at x* guarantees that, if x* is optimal, we can take µ 0 = 1 in the F-J conditions (so there exists an informative Lagrange multiplier) u 2 u 2 0 λ Tε u 1 λ 0 Tε u 1 Map an ε -ball around x* onto the constraint space T ε = {h(x) x-x* < ε} Pseudonormal h i : Affine Not Pseudonormal
29 29 INFORMATIVE LAGRANGE MULTIPLIERS The Lagrange multipliers obtained from the enhanced Fritz John conditions have a special sensitivity property: They indicate the constraints to violate in order to improve the cost We call such multipliers informative Proposition: If there exists at least one Lagrange multiplier vector, there exists one that is informative
30 30 EXTENSIONS/CONNECTIONS TO NONSMOOTH ANALAYSIS F-J conditions for an additional constraint x X The stationarity condition becomes - ( f(x*) + Σ I λ i h i (x*)) (normal cone of X at x*) X is called regular at x* if the normal cone is equal to the polar of its tangent cone at x* (example: X convex) If X is not regular at x*, the Lagrangian may have negative slope along some feasible directions Regularity is the fault line beyond which there is no satisfactory Lagrange multiplier theory
31 31 THE STRUCTURE OF THE THEORY FOR X = R n Independent Constraint Gradients Linear Constraints Mangasarian Fromovitz Condition Pseudonormality Existence of Lagrange Multipliers Existence of Informative Multipliers
32 32 THE STRUCTURE OF THE THEORY FOR X: REGULAR New Mangasarian Fromovitz-Like Condition Slater Condition Pseudonormality Existence of Lagrange Multipliers Existence of Informative Multipliers
33 33 EXTENSIONS Enhanced Fritz John conditions and pseudonormality for convex problems, when existence of a primal optimal solution is not assumed Connection of pseudonormality and exact penalty functions Connection of pseudonormality and the classical notion of quasiregularity
2.3 Convex Constrained Optimization Problems
42 CHAPTER 2. FUNDAMENTAL CONCEPTS IN CONVEX OPTIMIZATION Theorem 15 Let f : R n R and h : R R. Consider g(x) = h(f(x)) for all x R n. The function g is convex if either of the following two conditions
Date: April 12, 2001. Contents
2 Lagrange Multipliers Date: April 12, 2001 Contents 2.1. Introduction to Lagrange Multipliers......... p. 2 2.2. Enhanced Fritz John Optimality Conditions...... p. 12 2.3. Informative Lagrange Multipliers...........
Duality of linear conic problems
Duality of linear conic problems Alexander Shapiro and Arkadi Nemirovski Abstract It is well known that the optimal values of a linear programming problem and its dual are equal to each other if at least
Numerisches Rechnen. (für Informatiker) M. Grepl J. Berger & J.T. Frings. Institut für Geometrie und Praktische Mathematik RWTH Aachen
(für Informatiker) M. Grepl J. Berger & J.T. Frings Institut für Geometrie und Praktische Mathematik RWTH Aachen Wintersemester 2010/11 Problem Statement Unconstrained Optimality Conditions Constrained
Further Study on Strong Lagrangian Duality Property for Invex Programs via Penalty Functions 1
Further Study on Strong Lagrangian Duality Property for Invex Programs via Penalty Functions 1 J. Zhang Institute of Applied Mathematics, Chongqing University of Posts and Telecommunications, Chongqing
Convex analysis and profit/cost/support functions
CALIFORNIA INSTITUTE OF TECHNOLOGY Division of the Humanities and Social Sciences Convex analysis and profit/cost/support functions KC Border October 2004 Revised January 2009 Let A be a subset of R m
Some representability and duality results for convex mixed-integer programs.
Some representability and duality results for convex mixed-integer programs. Santanu S. Dey Joint work with Diego Morán and Juan Pablo Vielma December 17, 2012. Introduction About Motivation Mixed integer
10. Proximal point method
L. Vandenberghe EE236C Spring 2013-14) 10. Proximal point method proximal point method augmented Lagrangian method Moreau-Yosida smoothing 10-1 Proximal point method a conceptual algorithm for minimizing
SECOND DERIVATIVE TEST FOR CONSTRAINED EXTREMA
SECOND DERIVATIVE TEST FOR CONSTRAINED EXTREMA This handout presents the second derivative test for a local extrema of a Lagrange multiplier problem. The Section 1 presents a geometric motivation for the
Duality in General Programs. Ryan Tibshirani Convex Optimization 10-725/36-725
Duality in General Programs Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: duality in linear programs Given c R n, A R m n, b R m, G R r n, h R r : min x R n c T x max u R m, v R r b T
Several Views of Support Vector Machines
Several Views of Support Vector Machines Ryan M. Rifkin Honda Research Institute USA, Inc. Human Intention Understanding Group 2007 Tikhonov Regularization We are considering algorithms of the form min
A FIRST COURSE IN OPTIMIZATION THEORY
A FIRST COURSE IN OPTIMIZATION THEORY RANGARAJAN K. SUNDARAM New York University CAMBRIDGE UNIVERSITY PRESS Contents Preface Acknowledgements page xiii xvii 1 Mathematical Preliminaries 1 1.1 Notation
(Basic definitions and properties; Separation theorems; Characterizations) 1.1 Definition, examples, inner description, algebraic properties
Lecture 1 Convex Sets (Basic definitions and properties; Separation theorems; Characterizations) 1.1 Definition, examples, inner description, algebraic properties 1.1.1 A convex set In the school geometry
Nonlinear Optimization: Algorithms 3: Interior-point methods
Nonlinear Optimization: Algorithms 3: Interior-point methods INSEAD, Spring 2006 Jean-Philippe Vert Ecole des Mines de Paris [email protected] Nonlinear optimization c 2006 Jean-Philippe Vert,
Big Data - Lecture 1 Optimization reminders
Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Schedule Introduction Major issues Examples Mathematics
CONSTRAINED NONLINEAR PROGRAMMING
149 CONSTRAINED NONLINEAR PROGRAMMING We now turn to methods for general constrained nonlinear programming. These may be broadly classified into two categories: 1. TRANSFORMATION METHODS: In this approach
Convex Programming Tools for Disjunctive Programs
Convex Programming Tools for Disjunctive Programs João Soares, Departamento de Matemática, Universidade de Coimbra, Portugal Abstract A Disjunctive Program (DP) is a mathematical program whose feasible
Introduction to Support Vector Machines. Colin Campbell, Bristol University
Introduction to Support Vector Machines Colin Campbell, Bristol University 1 Outline of talk. Part 1. An Introduction to SVMs 1.1. SVMs for binary classification. 1.2. Soft margins and multi-class classification.
MINIMIZATION OF ENTROPY FUNCTIONALS UNDER MOMENT CONSTRAINTS. denote the family of probability density functions g on X satisfying
MINIMIZATION OF ENTROPY FUNCTIONALS UNDER MOMENT CONSTRAINTS I. Csiszár (Budapest) Given a σ-finite measure space (X, X, µ) and a d-tuple ϕ = (ϕ 1,..., ϕ d ) of measurable functions on X, for a = (a 1,...,
Walrasian Demand. u(x) where B(p, w) = {x R n + : p x w}.
Walrasian Demand Econ 2100 Fall 2015 Lecture 5, September 16 Outline 1 Walrasian Demand 2 Properties of Walrasian Demand 3 An Optimization Recipe 4 First and Second Order Conditions Definition Walrasian
GI01/M055 Supervised Learning Proximal Methods
GI01/M055 Supervised Learning Proximal Methods Massimiliano Pontil (based on notes by Luca Baldassarre) (UCL) Proximal Methods 1 / 20 Today s Plan Problem setting Convex analysis concepts Proximal operators
15 Kuhn -Tucker conditions
5 Kuhn -Tucker conditions Consider a version of the consumer problem in which quasilinear utility x 2 + 4 x 2 is maximised subject to x +x 2 =. Mechanically applying the Lagrange multiplier/common slopes
Research Article Stability Analysis for Higher-Order Adjacent Derivative in Parametrized Vector Optimization
Hindawi Publishing Corporation Journal of Inequalities and Applications Volume 2010, Article ID 510838, 15 pages doi:10.1155/2010/510838 Research Article Stability Analysis for Higher-Order Adjacent Derivative
Summer course on Convex Optimization. Fifth Lecture Interior-Point Methods (1) Michel Baes, K.U.Leuven Bharath Rangarajan, U.
Summer course on Convex Optimization Fifth Lecture Interior-Point Methods (1) Michel Baes, K.U.Leuven Bharath Rangarajan, U.Minnesota Interior-Point Methods: the rebirth of an old idea Suppose that f is
1 Norms and Vector Spaces
008.10.07.01 1 Norms and Vector Spaces Suppose we have a complex vector space V. A norm is a function f : V R which satisfies (i) f(x) 0 for all x V (ii) f(x + y) f(x) + f(y) for all x,y V (iii) f(λx)
Adaptive Online Gradient Descent
Adaptive Online Gradient Descent Peter L Bartlett Division of Computer Science Department of Statistics UC Berkeley Berkeley, CA 94709 bartlett@csberkeleyedu Elad Hazan IBM Almaden Research Center 650
Nonlinear Programming Methods.S2 Quadratic Programming
Nonlinear Programming Methods.S2 Quadratic Programming Operations Research Models and Methods Paul A. Jensen and Jonathan F. Bard A linearly constrained optimization problem with a quadratic objective
Support Vector Machines
Support Vector Machines Charlie Frogner 1 MIT 2011 1 Slides mostly stolen from Ryan Rifkin (Google). Plan Regularization derivation of SVMs. Analyzing the SVM problem: optimization, duality. Geometric
Max-Min Representation of Piecewise Linear Functions
Beiträge zur Algebra und Geometrie Contributions to Algebra and Geometry Volume 43 (2002), No. 1, 297-302. Max-Min Representation of Piecewise Linear Functions Sergei Ovchinnikov Mathematics Department,
Cost Minimization and the Cost Function
Cost Minimization and the Cost Function Juan Manuel Puerta October 5, 2009 So far we focused on profit maximization, we could look at a different problem, that is the cost minimization problem. This is
Lecture 2: August 29. Linear Programming (part I)
10-725: Convex Optimization Fall 2013 Lecture 2: August 29 Lecturer: Barnabás Póczos Scribes: Samrachana Adhikari, Mattia Ciollaro, Fabrizio Lecci Note: LaTeX template courtesy of UC Berkeley EECS dept.
The Heat Equation. Lectures INF2320 p. 1/88
The Heat Equation Lectures INF232 p. 1/88 Lectures INF232 p. 2/88 The Heat Equation We study the heat equation: u t = u xx for x (,1), t >, (1) u(,t) = u(1,t) = for t >, (2) u(x,) = f(x) for x (,1), (3)
Support Vector Machine (SVM)
Support Vector Machine (SVM) CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
Separation Properties for Locally Convex Cones
Journal of Convex Analysis Volume 9 (2002), No. 1, 301 307 Separation Properties for Locally Convex Cones Walter Roth Department of Mathematics, Universiti Brunei Darussalam, Gadong BE1410, Brunei Darussalam
1 Introduction. Linear Programming. Questions. A general optimization problem is of the form: choose x to. max f(x) subject to x S. where.
Introduction Linear Programming Neil Laws TT 00 A general optimization problem is of the form: choose x to maximise f(x) subject to x S where x = (x,..., x n ) T, f : R n R is the objective function, S
Error Bound for Classes of Polynomial Systems and its Applications: A Variational Analysis Approach
Outline Error Bound for Classes of Polynomial Systems and its Applications: A Variational Analysis Approach The University of New South Wales SPOM 2013 Joint work with V. Jeyakumar, B.S. Mordukhovich and
ECONOMIC THEORY AND OPERATIONS ANALYSIS
WILLIAM J. BAUMOL Professor of Economics Princeton University ECONOMIC THEORY AND OPERATIONS ANALYSIS Second Edition Prentice-Hall, I Inc. Engkwood Cliffs, New Jersey CONTENTS PART 7 ANALYTIC TOOLS OF
TOPIC 4: DERIVATIVES
TOPIC 4: DERIVATIVES 1. The derivative of a function. Differentiation rules 1.1. The slope of a curve. The slope of a curve at a point P is a measure of the steepness of the curve. If Q is a point on the
Lecture 2: The SVM classifier
Lecture 2: The SVM classifier C19 Machine Learning Hilary 2015 A. Zisserman Review of linear classifiers Linear separability Perceptron Support Vector Machine (SVM) classifier Wide margin Cost function
24. The Branch and Bound Method
24. The Branch and Bound Method It has serious practical consequences if it is known that a combinatorial problem is NP-complete. Then one can conclude according to the present state of science that no
CHAPTER 9. Integer Programming
CHAPTER 9 Integer Programming An integer linear program (ILP) is, by definition, a linear program with the additional constraint that all variables take integer values: (9.1) max c T x s t Ax b and x integral
Recovery of primal solutions from dual subgradient methods for mixed binary linear programming; a branch-and-bound approach
MASTER S THESIS Recovery of primal solutions from dual subgradient methods for mixed binary linear programming; a branch-and-bound approach PAULINE ALDENVIK MIRJAM SCHIERSCHER Department of Mathematical
No: 10 04. Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics
No: 10 04 Bilkent University Monotonic Extension Farhad Husseinov Discussion Papers Department of Economics The Discussion Papers of the Department of Economics are intended to make the initial results
The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method
The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method Robert M. Freund February, 004 004 Massachusetts Institute of Technology. 1 1 The Algorithm The problem
GenOpt (R) Generic Optimization Program User Manual Version 3.0.0β1
(R) User Manual Environmental Energy Technologies Division Berkeley, CA 94720 http://simulationresearch.lbl.gov Michael Wetter [email protected] February 20, 2009 Notice: This work was supported by the U.S.
Optimization. J(f) := λω(f) + R emp (f) (5.1) m l(f(x i ) y i ). (5.2) i=1
5 Optimization Optimization plays an increasingly important role in machine learning. For instance, many machine learning algorithms minimize a regularized risk functional: with the empirical risk min
A Simple Introduction to Support Vector Machines
A Simple Introduction to Support Vector Machines Martin Law Lecture for CSE 802 Department of Computer Science and Engineering Michigan State University Outline A brief history of SVM Large-margin linear
CURVES WHOSE SECANT DEGREE IS ONE IN POSITIVE CHARACTERISTIC. 1. Introduction
Acta Math. Univ. Comenianae Vol. LXXXI, 1 (2012), pp. 71 77 71 CURVES WHOSE SECANT DEGREE IS ONE IN POSITIVE CHARACTERISTIC E. BALLICO Abstract. Here we study (in positive characteristic) integral curves
Downloaded 08/02/12 to 132.68.160.18. Redistribution subject to SIAM license or copyright; see http://www.siam.org/journals/ojsa.
SIAM J. OPTIM. Vol. 22, No. 2, pp. 557 580 c 2012 Society for Industrial and Applied Mathematics SMOOTHING AND FIRST ORDER METHODS: A UNIFIED FRAMEWORK AMIR BECK AND MARC TEBOULLE Abstract. We propose
Two-Stage Stochastic Linear Programs
Two-Stage Stochastic Linear Programs Operations Research Anthony Papavasiliou 1 / 27 Two-Stage Stochastic Linear Programs 1 Short Reviews Probability Spaces and Random Variables Convex Analysis 2 Deterministic
Lecture 7: Finding Lyapunov Functions 1
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.243j (Fall 2003): DYNAMICS OF NONLINEAR SYSTEMS by A. Megretski Lecture 7: Finding Lyapunov Functions 1
Numerical Verification of Optimality Conditions in Optimal Control Problems
Numerical Verification of Optimality Conditions in Optimal Control Problems Dissertation zur Erlangung des naturwissenschaftlichen Doktorgrades der Julius-Maximilians-Universität Würzburg vorgelegt von
Critical points of once continuously differentiable functions are important because they are the only points that can be local maxima or minima.
Lecture 0: Convexity and Optimization We say that if f is a once continuously differentiable function on an interval I, and x is a point in the interior of I that x is a critical point of f if f (x) =
Metric Spaces. Chapter 1
Chapter 1 Metric Spaces Many of the arguments you have seen in several variable calculus are almost identical to the corresponding arguments in one variable calculus, especially arguments concerning convergence
Solutions Of Some Non-Linear Programming Problems BIJAN KUMAR PATEL. Master of Science in Mathematics. Prof. ANIL KUMAR
Solutions Of Some Non-Linear Programming Problems A PROJECT REPORT submitted by BIJAN KUMAR PATEL for the partial fulfilment for the award of the degree of Master of Science in Mathematics under the supervision
THE FUNDAMENTAL THEOREM OF ALGEBRA VIA PROPER MAPS
THE FUNDAMENTAL THEOREM OF ALGEBRA VIA PROPER MAPS KEITH CONRAD 1. Introduction The Fundamental Theorem of Algebra says every nonconstant polynomial with complex coefficients can be factored into linear
Optimization in R n Introduction
Optimization in R n Introduction Rudi Pendavingh Eindhoven Technical University Optimization in R n, lecture Rudi Pendavingh (TUE) Optimization in R n Introduction ORN / 4 Some optimization problems designing
t := maxγ ν subject to ν {0,1,2,...} and f(x c +γ ν d) f(x c )+cγ ν f (x c ;d).
1. Line Search Methods Let f : R n R be given and suppose that x c is our current best estimate of a solution to P min x R nf(x). A standard method for improving the estimate x c is to choose a direction
Lecture 3. Linear Programming. 3B1B Optimization Michaelmas 2015 A. Zisserman. Extreme solutions. Simplex method. Interior point method
Lecture 3 3B1B Optimization Michaelmas 2015 A. Zisserman Linear Programming Extreme solutions Simplex method Interior point method Integer programming and relaxation The Optimization Tree Linear Programming
ALMOST COMMON PRIORS 1. INTRODUCTION
ALMOST COMMON PRIORS ZIV HELLMAN ABSTRACT. What happens when priors are not common? We introduce a measure for how far a type space is from having a common prior, which we term prior distance. If a type
Metric Spaces. Chapter 7. 7.1. Metrics
Chapter 7 Metric Spaces A metric space is a set X that has a notion of the distance d(x, y) between every pair of points x, y X. The purpose of this chapter is to introduce metric spaces and give some
Notes on metric spaces
Notes on metric spaces 1 Introduction The purpose of these notes is to quickly review some of the basic concepts from Real Analysis, Metric Spaces and some related results that will be used in this course.
Introduction to Algebraic Geometry. Bézout s Theorem and Inflection Points
Introduction to Algebraic Geometry Bézout s Theorem and Inflection Points 1. The resultant. Let K be a field. Then the polynomial ring K[x] is a unique factorisation domain (UFD). Another example of a
Support Vector Machines Explained
March 1, 2009 Support Vector Machines Explained Tristan Fletcher www.cs.ucl.ac.uk/staff/t.fletcher/ Introduction This document has been written in an attempt to make the Support Vector Machines (SVM),
Optimal shift scheduling with a global service level constraint
Optimal shift scheduling with a global service level constraint Ger Koole & Erik van der Sluis Vrije Universiteit Division of Mathematics and Computer Science De Boelelaan 1081a, 1081 HV Amsterdam The
Mathematical Methods of Engineering Analysis
Mathematical Methods of Engineering Analysis Erhan Çinlar Robert J. Vanderbei February 2, 2000 Contents Sets and Functions 1 1 Sets................................... 1 Subsets.............................
The Equivalence of Linear Programs and Zero-Sum Games
The Equivalence of Linear Programs and Zero-Sum Games Ilan Adler, IEOR Dep, UC Berkeley [email protected] Abstract In 1951, Dantzig showed the equivalence of linear programming problems and two-person
constraint. Let us penalize ourselves for making the constraint too big. We end up with a
Chapter 4 Constrained Optimization 4.1 Equality Constraints (Lagrangians) Suppose we have a problem: Maximize 5, (x 1, 2) 2, 2(x 2, 1) 2 subject to x 1 +4x 2 =3 If we ignore the constraint, we get the
Third CI 2 MA Focus Seminar Optimization: Asymptotic Analysis, Strong Duality and Relaxation
Third CI 2 MA Focus Seminar Optimization: Asymptotic Analysis, Strong Duality and Relaxation April 17, 2012 Auditorio Alamiro Robledo Facultad de Ciencias Físicas y Matemáticas Universidad de Concepción
Class Meeting # 1: Introduction to PDEs
MATH 18.152 COURSE NOTES - CLASS MEETING # 1 18.152 Introduction to PDEs, Fall 2011 Professor: Jared Speck Class Meeting # 1: Introduction to PDEs 1. What is a PDE? We will be studying functions u = u(x
3. Linear Programming and Polyhedral Combinatorics
Massachusetts Institute of Technology Handout 6 18.433: Combinatorial Optimization February 20th, 2009 Michel X. Goemans 3. Linear Programming and Polyhedral Combinatorics Summary of what was seen in the
Chapter 4. Duality. 4.1 A Graphical Example
Chapter 4 Duality Given any linear program, there is another related linear program called the dual. In this chapter, we will develop an understanding of the dual linear program. This understanding translates
Lecture 6: Logistic Regression
Lecture 6: CS 194-10, Fall 2011 Laurent El Ghaoui EECS Department UC Berkeley September 13, 2011 Outline Outline Classification task Data : X = [x 1,..., x m]: a n m matrix of data points in R n. y { 1,
Envelope Theorem. Kevin Wainwright. Mar 22, 2004
Envelope Theorem Kevin Wainwright Mar 22, 2004 1 Maximum Value Functions A maximum (or minimum) value function is an objective function where the choice variables have been assigned their optimal values.
Geometrical Characterization of RN-operators between Locally Convex Vector Spaces
Geometrical Characterization of RN-operators between Locally Convex Vector Spaces OLEG REINOV St. Petersburg State University Dept. of Mathematics and Mechanics Universitetskii pr. 28, 198504 St, Petersburg
Choice under Uncertainty
Choice under Uncertainty Part 1: Expected Utility Function, Attitudes towards Risk, Demand for Insurance Slide 1 Choice under Uncertainty We ll analyze the underlying assumptions of expected utility theory
An Overview Of Software For Convex Optimization. Brian Borchers Department of Mathematics New Mexico Tech Socorro, NM 87801 borchers@nmt.
An Overview Of Software For Convex Optimization Brian Borchers Department of Mathematics New Mexico Tech Socorro, NM 87801 [email protected] In fact, the great watershed in optimization isn t between linearity
HOMEWORK 5 SOLUTIONS. n!f n (1) lim. ln x n! + xn x. 1 = G n 1 (x). (2) k + 1 n. (n 1)!
Math 7 Fall 205 HOMEWORK 5 SOLUTIONS Problem. 2008 B2 Let F 0 x = ln x. For n 0 and x > 0, let F n+ x = 0 F ntdt. Evaluate n!f n lim n ln n. By directly computing F n x for small n s, we obtain the following
Least-Squares Intersection of Lines
Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a
International Doctoral School Algorithmic Decision Theory: MCDA and MOO
International Doctoral School Algorithmic Decision Theory: MCDA and MOO Lecture 2: Multiobjective Linear Programming Department of Engineering Science, The University of Auckland, New Zealand Laboratoire
. P. 4.3 Basic feasible solutions and vertices of polyhedra. x 1. x 2
4. Basic feasible solutions and vertices of polyhedra Due to the fundamental theorem of Linear Programming, to solve any LP it suffices to consider the vertices (finitely many) of the polyhedron P of the
The Envelope Theorem 1
John Nachbar Washington University April 2, 2015 1 Introduction. The Envelope Theorem 1 The Envelope theorem is a corollary of the Karush-Kuhn-Tucker theorem (KKT) that characterizes changes in the value
A QUICK GUIDE TO THE FORMULAS OF MULTIVARIABLE CALCULUS
A QUIK GUIDE TO THE FOMULAS OF MULTIVAIABLE ALULUS ontents 1. Analytic Geometry 2 1.1. Definition of a Vector 2 1.2. Scalar Product 2 1.3. Properties of the Scalar Product 2 1.4. Length and Unit Vectors
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem
Constrained optimization.
ams/econ 11b supplementary notes ucsc Constrained optimization. c 2010, Yonatan Katznelson 1. Constraints In many of the optimization problems that arise in economics, there are restrictions on the values
Metric Spaces Joseph Muscat 2003 (Last revised May 2009)
1 Distance J Muscat 1 Metric Spaces Joseph Muscat 2003 (Last revised May 2009) (A revised and expanded version of these notes are now published by Springer.) 1 Distance A metric space can be thought of
Linear Algebra I. Ronald van Luijk, 2012
Linear Algebra I Ronald van Luijk, 2012 With many parts from Linear Algebra I by Michael Stoll, 2007 Contents 1. Vector spaces 3 1.1. Examples 3 1.2. Fields 4 1.3. The field of complex numbers. 6 1.4.
Notes from Week 1: Algorithms for sequential prediction
CS 683 Learning, Games, and Electronic Markets Spring 2007 Notes from Week 1: Algorithms for sequential prediction Instructor: Robert Kleinberg 22-26 Jan 2007 1 Introduction In this course we will be looking
Proximal mapping via network optimization
L. Vandenberghe EE236C (Spring 23-4) Proximal mapping via network optimization minimum cut and maximum flow problems parametric minimum cut problem application to proximal mapping Introduction this lecture:
Spatial Decomposition/Coordination Methods for Stochastic Optimal Control Problems. Practical aspects and theoretical questions
Spatial Decomposition/Coordination Methods for Stochastic Optimal Control Problems Practical aspects and theoretical questions P. Carpentier, J-Ph. Chancelier, M. De Lara, V. Leclère École des Ponts ParisTech
What is Linear Programming?
Chapter 1 What is Linear Programming? An optimization problem usually has three essential ingredients: a variable vector x consisting of a set of unknowns to be determined, an objective function of x to
9 More on differentiation
Tel Aviv University, 2013 Measure and category 75 9 More on differentiation 9a Finite Taylor expansion............... 75 9b Continuous and nowhere differentiable..... 78 9c Differentiable and nowhere monotone......
BANACH AND HILBERT SPACE REVIEW
BANACH AND HILBET SPACE EVIEW CHISTOPHE HEIL These notes will briefly review some basic concepts related to the theory of Banach and Hilbert spaces. We are not trying to give a complete development, but
Math 120 Final Exam Practice Problems, Form: A
Math 120 Final Exam Practice Problems, Form: A Name: While every attempt was made to be complete in the types of problems given below, we make no guarantees about the completeness of the problems. Specifically,
