An Idiot s guide to Support vector machines (SVMs)
|
|
|
- Milo Cook
- 9 years ago
- Views:
Transcription
1 An Idiot s guide to Support vector machines (SVMs) R. Berwick, Viage Idiot SVMs: A New Generation of Learning Agorithms Pre 1980: Amost a earning methods earned inear decision surfaces. Linear earning methods have nice theoretica properties 1980 s Decision trees and NNs aowed efficient earning of noninear decision surfaces Litte theoretica basis and a suffer from oca minima 1990 s Efficient earning agorithms for non-inear functions based on computationa earning theory deveoped Nice theoretica properties. 1
2 Key Ideas Two independent deveopments within ast decade New efficient separabiity of non-inear regions that use kerne functions : generaization of simiarity to new kinds of simiarity measures based on dot products Use of quadratic optimization probem to avoid oca minimum issues with neura nets The resuting earning agorithm is an optimization agorithm rather than a greedy search Organization Basic idea of support vector machines: just ike 1- ayer or muti-ayer neura nets Optima hyperpane for ineary separabe patterns Extend to patterns that are not ineary separabe by transformations of origina data to map into new space the Kerne function SVM agorithm for pattern recognition 2
3 Support Vectors Support vectors are the data points that ie cosest to the decision surface (or hyperpane) They are the data points most difficut to cassify They have direct bearing on the optimum ocation of the decision surface We can show that the optima hyperpane stems from the function cass with the owest capacity = # of independent features/parameters we can twidde [note this is extra materia not covered in the ectures you don t have to know this] Reca from 1-ayer nets : Which Separating Hyperpane? In genera, ots of possibe soutions for a,b,c (an infinite number!) Support Vector Machine (SVM) finds an optima soution 3
4 Support Vector Machine (SVM) SVMs maximize the margin (Winston terminoogy: the street ) around the separating hyperpane. The decision function is fuy specified by a (usuay very sma) subset of training sampes, the support vectors. This becomes a Quadratic programming probem that is easy to sove by standard methods Support vectors Maximize margin Separation by Hyperpanes Assume inear separabiity for now (we wi reax this ater) in 2 dimensions, can separate by a ine in higher dimensions, need hyperpanes 4
5 Genera input/output for SVMs just ike for neura nets, but for one important addition Input: set of (input, output) training pair sampes; ca the input sampe features x 1, x 2 x n, and the output resut y. Typicay, there can be ots of input features x i. Output: set of weights w (or w i ), one for each feature, whose inear combination predicts the vaue of y. (So far, just ike neura nets ) Important difference: we use the optimization of maximizing the margin ( street width ) to reduce the number of weights that are nonzero to just a few that correspond to the important features that matter in deciding the separating ine(hyperpane) these nonzero weights correspond to the support vectors (because they support the separating hyperpane) 2-D Case Find a,b,c, such that ax + by c for red points ax + by (or < ) c for green points. 5
6 Which Hyperpane to pick? Lots of possibe soutions for a,b,c. Some methods find a separating hyperpane, but not the optima one (e.g., neura net) But: Which points shoud infuence optimaity? A points? Linear regression Neura nets Or ony difficut points cose to decision boundary Support vector machines Support Vectors again for ineary separabe case Support vectors are the eements of the training set that woud change the position of the dividing hyperpane if removed. Support vectors are the critica eements of the training set The probem of finding the optima hyper pane is an optimization probem and can be soved by optimization techniques (we use Lagrange mutipiers to get this probem into a form that can be soved anayticay). 6
7 Support Vectors: Input vectors that just touch the boundary of the margin (street) circed beow, there are 3 of them (or, rather, the tips of the vectors w 0T x + b 0 = 1 or w 0T x + b 0 = 1 d X X X X X X Here, we have shown the actua support vectors, v 1, v 2, v 3, instead of just the 3 circed points at the tai ends of the support vectors. d denotes 1/2 of the street width d X X v 1 v 2 v 3 X X X X 7
8 Definitions Define the hyperpanes H such that: w x i +b +1 when =+1 w x i +b -1 when = 1 H 1 and H 2 are the panes: d + H 1 : w x i +b = +1 d - H H 2 : w x i +b = 1 The points on the panes H 1 and H 2 are the tips of the Support Vectors The pane H 0 is the median in between, where w x i +b =0 d+ = the shortest distance to the cosest positive point d- = the shortest distance to the cosest negative point The margin (gutter) of a separating hyperpane is d+ + d. H 2 H 0 H 1 Moving a support vector moves the decision boundary Moving the other vectors has no effect The optimization agorithm to generate the weights proceeds in such a way that ony the support vectors determine the weights and thus the boundary 8
9 Defining the separating Hyperpane Form of equation defining the decision surface separating the casses is a hyperpane of the form: w T x + b = 0 w is a weight vector x is input vector b is bias Aows us to write w T x + b 0 for d i = +1 w T x + b < 0 for d i = 1 Some fina definitions Margin of Separation (d): the separation between the hyperpane and the cosest data point for a given weight vector w and bias b. Optima Hyperpane (maxima margin): the particuar hyperpane for which the margin of separation d is maximized. 9
10 Maximizing the margin (aka street width) We want a cassifier (inear separator) with as big a margin as possibe. H 1 Reca the distance from a point(x 0,y 0 ) to a ine: Ax+By+c = 0 is: Ax 0 +By 0 +c /sqrt(a 2 +B 2 ), so, The distance between H 0 and H 1 is then: w x+b / w =1/ w, so The tota distance between H 1 and H 2 is thus: 2/ w H 2 H 0 d- d+ In order to maximize the margin, we thus need to minimize w. With the condition that there are no datapoints between H 1 and H 2 : x i w+b +1 when =+1 x i w+b 1 when = 1 Can be combined into: (x i w) 1 We now must sove a quadratic programming probem Probem is: minimize w, s.t. discrimination boundars obeyed, i.e., min f(x) s.t. g(x)=0, which we can rewrite as: min f: ½ w 2 (Note this is a quadratic function) s.t. g: (w x i ) b = 1 or [ (w x i ) b] 1 =0 This is a constrained optimization probem It can be soved by the Lagrangian mutiper method Because it is quadratic, the surface is a parabooid, with just a singe goba minimum (thus avoiding a probem we had with neura nets!) 10
11 fatten Exampe: parabooid 2+x 2 +2y 2 s.t. x+y=1 Intuition: find intersection of two functions f, g at a tangent point (intersection = both constraints satisfied; tangent = derivative is 0); this wi be a min (or max) for f s.t. the contraint g is satisfied Fattened parabooid f: 2x 2 +2y 2 =0 with superimposed constraint g: x +y = 1 Minimize when the constraint ine g (shown in green) is tangent to the inner eipse contour inez of f (shown in red) note direction of gradient arrows. 11
12 fattened parabooid f: 2+x 2 +2y 2 =0 with superimposed constraint g: x +y = 1; at tangent soution p, gradient vectors of f,g are parae (no possibe move to increment f that aso keeps you in region g) Minimize when the constraint ine g is tangent to the inner eipse contour ine of f Two constraints 1. Parae norma constraint (= gradient constraint on f, g s.t. soution is a max, or a min) 2. g(x)=0 (soution is on the constraint ine as we) We now recast these by combining f, g as the new Lagrangian function bntroducing new sack variabes denoted a or (more usuay, denoted α in the iterature) 12
13 Redescribing these conditions Want to ook for soution point p where " f ( p) = "! g( p) g( x) = 0 Or, combining these two as the Langrangian L & requiring derivative of L be zero: L(x,a) = f (x)! ag(x) "(x,a) = 0 At a soution p The the constraint ine g and the contour ines of f must be tangent If they are tangent, their gradient vectors (perpendicuars) are parae Gradient of g must be 0 i.e., steepest ascent & so perpendicuar to f Gradient of f must aso be in the same direction as g 13
14 How Langrangian soves constrained optimization L(x,a) = f (x)! ag(x) where "(x,a) = 0 Partia derivatives wrt x recover the parae norma constraint Partia derivatives wrt λ recover the g(x,y)=0 In genera, L(x,a) = f (x) +! a i i g i (x) In genera Gradient min of f constraint condition g L(x,a) = f (x) +! a i i g i (x) a function of n + m variabes n for the x' s, m for the a. Differentiating gives n + m equations, each set to 0. The n eqns differentiated wrt each x i give the gradient conditions; the m eqns differentiated wrt each a i recover the constraints g i In our case, f(x): ½ w 2 ; g(x): (w x i +b) 1=0 so Lagrangian is: min L= ½ w 2 Σa i [ (w x i +b) 1] wrt w, b We expand the ast to get the foowing L form: min L= ½ w 2 Σa i (w x i +b) +Σa i wrt w, b 14
15 Lagrangian Formuation So in the SVM probem the Lagrangian is min L P = 1 w 2! a 2 " i ( x i # w + b) + " a i s.t. $i, a i % 0 where is the # of training points From the property that the derivatives at min = 0 we get:!l P!w = w " # a i x i = 0!L P!b = " a i = 0 so w =! a i x i,! a i = 0 What s with this L p business? This indicates that this is the prima form of the optimization probem We wi actuay sove the optimization probem by now soving for the dua of this origina probem What is this dua formuation? 15
16 The Lagrangian Dua Probem: instead of minimizing over w, b, subject to constraints invoving a s, we can maximize over a (the dua variabe) subject to the reations obtained previousy for w and b Our soution must satisfy these two reations: w =! a i x i,! a i = 0 By substituting for w and b back in the origina eqn we can get rid of the dependence on w and b. Note first that we aready now have our answer for what the weights w must be: they are a inear combination of the training inputs and the training outputs, x i and and the vaues of a. We wi now sove for the a s by differentiating the dua probem wrt a, and setting it to zero. Most of the a s wi turn out to have the vaue zero. The non-zero a s wi correspond to the support vectors Prima probem: min L P = 1 w 2! a 2 " i x i # w + b s.t. $i a i % 0 ( ) + " a i w =! a i x i,! a i = 0 Dua probem: max L D (a i ) =! a i " 1! a 2 i a j y j x i # x j s.t.! a i = 0 & a i $ 0 ( ) (note that we have removed the dependence on w and b) 16
17 The Dua probem Kuhn-Tucker theorem: the soution we find here wi be the same as the soution to the origina probem Q: But why are we doing this???? (why not just sove the origina probem????) Ans: Because this wi et us sove the probem by computing the just the inner products of x i, x j (which wi be vermportant ater on when we want to sove non-ineary separabe cassification probems) Dua probem: The Dua Probem max L D (a i ) =! a i " 1! a 2 i a j y j x i # x j s.t.! a i = 0 & a i $ 0 ( ) Notice that a we have are the dot products of x i,x j If we take the derivative wrt a and set it equa to zero, we get the foowing soution, so we can sove for a i :! a i = 0 0 " a i " C 17
18 Now knowing the a i we can find the weights w for the maxima margin separating hyperpane: w =! a i x i And now, after training and finding the w by this method, given an unknown point u measured on features x i we can cassift by ooking at the sign of: f (x) = wiu + b = ( a i x i iu) + b Remember: most of the weights w i, i.e., the a s, wi be zero Ony the support vectors (on the gutters or margin) wi have nonzero weights or a s this reduces the dimensionaity of the soution! Inner products, simiarity, and SVMs Why shoud inner product kernes be invoved in pattern recognition using SVMs, or at a? Intuition is that inner products provide some measure of simiarity Inner product in 2D between 2 vectors of unit ength returns the cosine of the ange between them = how far apart they are e.g. x = [1, 0] T, y = [0, 1] T i.e. if they are parae their inner product is 1 (competey simiar) x T y = x y = 1 If they are perpendicuar (competey unike) their inner product is 0 (so shoud not contribute to the correct cassifier) x T y = x y = 0 18
19 Insight into inner products Consider that we are trying to maximize the form: L D (a i ) = a i! a i a j y j ( x i # x j )! " 1 2 s.t.! a i = 0 & a i $ 0 The caim is that this function wi be maximized if we give nonzero vaues to a s that correspond to the support vectors, ie, those that matter in fixing the maximum width margin ( street ). We, consider what this ooks ike. Note first from the constraint condition that a the a s are positive. Now et s think about a few cases. Case 1. If two features x i, x j are competey dissimiar, their dot product is 0, and they don t contribute to L. Case 2. If two features x i,x j are competey aike, their dot product is 0. There are 2 subcases. Subcase 1: both x i,and x j predict the same output vaue (either +1 or 1). Then x y j is aways 1, and the vaue of a i a j y j x i x j wi be positive. But this woud decrease the vaue of L (since it woud subtract from the first term sum). So, the agorithm downgrades simiar feature vectors that make the same prediction. Subcase 2: x i,and x j make opposite predictions about the output vaue (ie, one is +1, the other 1), but are otherwise very cosey simiar: then the product a i a j y j x i x is negative and we are subtracting it, so this adds to the sum, maximizing it. This is precisey the exampes we are ooking for: the critica ones that te the two cassses apart. Insight into inner products, graphicay: 2 very very simiar x i, x j vectors that predict difft casses tend to maximize the margin width x i x j 19
20 2 vectors that are simiar but predict the same cass are redundant x i x j 2 dissimiar (orthogona) vectors don t count at a x j x i 20
21 But are we done??? Not Lineary Separabe! Find a ine that penaizes points on the wrong side 21
22 Transformation to separate o o x x o x o o x x x o x ϕ ϕ (x) ϕ (x) ϕ (o) ϕ (x) ϕ (x) ϕ (o) ϕ (x) ϕ (o) ϕ (o) ϕ (x) ϕ (o) ϕ (x) ϕ (o) ϕ (o) X F Non Linear SVMs The idea is to gain ineary separation by mapping the data to a higher dimensiona space The foowing set can t be separated by a inear function, but can be separated by a quadratic one 2 ( )( ) ( ) x! a x! b = x! a + b x + ab a b { 2, } x! x x So if we map we gain inear separation 22
23 Probems with inear SVM =-1 =+1 What if the decision function is not inear? What transform woud separate these? Ans: poar coordinates! Non-inear SVM The Kerne trick Imagine a function φ that maps the data into another space: φ=radia Η Radia Η =-1 =+1 φ =-1 =+1 Remember the function we want to optimize: L d = a i ½ a i a j y j (x i x j ) where (x i x j ) is the dot product of the two feature vectors. If we now transform to φ, instead of computing this dot product (x i x j ) we wi have to compute (φ (x i ) φ (x j )). But how can we do this? This is expensive and time consuming (suppose φ is a quartic poynomia or worse, we don t know the function expicity. We, here is the neat thing: If there is a kerne function K such that K(x i,x j ) = φ (x i ) φ (x j ), then we do not need to know or compute φ at a!! That is, the kerne function defines inner products in the transformed space. Or, it defines simiaritn the transformed space. 23
24 Non-inear SVMs So, the function we end up optimizing is: L d = a i ½ a i a j y j K(x i x j ), Kerne exampe: The poynomia kerne K(xi,xj) = (x i x j + 1) p, where p is a tunabe parameter Note: Evauating K ony requires one addition and one exponentiation more than the origina dot product Exampes for Non Linear SVMs ( x y) ( x y ) K, =! + 1 p 2 " ( x, y) = exp{ " 2 2 } x y K! ( x, y) = tanh( x # y $ ) K! " 1 st is poynomia (incudes x x as specia case) 2 nd is radia basis function (gaussians) 3 rd is sigmoid (neura net activation function) 24
25 We ve aready seen such noninear transforms What is it??? tanh(β 0 x T x i + β 1 ) It s the sigmoid transform (for neura nets) So, SVMs subsume neura nets! (but w/o their probems ) Inner Product Kernes Type of Support Vector Machine Inner Product Kerne K(x,x i ), I = 1, 2,, N Usua inner product Poynomia earning machine (x T x i + 1) p Power p is specified a priori by the user Radia-basis function (RBF) exp(1/(2σ 2 ) x-x i 2 ) The width σ 2 is specified a priori Two ayer neura net tanh(β 0 x T x i + β 1 ) Actuay works ony for some vaues of β 0 and β 1 25
26 Kernes generaize the notion of inner product simiarity Note that one can define kernes over more than just vectors: strings, trees, structures, in fact, just about anything A very powerfu idea: used in comparing DNA, protein structure, sentence structures, etc. Exampes for Non Linear SVMs 2 Gaussian Kerne Linear Gaussian 26
27 Noninear rbf kerne Admira s deight w/ difft kerne functions 27
28 Overfitting by SVM Every point is a support vector too much freedom to bend to fit the training data no generaization. In fact, SVMs have an automatic way to avoid such issues, but we won t cover it here see the book by Vapnik, (We add a penaty function for mistakes made after training by over-fitting: reca that if one over-fits, then one wi tend to make errors on new data. This penaty fn can be put into the quadratic programming probem directy. You don t need to know this for this course.) 28
SAT Math Facts & Formulas
Numbers, Sequences, Factors SAT Mat Facts & Formuas Integers:..., -3, -2, -1, 0, 1, 2, 3,... Reas: integers pus fractions, decimas, and irrationas ( 2, 3, π, etc.) Order Of Operations: Aritmetic Sequences:
SAT Math Must-Know Facts & Formulas
SAT Mat Must-Know Facts & Formuas Numbers, Sequences, Factors Integers:..., -3, -2, -1, 0, 1, 2, 3,... Rationas: fractions, tat is, anyting expressabe as a ratio of integers Reas: integers pus rationas
1 Basic concepts in geometry
1 asic concepts in geometry 1.1 Introduction We start geometry with the simpest idea a point. It is shown using a dot, which is abeed with a capita etter. The exampe above is the point. straight ine is
Energy Density / Energy Flux / Total Energy in 3D
Lecture 5 Phys 75 Energy Density / Energy Fux / Tota Energy in D Overview and Motivation: In this ecture we extend the discussion of the energy associated with wave otion to waves described by the D wave
Physics 100A Homework 11- Chapter 11 (part 1) The force passes through the point A, so there is no arm and the torque is zero.
Physics A Homework - Chapter (part ) Finding Torque A orce F o magnitude F making an ange with the x axis is appied to a partice ocated aong axis o rotation A, at Cartesian coordinates (,) in the igure.
Angles formed by 2 Lines being cut by a Transversal
Chapter 4 Anges fored by 2 Lines being cut by a Transversa Now we are going to nae anges that are fored by two ines being intersected by another ine caed a transversa. 1 2 3 4 t 5 6 7 8 If I asked you
Math 447/547 Partial Differential Equations Prof. Carlson Homework 7 Text section 4.2 1. Solve the diffusion problem. u(t,0) = 0 = u x (t,l).
Math 447/547 Partia Differentia Equations Prof. Carson Homework 7 Text section 4.2 1. Sove the diffusion probem with the mixed boundary conditions u t = ku xx, < x
A Simple Introduction to Support Vector Machines
A Simple Introduction to Support Vector Machines Martin Law Lecture for CSE 802 Department of Computer Science and Engineering Michigan State University Outline A brief history of SVM Large-margin linear
Simultaneous Routing and Power Allocation in CDMA Wireless Data Networks
Simutaneous Routing and Power Aocation in CDMA Wireess Data Networks Mikae Johansson *,LinXiao and Stephen Boyd * Department of Signas, Sensors and Systems Roya Institute of Technoogy, SE 00 Stockhom,
5. Introduction to Robot Geometry and Kinematics
V. Kumar 5. Introduction to Robot Geometry and Kinematics The goa of this chapter is to introduce the basic terminoogy and notation used in robot geometry and kinematics, and to discuss the methods used
Finance 360 Problem Set #6 Solutions
Finance 360 Probem Set #6 Soutions 1) Suppose that you are the manager of an opera house. You have a constant margina cost of production equa to $50 (i.e. each additiona person in the theatre raises your
FRAME BASED TEXTURE CLASSIFICATION BY CONSIDERING VARIOUS SPATIAL NEIGHBORHOODS. Karl Skretting and John Håkon Husøy
FRAME BASED TEXTURE CLASSIFICATION BY CONSIDERING VARIOUS SPATIAL NEIGHBORHOODS Kar Skretting and John Håkon Husøy University of Stavanger, Department of Eectrica and Computer Engineering N-4036 Stavanger,
SECOND DERIVATIVE TEST FOR CONSTRAINED EXTREMA
SECOND DERIVATIVE TEST FOR CONSTRAINED EXTREMA This handout presents the second derivative test for a local extrema of a Lagrange multiplier problem. The Section 1 presents a geometric motivation for the
Several Views of Support Vector Machines
Several Views of Support Vector Machines Ryan M. Rifkin Honda Research Institute USA, Inc. Human Intention Understanding Group 2007 Tikhonov Regularization We are considering algorithms of the form min
Secure Network Coding with a Cost Criterion
Secure Network Coding with a Cost Criterion Jianong Tan, Murie Médard Laboratory for Information and Decision Systems Massachusetts Institute of Technoogy Cambridge, MA 0239, USA E-mai: {jianong, medard}@mit.edu
Increasing for all. Convex for all. ( ) Increasing for all (remember that the log function is only defined for ). ( ) Concave for all.
1. Differentiation The first derivative of a function measures by how much changes in reaction to an infinitesimal shift in its argument. The largest the derivative (in absolute value), the faster is evolving.
7. Dry Lab III: Molecular Symmetry
0 7. Dry Lab III: Moecuar Symmetry Topics: 1. Motivation. Symmetry Eements and Operations. Symmetry Groups 4. Physica Impications of Symmetry 1. Motivation Finite symmetries are usefu in the study of moecues.
3.5 Pendulum period. 2009-02-10 19:40:05 UTC / rev 4d4a39156f1e. g = 4π2 l T 2. g = 4π2 x1 m 4 s 2 = π 2 m s 2. 3.5 Pendulum period 68
68 68 3.5 Penduum period 68 3.5 Penduum period Is it coincidence that g, in units of meters per second squared, is 9.8, very cose to 2 9.87? Their proximity suggests a connection. Indeed, they are connected
Support Vector Machines Explained
March 1, 2009 Support Vector Machines Explained Tristan Fletcher www.cs.ucl.ac.uk/staff/t.fletcher/ Introduction This document has been written in an attempt to make the Support Vector Machines (SVM),
Normalization of Database Tables. Functional Dependency. Examples of Functional Dependencies: So Now what is Normalization? Transitive Dependencies
ISM 602 Dr. Hamid Nemati Objectives The idea Dependencies Attributes and Design Understand concepts normaization (Higher-Leve Norma Forms) Learn how to normaize tabes Understand normaization and database
Vector and Matrix Norms
Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a non-empty
A Supplier Evaluation System for Automotive Industry According To Iso/Ts 16949 Requirements
A Suppier Evauation System for Automotive Industry According To Iso/Ts 16949 Requirements DILEK PINAR ÖZTOP 1, ASLI AKSOY 2,*, NURSEL ÖZTÜRK 2 1 HONDA TR Purchasing Department, 41480, Çayırova - Gebze,
A Latent Variable Pairwise Classification Model of a Clustering Ensemble
A atent Variabe Pairwise Cassification Mode of a Custering Ensembe Vadimir Berikov Soboev Institute of mathematics, Novosibirsk State University, Russia [email protected] http://www.math.nsc.ru Abstract.
Support Vector Machine (SVM)
Support Vector Machine (SVM) CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
Face Hallucination and Recognition
Face Haucination and Recognition Xiaogang Wang and Xiaoou Tang Department of Information Engineering, The Chinese University of Hong Kong {xgwang1, xtang}@ie.cuhk.edu.hk http://mmab.ie.cuhk.edu.hk Abstract.
Introduction to Support Vector Machines. Colin Campbell, Bristol University
Introduction to Support Vector Machines Colin Campbell, Bristol University 1 Outline of talk. Part 1. An Introduction to SVMs 1.1. SVMs for binary classification. 1.2. Soft margins and multi-class classification.
Big Data Analytics CSCI 4030
High dim. data Graph data Infinite data Machine learning Apps Locality sensitive hashing PageRank, SimRank Filtering data streams SVM Recommen der systems Clustering Community Detection Web advertising
Chapter 3: JavaScript in Action Page 1 of 10. How to practice reading and writing JavaScript on a Web page
Chapter 3: JavaScript in Action Page 1 of 10 Chapter 3: JavaScript in Action In this chapter, you get your first opportunity to write JavaScript! This chapter introduces you to JavaScript propery. In addition,
Maintenance activities planning and grouping for complex structure systems
Maintenance activities panning and grouping for compex structure systems Hai Canh u, Phuc Do an, Anne Barros, Christophe Berenguer To cite this version: Hai Canh u, Phuc Do an, Anne Barros, Christophe
Least-Squares Intersection of Lines
Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a
ACO and SVM Selection Feature Weighting of Network Intrusion Detection Method
, pp. 129-270 http://dx.doi.org/10.14257/ijsia.2015.9.4.24 ACO and SVM Seection Feature Weighting of Network Intrusion Detection Method Wang Xingzhu Furong Coege Hunan, University of Arts and Science,
Linear Algebra Notes for Marsden and Tromba Vector Calculus
Linear Algebra Notes for Marsden and Tromba Vector Calculus n-dimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of
Statistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
Fast Robust Hashing. ) [7] will be re-mapped (and therefore discarded), due to the load-balancing property of hashing.
Fast Robust Hashing Manue Urueña, David Larrabeiti and Pabo Serrano Universidad Caros III de Madrid E-89 Leganés (Madrid), Spain Emai: {muruenya,darra,pabo}@it.uc3m.es Abstract As statefu fow-aware services
To give it a definition, an implicit function of x and y is simply any relationship that takes the form:
2 Implicit function theorems and applications 21 Implicit functions The implicit function theorem is one of the most useful single tools you ll meet this year After a while, it will be second nature to
Cooperative Content Distribution and Traffic Engineering in an ISP Network
Cooperative Content Distribution and Traffic Engineering in an ISP Network Wenjie Jiang, Rui Zhang-Shen, Jennifer Rexford, Mung Chiang Department of Computer Science, and Department of Eectrica Engineering
Multi-Robot Task Scheduling
Proc of IEEE Internationa Conference on Robotics and Automation, Karsruhe, Germany, 013 Muti-Robot Tas Scheduing Yu Zhang and Lynne E Parer Abstract The scheduing probem has been studied extensivey in
Discounted Cash Flow Analysis (aka Engineering Economy)
Discounted Cash Fow Anaysis (aka Engineering Economy) Objective: To provide economic comparison of benefits and costs that occur over time Assumptions: Future benefits and costs can be predicted A Benefits,
Lecture 2. Marginal Functions, Average Functions, Elasticity, the Marginal Principle, and Constrained Optimization
Lecture 2. Marginal Functions, Average Functions, Elasticity, the Marginal Principle, and Constrained Optimization 2.1. Introduction Suppose that an economic relationship can be described by a real-valued
CONDENSATION. Prabal Talukdar. Associate Professor Department of Mechanical Engineering IIT Delhi E-mail: [email protected]
CONDENSATION Praba Taukdar Associate Professor Department of Mechanica Engineering IIT Dehi E-mai: [email protected] Condensation When a vapor is exposed to a surface at a temperature beow T sat, condensation
LADDER SAFETY Table of Contents
Tabe of Contents SECTION 1. TRAINING PROGRAM INTRODUCTION..................3 Training Objectives...........................................3 Rationae for Training.........................................3
Pricing and Revenue Sharing Strategies for Internet Service Providers
Pricing and Revenue Sharing Strategies for Internet Service Providers Linhai He and Jean Warand Department of Eectrica Engineering and Computer Sciences University of Caifornia at Berkeey {inhai,wr}@eecs.berkeey.edu
Support Vector Machines
Support Vector Machines Charlie Frogner 1 MIT 2011 1 Slides mostly stolen from Ryan Rifkin (Google). Plan Regularization derivation of SVMs. Analyzing the SVM problem: optimization, duality. Geometric
SOLUTIONS. f x = 6x 2 6xy 24x, f y = 3x 2 6y. To find the critical points, we solve
SOLUTIONS Problem. Find the critical points of the function f(x, y = 2x 3 3x 2 y 2x 2 3y 2 and determine their type i.e. local min/local max/saddle point. Are there any global min/max? Partial derivatives
Lecture 3: Linear methods for classification
Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,
Chapter 1 Structural Mechanics
Chapter Structura echanics Introduction There are many different types of structures a around us. Each structure has a specific purpose or function. Some structures are simpe, whie others are compex; however
Risk Margin for a Non-Life Insurance Run-Off
Risk Margin for a Non-Life Insurance Run-Off Mario V. Wüthrich, Pau Embrechts, Andreas Tsanakas February 2, 2011 Abstract For sovency purposes insurance companies need to cacuate so-caed best-estimate
Lecture 2: The SVM classifier
Lecture 2: The SVM classifier C19 Machine Learning Hilary 2015 A. Zisserman Review of linear classifiers Linear separability Perceptron Support Vector Machine (SVM) classifier Wide margin Cost function
500 IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, VOL. 31, NO. 3, MARCH 2013
500 IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, VOL. 3, NO. 3, MARCH 203 Cognitive Radio Network Duaity Agorithms for Utiity Maximization Liang Zheng Chee Wei Tan, Senior Member, IEEE Abstract We
ASYMPTOTIC DIRECTION FOR RANDOM WALKS IN RANDOM ENVIRONMENTS arxiv:math/0512388v2 [math.pr] 11 Dec 2007
ASYMPTOTIC DIRECTION FOR RANDOM WALKS IN RANDOM ENVIRONMENTS arxiv:math/0512388v2 [math.pr] 11 Dec 2007 FRANÇOIS SIMENHAUS Université Paris 7, Mathématiques, case 7012, 2, pace Jussieu, 75251 Paris, France
Microeconomic Theory: Basic Math Concepts
Microeconomic Theory: Basic Math Concepts Matt Van Essen University of Alabama Van Essen (U of A) Basic Math Concepts 1 / 66 Basic Math Concepts In this lecture we will review some basic mathematical concepts
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, VOL. 31, NO. 12, DECEMBER 2013 1
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, VOL. 31, NO. 12, DECEMBER 2013 1 Scaabe Muti-Cass Traffic Management in Data Center Backbone Networks Amitabha Ghosh, Sangtae Ha, Edward Crabbe, and Jennifer
Market Design & Analysis for a P2P Backup System
Market Design & Anaysis for a P2P Backup System Sven Seuken Schoo of Engineering & Appied Sciences Harvard University, Cambridge, MA [email protected] Denis Chares, Max Chickering, Sidd Puri Microsoft
Budgeting Loans from the Social Fund
Budgeting Loans from the Socia Fund tes sheet Pease read these notes carefuy. They expain the circumstances when a budgeting oan can be paid. Budgeting Loans You may be abe to get a Budgeting Loan if:
A Branch-and-Price Algorithm for Parallel Machine Scheduling with Time Windows and Job Priorities
A Branch-and-Price Agorithm for Parae Machine Scheduing with Time Windows and Job Priorities Jonathan F. Bard, 1 Siwate Rojanasoonthon 2 1 Graduate Program in Operations Research and Industria Engineering,
In this section, we will consider techniques for solving problems of this type.
Constrained optimisation roblems in economics typically involve maximising some quantity, such as utility or profit, subject to a constraint for example income. We shall therefore need techniques for solving
Christfried Webers. Canberra February June 2015
c Statistical Group and College of Engineering and Computer Science Canberra February June (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 829 c Part VIII Linear Classification 2 Logistic
DEGREES OF ORDERS ON TORSION-FREE ABELIAN GROUPS
DEGREES OF ORDERS ON TORSION-FREE ABELIAN GROUPS ASHER M. KACH, KAREN LANGE, AND REED SOLOMON Abstract. We show that if H is an effectivey competey decomposabe computabe torsion-free abeian group, then
Early access to FAS payments for members in poor health
Financia Assistance Scheme Eary access to FAS payments for members in poor heath Pension Protection Fund Protecting Peope s Futures The Financia Assistance Scheme is administered by the Pension Protection
Constrained optimization.
ams/econ 11b supplementary notes ucsc Constrained optimization. c 2010, Yonatan Katznelson 1. Constraints In many of the optimization problems that arise in economics, there are restrictions on the values
Artificial Neural Networks and Support Vector Machines. CS 486/686: Introduction to Artificial Intelligence
Artificial Neural Networks and Support Vector Machines CS 486/686: Introduction to Artificial Intelligence 1 Outline What is a Neural Network? - Perceptron learners - Multi-layer networks What is a Support
Linear Programming Notes V Problem Transformations
Linear Programming Notes V Problem Transformations 1 Introduction Any linear programming problem can be rewritten in either of two standard forms. In the first form, the objective is to maximize, the material
2013 MBA Jump Start Program
2013 MBA Jump Start Program Module 2: Mathematics Thomas Gilbert Mathematics Module Algebra Review Calculus Permutations and Combinations [Online Appendix: Basic Mathematical Concepts] 2 1 Equation of
Core Maths C1. Revision Notes
Core Maths C Revision Notes November 0 Core Maths C Algebra... Indices... Rules of indices... Surds... 4 Simplifying surds... 4 Rationalising the denominator... 4 Quadratic functions... 4 Completing the
COMPARISON OF DIFFUSION MODELS IN ASTRONOMICAL OBJECT LOCALIZATION
COMPARISON OF DIFFUSION MODELS IN ASTRONOMICAL OBJECT LOCALIZATION Františe Mojžíš Department of Computing and Contro Engineering, ICT Prague, Technicá, 8 Prague [email protected] Abstract This
Chapter 6 Signal Data Mining from Wearable Systems
Chapter 6 Signa Data Mining from Wearabe Systems Francois G. Meyer 6.1 Definition of the Subject 6.1.1 Introduction Sensors from wearabe systems can be anayzed in rea-time on-site, or can be transmitted
Spherical Correlation of Visual Representations for 3D Model Retrieval
Noname manuscript No. (wi be inserted by the editor) Spherica Correation of Visua Representations for 3D Mode Retrieva Ameesh Makadia Kostas Daniiidis the date of receipt and acceptance shoud be inserted
Class #6: Non-linear classification. ML4Bio 2012 February 17 th, 2012 Quaid Morris
Class #6: Non-linear classification ML4Bio 2012 February 17 th, 2012 Quaid Morris 1 Module #: Title of Module 2 Review Overview Linear separability Non-linear classification Linear Support Vector Machines
Teaching fractions in elementary school: A manual for teachers
Teaching fractions in eementary schoo: A manua for teachers H. Wu Apri 0, 998 [Added December, 200] I have decided to resurrect this fie of 998 because, as a reativey short summary of the basic eements
NCH Software Bolt PDF Printer
NCH Software Bot PDF Printer This user guide has been created for use with Bot PDF Printer Version 1.xx NCH Software Technica Support If you have difficuties using Bot PDF Printer pease read the appicabe
Mechanical Engineering Drawing Workshop. Sample Drawings. Sample 1. Sample 1. Sample 2. Sample 2 2013/7/19. JIS Drawing Part 1
Mechanica Engineering Drawing Workshop Sampe Drawings JIS Drawing Part Content:.Sampe Drawings 2.Drawing sheets 3. Views and projection methods 4. Line Conventions 5. Scaes 6. Sectiona drawings 7. Detai
This unit will lay the groundwork for later units where the students will extend this knowledge to quadratic and exponential functions.
Algebra I Overview View unit yearlong overview here Many of the concepts presented in Algebra I are progressions of concepts that were introduced in grades 6 through 8. The content presented in this course
Support Vector Machines for Classification and Regression
UNIVERSITY OF SOUTHAMPTON Support Vector Machines for Classification and Regression by Steve R. Gunn Technical Report Faculty of Engineering, Science and Mathematics School of Electronics and Computer
Advanced ColdFusion 4.0 Application Development - 3 - Server Clustering Using Bright Tiger
Advanced CodFusion 4.0 Appication Deveopment - CH 3 - Server Custering Using Bri.. Page 1 of 7 [Figures are not incuded in this sampe chapter] Advanced CodFusion 4.0 Appication Deveopment - 3 - Server
Australian Bureau of Statistics Management of Business Providers
Purpose Austraian Bureau of Statistics Management of Business Providers 1 The principa objective of the Austraian Bureau of Statistics (ABS) in respect of business providers is to impose the owest oad
Leakage detection in water pipe networks using a Bayesian probabilistic framework
Probabiistic Engineering Mechanics 18 (2003) 315 327 www.esevier.com/ocate/probengmech Leakage detection in water pipe networks using a Bayesian probabiistic framework Z. Pouakis, D. Vaougeorgis, C. Papadimitriou*
Walrasian Demand. u(x) where B(p, w) = {x R n + : p x w}.
Walrasian Demand Econ 2100 Fall 2015 Lecture 5, September 16 Outline 1 Walrasian Demand 2 Properties of Walrasian Demand 3 An Optimization Recipe 4 First and Second Order Conditions Definition Walrasian
1 3 4 = 8i + 20j 13k. x + w. y + w
) Find the point of intersection of the lines x = t +, y = 3t + 4, z = 4t + 5, and x = 6s + 3, y = 5s +, z = 4s + 9, and then find the plane containing these two lines. Solution. Solve the system of equations
Prot Maximization and Cost Minimization
Simon Fraser University Prof. Karaivanov Department of Economics Econ 0 COST MINIMIZATION Prot Maximization and Cost Minimization Remember that the rm's problem is maximizing prots by choosing the optimal
Multi-variable Calculus and Optimization
Multi-variable Calculus and Optimization Dudley Cooke Trinity College Dublin Dudley Cooke (Trinity College Dublin) Multi-variable Calculus and Optimization 1 / 51 EC2040 Topic 3 - Multi-variable Calculus
CERTIFICATE COURSE ON CLIMATE CHANGE AND SUSTAINABILITY. Course Offered By: Indian Environmental Society
CERTIFICATE COURSE ON CLIMATE CHANGE AND SUSTAINABILITY Course Offered By: Indian Environmenta Society INTRODUCTION The Indian Environmenta Society (IES) a dynamic and fexibe organization with a goba vision
INTRODUCTION TO THE FINITE ELEMENT METHOD
INTRODUCTION TO THE FINITE ELEMENT METHOD Evgen Barkanov Institute of Materias and Structures Facut of Civi Engineering Riga Technica Universit Riga, Preface Toda the finite eement method (FEM) is considered
EFFICIENT CLUSTERING OF VERY LARGE DOCUMENT COLLECTIONS
Chapter 1 EFFICIENT CLUSTERING OF VERY LARGE DOCUMENT COLLECTIONS Inderjit S. Dhion, James Fan and Yuqiang Guan Abstract An invauabe portion of scientific data occurs naturay in text form. Given a arge
Definition 8.1 Two inequalities are equivalent if they have the same solution set. Add or Subtract the same value on both sides of the inequality.
8 Inequalities Concepts: Equivalent Inequalities Linear and Nonlinear Inequalities Absolute Value Inequalities (Sections 4.6 and 1.1) 8.1 Equivalent Inequalities Definition 8.1 Two inequalities are equivalent
Scheduling in Multi-Channel Wireless Networks
Scheduing in Muti-Channe Wireess Networks Vartika Bhandari and Nitin H. Vaidya University of Iinois at Urbana-Champaign, USA [email protected], [email protected] Abstract. The avaiabiity of mutipe orthogona
(a) We have x = 3 + 2t, y = 2 t, z = 6 so solving for t we get the symmetric equations. x 3 2. = 2 y, z = 6. t 2 2t + 1 = 0,
Name: Solutions to Practice Final. Consider the line r(t) = 3 + t, t, 6. (a) Find symmetric equations for this line. (b) Find the point where the first line r(t) intersects the surface z = x + y. (a) We
Week 3: Consumer and Firm Behaviour: The Work-Leisure Decision and Profit Maximization
AROEOOIS 2006 Week 3: onsumer and Firm Behaviour: The Work-Leisure Decision and Profit aximization Questions for Review 1. How are a consumer s preferences over goods represented? By utiity functions:
NCH Software Warp Speed PC Tune-up Software
NCH Software Warp Speed PC Tune-up Software This user guide has been created for use with Warp Speed PC Tune-up Software Version 1.xx NCH Software Technica Support If you have difficuties using Warp Speed
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
PENALTY TAXES ON CORPORATE ACCUMULATIONS
H Chapter Six H PENALTY TAXES ON CORPORATE ACCUMULATIONS INTRODUCTION AND STUDY OBJECTIVES The accumuated earnings tax and the persona hoding company tax are penaty taxes designed to prevent taxpayers
Duality in General Programs. Ryan Tibshirani Convex Optimization 10-725/36-725
Duality in General Programs Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: duality in linear programs Given c R n, A R m n, b R m, G R r n, h R r : min x R n c T x max u R m, v R r b T
Breakeven analysis and short-term decision making
Chapter 20 Breakeven anaysis and short-term decision making REAL WORLD CASE This case study shows a typica situation in which management accounting can be hepfu. Read the case study now but ony attempt
The width of single glazing. The warmth of double glazing.
Therma Insuation CI/SfB (31) Ro5 (M5) September 2012 The width of singe gazing. The warmth of doube gazing. Pikington Spacia Revoutionary vacuum gazing. Image courtesy of Lumen Roofight Ltd. Pikington
The Use of Cooling-Factor Curves for Coordinating Fuses and Reclosers
he Use of ooing-factor urves for oordinating Fuses and Recosers arey J. ook Senior Member, IEEE S& Eectric ompany hicago, Iinois bstract his paper describes how to precisey coordinate distribution feeder
Lecture 3. Linear Programming. 3B1B Optimization Michaelmas 2015 A. Zisserman. Extreme solutions. Simplex method. Interior point method
Lecture 3 3B1B Optimization Michaelmas 2015 A. Zisserman Linear Programming Extreme solutions Simplex method Interior point method Integer programming and relaxation The Optimization Tree Linear Programming
Comparison of Traditional and Open-Access Appointment Scheduling for Exponentially Distributed Service Time
Journa of Heathcare Engineering Vo. 6 No. 3 Page 34 376 34 Comparison of Traditiona and Open-Access Appointment Scheduing for Exponentiay Distributed Service Chongjun Yan, PhD; Jiafu Tang *, PhD; Bowen
Hybrid Process Algebra
Hybrid Process Agebra P.J.L. Cuijpers M.A. Reniers Eindhoven University of Technoogy (TU/e) Den Doech 2 5600 MB Eindhoven, The Netherands Abstract We deveop an agebraic theory, caed hybrid process agebra
