A Novel Hyperbolic Smoothing Algorithm for Clustering Large Data Sets

Size: px
Start display at page:

Download "A Novel Hyperbolic Smoothing Algorithm for Clustering Large Data Sets"

Transcription

1 A Novel Hyperbolic Smoothing Algorithm for Clustering Large Data Sets Abstract Adilson Elias Xavier Vinicius Layter Xavier Dept. of Systems Engineering and Computer Science Graduate School of Engineering (COPPE) Federal University of Rio de Janeiro P.O. Box Rio de Janeiro,RJ , BRAZIL The minimum sum-of-squares clustering problem is considered. The mathematical modeling of this problem leads to a min sum min formulation which, in addition to its intrinsic bi-level nature, has the significant characteristic of being strongly nondifferentiable. To overcome these difficulties, the proposed resolution method, called Hyperbolic Smoothing, adopts a smoothing strategy using a special C differentiable class function. The final solution is obtained by solving a sequence of low dimension differentiable unconstrained optimization This paper presents an extended method based upon the partition of the set of observations in two non overlapping parts. This last approach engenders a drastic simplification of the computational tasks. Keywords: Cluster Analysis, Pattern Recognition, Min-Sum- Min Problems, Nondifferentiable Programming, Smoothing 1 XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1535

2 1 Introduction Cluster analysis deals with the problems of classification of a set of patterns or observations, in general represented as points in a multidimensional space, into clusters, following two basic and simultaneous objectives: patterns in the same clusters must be similar to another (homogeneity objective) and different from patterns of other clusters (separation objective), see Hartingan (1973) and Späth (1980). A particular clustering problem formulation is considered. Among many criteria used in cluster analysis, the most natural, intuitive and frequently adopted criterion is the minimum sum-of-squares clustering (MSSC). It is a criterion for both the homogeneity and the separation objectives. The minimum sum-of-squares clustering (MSSC) formulation produces a mathematical problem of global optimization. It is both a nondifferentiable and a nonconvex mathematical problem, with a large number of local minimizers. In the cluster analysis scope, algorithms use, traditionally, two main strategies: hierarchical clustering methods and partition clustering methods (Hansen and Jaumard (1997) and Jain et alli (1999)). This paper adopts a different strategy, a smooth approach, which was originally presented in Sousa(2005) and Xavier(2008). For the sake of completeness, we present first the smoothing of the min sum min problem engendered by the modeling of the clustering problem. In a sense, the process whereby this is achieved is an extension of a smoothing scheme, called Hyperbolic Smoothing. This technique was developed through an adaptation of the hyperbolic penalty method originally introduced by Xavier (1982). By smoothing, we fundamentally mean the substitution of an intrinsically nondifferentiable two-level problem by a C unconstrained differentiable single-level alternative. Additionally, the paper presents a extended faster procedure. The basic idea is the partition of the set of observations in two non overlapping parts. By using a conceptual presentation, the first set corresponds to the observation points relatively close to two or more centroids. This set of observations, named boundary band points, can be managed by using the previously presented smoothing approach. The second set corresponds to observation points significantly closer to a single centroid in comparison with others. The set of observations, named gravitational points, is managed in a direct and simple way, offering much faster performance. For the purpose of illustrating both the reliability and the efficiency of the method, a set of computational experiments was performed, making use of traditional test problems described in the literature. XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1536

3 2 The Clustering Problem Transformation Let S = {s 1,..., s m } denote a set of m patterns or observations from an Euclidean n-space, to be clustered into a given number q of disjoint clusters. To formulate the original clustering problem as a min sum min problem, we proceed as follows. Let x i, i = 1,..., q be the centroids of the clusters, where each x i R n. The set of these centroid coordinates will be represented by X R nq. Given a point s j of S, we initially calculate the distance from s j to the center in X that is nearest. This is given by z j = min xi X s j x i 2. The most frequent measurement of the quality of a clustering associated to a specific position of q centroids is provided by the sum of the squares of these distances, which determines the following problem: subject to minimize zj 2 (1) z j = min,...,q s j x i 2, j = 1,..., m. Considering its definition, each z j must necessarily satisfy the following set of inequalities: z j s j x i 2 0, i = 1,..., q. Substituting these inequalities for the equality constraints of problem (1), a relaxed problem would be obtained. Since the variables z j are not bounded from below, the optimum solution of the relaxed problem will be z j = 0, j = 1,..., m. In order to obtain a bounded formulation, we do so by first letting ϕ(y) denote max{0, y} and then observing that, from the set of constraints in (1), it follows that ϕ(z j s j x i 2 ) = 0, j = 1,..., m. (2) By using (2) in place of the set of constraints in (1) and by including a perturbation ɛ > 0 we would obtain the bounded problem: subject to minimize zj 2 (3) ϕ(z j s j x i 2 ) ε, j = 1,..., m. Since the feasible set of problem (1) is the limit of that of (3) when ɛ 0 +, we can then consider solving (1) by solving a sequence of problems like (3) for a sequence of decreasing values for ε that approaches 0. XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1537

4 Analyzing the problem (3), the definition of function ϕ endows it with an extremely rigid nondifferentiable structure, which makes its computational solution very hard. From this perspective, for y R and τ > 0, let us define the function: φ(y, τ) = (y + y 2 + τ 2 ) /2. (4) Function φ constitutes an approximation of function ϕ. To obtain a completely differentiable problem, it is yet necessary to smooth the Euclidean distance s j x i 2. For this purpose, let us use the function θ( s j, x i, γ ) = n (s l j xl i )2 + γ 2 (5) l=1 By using function θ in place of the distance the completely differentiable problem: s j x i 2, it is obtained subject to minimize zj 2 (6) φ(z j θ(s j, x i, γ), τ) ε, j = 1,..., m. So, the properties of functions φ and θ allow us to seek a solution to problem (3) by solving a sequence of subproblems like problem (6), produced by the decreasing of the parameters γ 0, τ 0, and ε 0. On the other hand, given any set of centroids x i, i = 1,..., q, due to the increasing convex property of the hyperbolic smoothing function φ, the constraints of problem (6) will certainly be active. So, it will at last be equivalent to problem: minimize subject to h j (z j, x) = zj 2 (7) φ(z j θ(s j, x i, γ), τ) ε = 0, j = 1,..., m. The dimension of variable domain space of problem (7) is (nq + m). However, it has a separable structure, because each variable z j appears only in one equality constraint. Therefore, as the partial derivative of h(z j, x) with respect to z j, j = 1,..., m is not equal to zero, it is possible to use the XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1538

5 Implicit Function Theorem to calculate each component z j, j = 1,..., m as a function of the centroid variables x i, i = 1,..., q. In this way, the unconstrained problem minimize f(x) = z j (x) 2 (8) is obtained, where each z j (x) results from the calculation of the unique zero of each equation h j (z j, x) = φ(z j θ(s j, x i, γ), τ) ε = 0, j = 1,..., m. (9) Again, due to the Implicit Function Theorem, the functions z j (x) have all derivatives with respect to the variables x i, i = 1,..., q, and therefore it is possible to calculate the gradient of the objective function of problem (8), where f(x) = 2 z j (x) z j (x), (10) z j (x) = h j (z j, x) / h j(z j, x) z j, (11) while h j (z j, x) and h j (z j, x)/ z j are obtained from equations (4), (5) and (9). In this way, it is easy to solve problem (8) by making use of any method based on first order derivative information. At last, it must be emphasized that problem (8) is defined on a (nq) dimensional space, so it is a small problem, since the number of clusters, q, is, in general, very small for real applications. 3 The Partition Procedure The calculation of the objective function of the problem (8) demands the determination of the zeros of m equations (9), one equation for each observation point. Now, it will be presented a faster procedure. The basic idea is the partition of the set of observations in two non overlapping parts. By using a conceptual presentation, the first set corresponds to the observation points that are relatively close to two or more centroids. The second set corresponds to the observation points that are significantly close to a unique centroid in comparison with the other ones. XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1539

6 So, the first part J B is the set of boundary observations and the second is the set J G of gravitational observations. Considering this partition, equation (8) can be expressed as a sum of two independent components. minimize f(x) = z j (x) 2 = z j (x) 2 + z j (x) 2. (12) j J B j J G minimize f(x) = f 1 (x) + f 2 (x). (13) The first part of expression (13), associated with the boundary observations, can be calculated by using the previous presented smoothing approach, see (8) and (9). The second part of expression (13) can be calculated by using a faster procedure, as we will show right away. Let us define the two parts in a more rigorous form. Let be x i, i = 1,..., q be a referential position of centroids of the clusters taken in the iterative process. The boundary concept in relation to the referential point x can be easily specified by defining a δ band zone between neighboring centroids. For a generic observation s R n, we define the first and second nearest distances from s to the centroids: d 1 (s, x) = s x i1 = min i s x i and d 2 (s, x) = s x i2 = min i i1 s x i, where i 1 and i 2 are the labeling indices of these two nearest centroids. By using the above definitions, let us define precisely the δ boundary band zone: Z δ (x) = {s R n d 2 (s, x) d 1 (s, x) < δ } (14) and the gravitational region, this is the complementary space: G δ (x) = {s R n Z δ (x) }. (15) Figure 1 illustrates in R 2 the Z δ (x) and G δ (x) partitions. The central lines form the Voronoy polygon associated with the referential centroids x i, i = 1,..., q. The region between two parallel lines to Voronoy lines constitutes the boundary band zone Z δ (x). Now, the two sets can be defined in a precise form: J B (x) = {j s j Z δ (x)} and J G (x) = {j s j G δ (x)}. Proposition 1: Let s be a generic point belonging to the gravitational region G δ (x), with nearest centroid i 1. Let x be the current position of the centroids. Let x = max i x i x i be the maximum displacement of the centroids. If x < δ then s will continue to be nearest to centroid x i1 than to any other. Since δ x, Proposition 1 makes it possible to calculate exactly expression (12) in a very fast way. First, let us define the subsets of gravitational observations associated with each referential centroid: J i (x) = { } j J G min s j x l = s j x i l=1,...,q (16) XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1540

7 Figure 1: The Z δ (x) and G δ (x) partitions. The gravity centers of the observations in each non-empty subset are given by v i = 1 J i s j J i s j, i = 1,..., q. (17) Let us consider the second sum in expression (12). It will be computed by taking into account the gravity centers defined above. minimize f 2 (x) = j J G z j (x) 2 = j J i s j x i 2 = j J i s j v i + v i x i 2 = s j v i 2 + J i x i v i 2 (18) j J i When the position of centroids x i, i = 1,..., q moves during iterative process, the value of the first sum in (18) assumes a constant value, since the vectors s and v are fixed. On the other hand, for the calculation of the second sum, it is only necessary to calculate q distances, namely: v i x i, i = 1,..., q. The gradient of the second part of the objective function is easily calculated by: f 2 (x) = 2 J i ( x i v i ). (19) So, if it is observed that δ x within the iterative process, j J G z j (x) 2 and its gradient can be computed exactly by very fast procedures. XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1541

8 Simplified Extended Hyperbolic Smoothing Clustering Algorithm Initialization Step: Choose start point: x 0 ; Choose parameter values: γ 1, τ 1, ε 1 ; Choose reduction factors: 0 < ρ 1 < 1, 0 < ρ 2 < 1, 0 < ρ 3 < 1; Specify the boundary band width: δ 1 ; Let k = 1. Main Step: Repeat until a stopping rule is attained For determining the Z δ (x) and G δ (x) partitions, given by (14) and (15), use x = x k 1 and δ = δ k. Calculate the gravity centers of the observations v i, i = 1,..., q of each non-empty subset by using (17). Solve problem (13) starting at the initial point x k 1 and let x k be the solution obtained: For solving the equations (9), associated to the first part given by (8), take the smoothing parameters: γ = γ k, τ = τ k and ε = ε k ; For solving the second part, given by (18), use the above calculated gravity centers of the observations. Updating procedure: Let γ k+1 = ρ 1 γ k, τ k+1 = ρ 2 τ k, ε k+1 = ρ 3 ε k Redefine the boundary value: δ k+1. Let k := k + 1. Just as in other smoothing methods, the solution to the clustering problem is obtained, in theory, by solving an infinite sequence of optimization problems. In the Extended Hyperbolic Smoothing Clustering (XHSC) algorithm, each problem to be minimized is unconstrained and of low dimension. Notice that the algorithm causes τ and γ to approach 0, so the constraints of the subproblems as given in (6) tend to those of (3). In addition, the algorithm causes ε to approach 0, so, in a simultaneous movement, the solved problem (3) gradually approaches the original MSSC problem (1). The efficiency of the XHSC Algorithm depends strongly on the choice of boundary band width δ. A small value choice will imply an improper definition of the set G δ (x) and the frequent violation of the basic condition, x < δ, for the valid of Proposition 1. Otherwise, a big value choice will imply in a decrease of the number of observations belonging to the gravitational region and, therefore, the computational advantages given by the formulation (18) will be reduced. XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1542

9 4 Computational Results The computational results presented below were obtained from a preliminary implementation of the Extended Hyperbolic Smoothing Clustering algorithm. The numerical experiments have been carried out on a PC Intel Celeron with 2.7GHz CPU and 512MB RAM. The programs are coded with Compac Visual FORTRAN, Version 6.1. The unconstrained minimization tasks were carried out by means of a Quasi-Newton algorithm employing the BFGS updating formula from the Harwell Library, obtained in the site: ( In order to show the distinct performance of the XHSC algorithm, results obtained by solving a set of the largest problems of the TSP collection are shown below ( Table 1 presents the results for the TSPLIB-3038 data set. It exhibits the results produced by XHSC algorithm and, for comparison, those of two algorithms presented in Bagirov(2008). The first two columns show the number of clusters (q) and the best known value for the global optimum (f opt ) taken from Bagirov(2008). The next columns show the error (E) for the best solution produced (f Best ) and the mean CPU time (T ime) given in seconds associated to three algorithms: multi-start k-means (MS k-means), modified global k-means (MGKM) and the proposed XHSM. The errors are calculated in the following way: E = 100 (f Best f opt ) / f opt. The multi-start k-means algorithm is the traditional k-means algorithm with multiple initial starting points. In this experiment, to find q clusters, 100 times q starting points were randomly chosen in the MS k-means algorithm. The global k-means (GKM) algorithm, introduced by Likas et alli (2003), is a significant improvement of the k-means algorithm. The MGKS is an improved version of the GKM algorithm proposed by Bagirov(2008). The XHSM solutions were produced by using 10 starting points in all cases, except q = 40 and q = 50, where 20 and 40 starting points were taken, respectively. q f opt MS k-means MGKM XHSC Algorithm E T ime E T ime E T ime E E E E E E E E E Table 1: Results for the TSPLIB-3038 Instance XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1543

10 It is possible to observe in each row of Table 1 that the best solution produced by the new XHSC Algorithm becomes significantly smaller than that by MS k-means when the number of clusters q increases. In fact, this algorithm do not perform well for big instances, in despite being one of most used algorithms. Wu et alli (2008) present the top 10 data mining algorithms identified by the IEEE International Conference on Data Mining in December The k-means assumes the second place in this list. The comparison between XHSC and MGKM solutions demonstrates similar superiority of proposed algorithm. In the same way, the comparison of time columns shows a consistent speed advantage of proposed algorithm over the older ones. On the other hand, the best solution produced by the XHSC Algorithm is very close to the putative global minimum, the best known solution of the TSPLIB-3038 instance. Moreover, in this preliminary experiment, by using a relatively small number of initial starting points, four new putative global minimum results (q = 40, q = 60, q = 80 and q = 100) have been established. Instance q = 5 q = 10 f XHSCBest E Mean T ime Mean f XHSCBest E Mean T ime Mean FL E E FNL E E RL E E RL E E Pla E E RL E E USA E E BRD E E D E E BRD E E Pla E E Pla E E Table 2: Results for larger instances of the TSPLIB collection Table 2 presents the computational results produced by the XHSC Algorithm for the largest instances of Symmetric Traveling Salesman Problem (TSP) collection: FL3795, FNL4461, RL5915, RL5934, Pla7397, RL11849, USA13509, BRD14051, D15112, BRD18512, Pla33810 and Pla For each instance, it is presented two cases: q = 5 and q = 10. Ten different randomly chosen starting points were used. For each case, the table presents: the best objective function value produced by the XHSC Algorithm (f XHSCBest ), the average error of the 10 solutions in relation to the best solution obtained (E Mean ) and CPU mean time given in seconds (T ime Mean ). It was impossible to perform a comparison on given the lack of records of solutions of these instances. Indeed, the clustering literature seldom considers instances with such number of observations. Only a possible remark, the low XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1544

11 values presented in columns (E Mean ) show a consistent performance of the proposed algorithm. 5 Conclusions In this paper, the Extended HSC algorithm, a new method for the solution of the minimum sum-of-squares clustering problem, has been proposed. The robustness performance of the XHSC Algorithm can be attributed to the complete differentiability of the approach. The speed performance of the XHSC Algorithm can be attributed to the partition of the set of observations in two non overlapping parts. This last approach engenders a drastic simplification of computational tasks. From the results presented in Tables 1 and 2, as well other results obtained for solving classical problems of the literature (see Sousa(2005)), it was possible to observe that the implementation of the XHSC algorithm reaches the best known solution, in most of the cases, otherwise calculates a solution which is close to the best. Moreover, some new best results in the cluster literature have been established, as these shown in the present paper for the TSPLIB-3038 instance. Finally, it must be remembered that the MSSC problem is a global optimization problem with a lot of local minima, so the algorithm only can produce local minima. However, the obtained computational results presented a deep local minima property, perfectly adequate to the necessities and demands of real applications. References BAGIROV, A. M. and YEARWOOD, J. (2006) A New Nonsmooth Optimization Algorithm for Minimum Sum-of-Squares Clustering Problems, European Journal of Operational Research, 170 pp BAGIROV, A. M. (2008) Modified Global k-means Algorithm for Minimum Sum-of-Squares Clustering Problems, Pattern Recognition, Vol 41 Issue 10 pp DUBES, R. C. and JAIN, A. K. (1976) Cluster Techniques: The User s Dilemma, Pattern Recognition No. 8 pp HANSEN, P. and JAUMARD, B. (1997) Cluster Analysis and Mathematical Programming, Mathematical Programming No. 79 pp HARTINGAN, J. A. (1975) Clustering Algorithms, John Wiley and Sons, Inc., New York, NY. JAIN, A. K. and DUBES, R. C. (1988) Algorithms for Clustering Data, Prentice-Hall Inc., Upper Saddle River, NJ. XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1545

12 JAIN, A. K.; MURTY, M. N. and FLYNN, P. J. (1999) Data Clustering: A Review, ACM Computing Surveys, Vol. 31, No. 3, Sept LIKAS, A.; VLASSIS, M. and VERBEEK, J. (2003) The Global k-means Clustering Algorithm, Pattern Recognition, Vol 36 pp SOUSA, L.C.F. (2005) Desempenho Computacional do Método de Agrupamento Via Suavização Hiperbólica, M.Sc. thesis - COPPE - UFRJ, Rio de Janeiro. SPÄTH, H. (1980) Cluster Analysis Algorithms for Data Reduction and Classification, Ellis Horwood, Upper Saddle River, NJ. WU, X. et alli (2008) Top 10 Algorithms in Data Mining, Knowledge and Information Systems, Vol 14 No. 1, pp 1-37 XAVIER, A.E. (1982) Penalização Hiperbólica: Um Novo Método para Resolução de Problemas de Otimização, M.Sc. Thesis - COPPE - UFRJ, Rio de Janeiro. XAVIER, A.E. (2008) The Hyperbolic Smothing Clustering Method, submited to Pattern Recognition. XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1546

Introduction to Support Vector Machines. Colin Campbell, Bristol University

Introduction to Support Vector Machines. Colin Campbell, Bristol University Introduction to Support Vector Machines Colin Campbell, Bristol University 1 Outline of talk. Part 1. An Introduction to SVMs 1.1. SVMs for binary classification. 1.2. Soft margins and multi-class classification.

More information

24. The Branch and Bound Method

24. The Branch and Bound Method 24. The Branch and Bound Method It has serious practical consequences if it is known that a combinatorial problem is NP-complete. Then one can conclude according to the present state of science that no

More information

Standardization and Its Effects on K-Means Clustering Algorithm

Standardization and Its Effects on K-Means Clustering Algorithm Research Journal of Applied Sciences, Engineering and Technology 6(7): 399-3303, 03 ISSN: 040-7459; e-issn: 040-7467 Maxwell Scientific Organization, 03 Submitted: January 3, 03 Accepted: February 5, 03

More information

Support Vector Machines with Clustering for Training with Very Large Datasets

Support Vector Machines with Clustering for Training with Very Large Datasets Support Vector Machines with Clustering for Training with Very Large Datasets Theodoros Evgeniou Technology Management INSEAD Bd de Constance, Fontainebleau 77300, France theodoros.evgeniou@insead.fr Massimiliano

More information

ARTIFICIAL INTELLIGENCE (CSCU9YE) LECTURE 6: MACHINE LEARNING 2: UNSUPERVISED LEARNING (CLUSTERING)

ARTIFICIAL INTELLIGENCE (CSCU9YE) LECTURE 6: MACHINE LEARNING 2: UNSUPERVISED LEARNING (CLUSTERING) ARTIFICIAL INTELLIGENCE (CSCU9YE) LECTURE 6: MACHINE LEARNING 2: UNSUPERVISED LEARNING (CLUSTERING) Gabriela Ochoa http://www.cs.stir.ac.uk/~goc/ OUTLINE Preliminaries Classification and Clustering Applications

More information

Adaptive Online Gradient Descent

Adaptive Online Gradient Descent Adaptive Online Gradient Descent Peter L Bartlett Division of Computer Science Department of Statistics UC Berkeley Berkeley, CA 94709 bartlett@csberkeleyedu Elad Hazan IBM Almaden Research Center 650

More information

No: 10 04. Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics

No: 10 04. Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics No: 10 04 Bilkent University Monotonic Extension Farhad Husseinov Discussion Papers Department of Economics The Discussion Papers of the Department of Economics are intended to make the initial results

More information

Big Data - Lecture 1 Optimization reminders

Big Data - Lecture 1 Optimization reminders Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Schedule Introduction Major issues Examples Mathematics

More information

Solving polynomial least squares problems via semidefinite programming relaxations

Solving polynomial least squares problems via semidefinite programming relaxations Solving polynomial least squares problems via semidefinite programming relaxations Sunyoung Kim and Masakazu Kojima August 2007, revised in November, 2007 Abstract. A polynomial optimization problem whose

More information

Level Set Framework, Signed Distance Function, and Various Tools

Level Set Framework, Signed Distance Function, and Various Tools Level Set Framework Geometry and Calculus Tools Level Set Framework,, and Various Tools Spencer Department of Mathematics Brigham Young University Image Processing Seminar (Week 3), 2010 Level Set Framework

More information

Several Views of Support Vector Machines

Several Views of Support Vector Machines Several Views of Support Vector Machines Ryan M. Rifkin Honda Research Institute USA, Inc. Human Intention Understanding Group 2007 Tikhonov Regularization We are considering algorithms of the form min

More information

Numerisches Rechnen. (für Informatiker) M. Grepl J. Berger & J.T. Frings. Institut für Geometrie und Praktische Mathematik RWTH Aachen

Numerisches Rechnen. (für Informatiker) M. Grepl J. Berger & J.T. Frings. Institut für Geometrie und Praktische Mathematik RWTH Aachen (für Informatiker) M. Grepl J. Berger & J.T. Frings Institut für Geometrie und Praktische Mathematik RWTH Aachen Wintersemester 2010/11 Problem Statement Unconstrained Optimality Conditions Constrained

More information

Solving Geometric Problems with the Rotating Calipers *

Solving Geometric Problems with the Rotating Calipers * Solving Geometric Problems with the Rotating Calipers * Godfried Toussaint School of Computer Science McGill University Montreal, Quebec, Canada ABSTRACT Shamos [1] recently showed that the diameter of

More information

Linear Threshold Units

Linear Threshold Units Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear

More information

Integrating Benders decomposition within Constraint Programming

Integrating Benders decomposition within Constraint Programming Integrating Benders decomposition within Constraint Programming Hadrien Cambazard, Narendra Jussien email: {hcambaza,jussien}@emn.fr École des Mines de Nantes, LINA CNRS FRE 2729 4 rue Alfred Kastler BP

More information

A NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION

A NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION 1 A NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION Dimitri Bertsekas M.I.T. FEBRUARY 2003 2 OUTLINE Convexity issues in optimization Historical remarks Our treatment of the subject Three unifying lines of

More information

Nonlinear Optimization: Algorithms 3: Interior-point methods

Nonlinear Optimization: Algorithms 3: Interior-point methods Nonlinear Optimization: Algorithms 3: Interior-point methods INSEAD, Spring 2006 Jean-Philippe Vert Ecole des Mines de Paris Jean-Philippe.Vert@mines.org Nonlinear optimization c 2006 Jean-Philippe Vert,

More information

Walrasian Demand. u(x) where B(p, w) = {x R n + : p x w}.

Walrasian Demand. u(x) where B(p, w) = {x R n + : p x w}. Walrasian Demand Econ 2100 Fall 2015 Lecture 5, September 16 Outline 1 Walrasian Demand 2 Properties of Walrasian Demand 3 An Optimization Recipe 4 First and Second Order Conditions Definition Walrasian

More information

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem

More information

GenOpt (R) Generic Optimization Program User Manual Version 3.0.0β1

GenOpt (R) Generic Optimization Program User Manual Version 3.0.0β1 (R) User Manual Environmental Energy Technologies Division Berkeley, CA 94720 http://simulationresearch.lbl.gov Michael Wetter MWetter@lbl.gov February 20, 2009 Notice: This work was supported by the U.S.

More information

A Study of Web Log Analysis Using Clustering Techniques

A Study of Web Log Analysis Using Clustering Techniques A Study of Web Log Analysis Using Clustering Techniques Hemanshu Rana 1, Mayank Patel 2 Assistant Professor, Dept of CSE, M.G Institute of Technical Education, Gujarat India 1 Assistant Professor, Dept

More information

Branch-and-Price Approach to the Vehicle Routing Problem with Time Windows

Branch-and-Price Approach to the Vehicle Routing Problem with Time Windows TECHNISCHE UNIVERSITEIT EINDHOVEN Branch-and-Price Approach to the Vehicle Routing Problem with Time Windows Lloyd A. Fasting May 2014 Supervisors: dr. M. Firat dr.ir. M.A.A. Boon J. van Twist MSc. Contents

More information

SPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING

SPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING AAS 07-228 SPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING INTRODUCTION James G. Miller * Two historical uncorrelated track (UCT) processing approaches have been employed using general perturbations

More information

Duality of linear conic problems

Duality of linear conic problems Duality of linear conic problems Alexander Shapiro and Arkadi Nemirovski Abstract It is well known that the optimal values of a linear programming problem and its dual are equal to each other if at least

More information

The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method

The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method Robert M. Freund February, 004 004 Massachusetts Institute of Technology. 1 1 The Algorithm The problem

More information

15.062 Data Mining: Algorithms and Applications Matrix Math Review

15.062 Data Mining: Algorithms and Applications Matrix Math Review .6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop

More information

2.3 Convex Constrained Optimization Problems

2.3 Convex Constrained Optimization Problems 42 CHAPTER 2. FUNDAMENTAL CONCEPTS IN CONVEX OPTIMIZATION Theorem 15 Let f : R n R and h : R R. Consider g(x) = h(f(x)) for all x R n. The function g is convex if either of the following two conditions

More information

Date: April 12, 2001. Contents

Date: April 12, 2001. Contents 2 Lagrange Multipliers Date: April 12, 2001 Contents 2.1. Introduction to Lagrange Multipliers......... p. 2 2.2. Enhanced Fritz John Optimality Conditions...... p. 12 2.3. Informative Lagrange Multipliers...........

More information

Derivative Free Optimization

Derivative Free Optimization Department of Mathematics Derivative Free Optimization M.J.D. Powell LiTH-MAT-R--2014/02--SE Department of Mathematics Linköping University S-581 83 Linköping, Sweden. Three lectures 1 on Derivative Free

More information

t := maxγ ν subject to ν {0,1,2,...} and f(x c +γ ν d) f(x c )+cγ ν f (x c ;d).

t := maxγ ν subject to ν {0,1,2,...} and f(x c +γ ν d) f(x c )+cγ ν f (x c ;d). 1. Line Search Methods Let f : R n R be given and suppose that x c is our current best estimate of a solution to P min x R nf(x). A standard method for improving the estimate x c is to choose a direction

More information

DATA MINING CLUSTER ANALYSIS: BASIC CONCEPTS

DATA MINING CLUSTER ANALYSIS: BASIC CONCEPTS DATA MINING CLUSTER ANALYSIS: BASIC CONCEPTS 1 AND ALGORITHMS Chiara Renso KDD-LAB ISTI- CNR, Pisa, Italy WHAT IS CLUSTER ANALYSIS? Finding groups of objects such that the objects in a group will be similar

More information

Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets

Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets Macario O. Cordel II and Arnulfo P. Azcarraga College of Computer Studies *Corresponding Author: macario.cordel@dlsu.edu.ph

More information

ALMOST COMMON PRIORS 1. INTRODUCTION

ALMOST COMMON PRIORS 1. INTRODUCTION ALMOST COMMON PRIORS ZIV HELLMAN ABSTRACT. What happens when priors are not common? We introduce a measure for how far a type space is from having a common prior, which we term prior distance. If a type

More information

Robust Outlier Detection Technique in Data Mining: A Univariate Approach

Robust Outlier Detection Technique in Data Mining: A Univariate Approach Robust Outlier Detection Technique in Data Mining: A Univariate Approach Singh Vijendra and Pathak Shivani Faculty of Engineering and Technology Mody Institute of Technology and Science Lakshmangarh, Sikar,

More information

Clustering. Danilo Croce Web Mining & Retrieval a.a. 2015/201 16/03/2016

Clustering. Danilo Croce Web Mining & Retrieval a.a. 2015/201 16/03/2016 Clustering Danilo Croce Web Mining & Retrieval a.a. 2015/201 16/03/2016 1 Supervised learning vs. unsupervised learning Supervised learning: discover patterns in the data that relate data attributes with

More information

Solving Method for a Class of Bilevel Linear Programming based on Genetic Algorithms

Solving Method for a Class of Bilevel Linear Programming based on Genetic Algorithms Solving Method for a Class of Bilevel Linear Programming based on Genetic Algorithms G. Wang, Z. Wan and X. Wang Abstract The paper studies and designs an genetic algorithm (GA) of the bilevel linear programming

More information

A simplified implementation of the least squares solution for pairwise comparisons matrices

A simplified implementation of the least squares solution for pairwise comparisons matrices A simplified implementation of the least squares solution for pairwise comparisons matrices Marcin Anholcer Poznań University of Economics Al. Niepodleg lości 10, 61-875 Poznań, Poland V. Babiy McMaster

More information

Elasticity Theory Basics

Elasticity Theory Basics G22.3033-002: Topics in Computer Graphics: Lecture #7 Geometric Modeling New York University Elasticity Theory Basics Lecture #7: 20 October 2003 Lecturer: Denis Zorin Scribe: Adrian Secord, Yotam Gingold

More information

Building an Advanced Invariant Real-Time Human Tracking System

Building an Advanced Invariant Real-Time Human Tracking System UDC 004.41 Building an Advanced Invariant Real-Time Human Tracking System Fayez Idris 1, Mazen Abu_Zaher 2, Rashad J. Rasras 3, and Ibrahiem M. M. El Emary 4 1 School of Informatics and Computing, German-Jordanian

More information

Pacific Journal of Mathematics

Pacific Journal of Mathematics Pacific Journal of Mathematics GLOBAL EXISTENCE AND DECREASING PROPERTY OF BOUNDARY VALUES OF SOLUTIONS TO PARABOLIC EQUATIONS WITH NONLOCAL BOUNDARY CONDITIONS Sangwon Seo Volume 193 No. 1 March 2000

More information

Increasing for all. Convex for all. ( ) Increasing for all (remember that the log function is only defined for ). ( ) Concave for all.

Increasing for all. Convex for all. ( ) Increasing for all (remember that the log function is only defined for ). ( ) Concave for all. 1. Differentiation The first derivative of a function measures by how much changes in reaction to an infinitesimal shift in its argument. The largest the derivative (in absolute value), the faster is evolving.

More information

A hierarchical multicriteria routing model with traffic splitting for MPLS networks

A hierarchical multicriteria routing model with traffic splitting for MPLS networks A hierarchical multicriteria routing model with traffic splitting for MPLS networks João Clímaco, José Craveirinha, Marta Pascoal jclimaco@inesccpt, jcrav@deecucpt, marta@matucpt University of Coimbra

More information

K-Means Clustering Tutorial

K-Means Clustering Tutorial K-Means Clustering Tutorial By Kardi Teknomo,PhD Preferable reference for this tutorial is Teknomo, Kardi. K-Means Clustering Tutorials. http:\\people.revoledu.com\kardi\ tutorial\kmean\ Last Update: July

More information

CLUSTERING LARGE DATA SETS WITH MIXED NUMERIC AND CATEGORICAL VALUES *

CLUSTERING LARGE DATA SETS WITH MIXED NUMERIC AND CATEGORICAL VALUES * CLUSTERING LARGE DATA SETS WITH MIED NUMERIC AND CATEGORICAL VALUES * ZHEUE HUANG CSIRO Mathematical and Information Sciences GPO Box Canberra ACT, AUSTRALIA huang@cmis.csiro.au Efficient partitioning

More information

The Heat Equation. Lectures INF2320 p. 1/88

The Heat Equation. Lectures INF2320 p. 1/88 The Heat Equation Lectures INF232 p. 1/88 Lectures INF232 p. 2/88 The Heat Equation We study the heat equation: u t = u xx for x (,1), t >, (1) u(,t) = u(1,t) = for t >, (2) u(x,) = f(x) for x (,1), (3)

More information

GENERALIZED INTEGER PROGRAMMING

GENERALIZED INTEGER PROGRAMMING Professor S. S. CHADHA, PhD University of Wisconsin, Eau Claire, USA E-mail: schadha@uwec.edu Professor Veena CHADHA University of Wisconsin, Eau Claire, USA E-mail: chadhav@uwec.edu GENERALIZED INTEGER

More information

Clustering through Decision Tree Construction in Geology

Clustering through Decision Tree Construction in Geology Nonlinear Analysis: Modelling and Control, 2001, v. 6, No. 2, 29-41 Clustering through Decision Tree Construction in Geology Received: 22.10.2001 Accepted: 31.10.2001 A. Juozapavičius, V. Rapševičius Faculty

More information

Least-Squares Intersection of Lines

Least-Squares Intersection of Lines Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a

More information

Classification of Fingerprints. Sarat C. Dass Department of Statistics & Probability

Classification of Fingerprints. Sarat C. Dass Department of Statistics & Probability Classification of Fingerprints Sarat C. Dass Department of Statistics & Probability Fingerprint Classification Fingerprint classification is a coarse level partitioning of a fingerprint database into smaller

More information

SEARCH ENGINE WITH PARALLEL PROCESSING AND INCREMENTAL K-MEANS FOR FAST SEARCH AND RETRIEVAL

SEARCH ENGINE WITH PARALLEL PROCESSING AND INCREMENTAL K-MEANS FOR FAST SEARCH AND RETRIEVAL SEARCH ENGINE WITH PARALLEL PROCESSING AND INCREMENTAL K-MEANS FOR FAST SEARCH AND RETRIEVAL Krishna Kiran Kattamuri 1 and Rupa Chiramdasu 2 Department of Computer Science Engineering, VVIT, Guntur, India

More information

Mathematical finance and linear programming (optimization)

Mathematical finance and linear programming (optimization) Mathematical finance and linear programming (optimization) Geir Dahl September 15, 2009 1 Introduction The purpose of this short note is to explain how linear programming (LP) (=linear optimization) may

More information

E3: PROBABILITY AND STATISTICS lecture notes

E3: PROBABILITY AND STATISTICS lecture notes E3: PROBABILITY AND STATISTICS lecture notes 2 Contents 1 PROBABILITY THEORY 7 1.1 Experiments and random events............................ 7 1.2 Certain event. Impossible event............................

More information

Simultaneous Gamma Correction and Registration in the Frequency Domain

Simultaneous Gamma Correction and Registration in the Frequency Domain Simultaneous Gamma Correction and Registration in the Frequency Domain Alexander Wong a28wong@uwaterloo.ca William Bishop wdbishop@uwaterloo.ca Department of Electrical and Computer Engineering University

More information

Fairness in Routing and Load Balancing

Fairness in Routing and Load Balancing Fairness in Routing and Load Balancing Jon Kleinberg Yuval Rabani Éva Tardos Abstract We consider the issue of network routing subject to explicit fairness conditions. The optimization of fairness criteria

More information

THE BANACH CONTRACTION PRINCIPLE. Contents

THE BANACH CONTRACTION PRINCIPLE. Contents THE BANACH CONTRACTION PRINCIPLE ALEX PONIECKI Abstract. This paper will study contractions of metric spaces. To do this, we will mainly use tools from topology. We will give some examples of contractions,

More information

THREE DIMENSIONAL REPRESENTATION OF AMINO ACID CHARAC- TERISTICS

THREE DIMENSIONAL REPRESENTATION OF AMINO ACID CHARAC- TERISTICS THREE DIMENSIONAL REPRESENTATION OF AMINO ACID CHARAC- TERISTICS O.U. Sezerman 1, R. Islamaj 2, E. Alpaydin 2 1 Laborotory of Computational Biology, Sabancı University, Istanbul, Turkey. 2 Computer Engineering

More information

A QUICK GUIDE TO THE FORMULAS OF MULTIVARIABLE CALCULUS

A QUICK GUIDE TO THE FORMULAS OF MULTIVARIABLE CALCULUS A QUIK GUIDE TO THE FOMULAS OF MULTIVAIABLE ALULUS ontents 1. Analytic Geometry 2 1.1. Definition of a Vector 2 1.2. Scalar Product 2 1.3. Properties of the Scalar Product 2 1.4. Length and Unit Vectors

More information

An Overview of Knowledge Discovery Database and Data mining Techniques

An Overview of Knowledge Discovery Database and Data mining Techniques An Overview of Knowledge Discovery Database and Data mining Techniques Priyadharsini.C 1, Dr. Antony Selvadoss Thanamani 2 M.Phil, Department of Computer Science, NGM College, Pollachi, Coimbatore, Tamilnadu,

More information

INTEREST RATES AND FX MODELS

INTEREST RATES AND FX MODELS INTEREST RATES AND FX MODELS 8. Portfolio greeks Andrew Lesniewski Courant Institute of Mathematical Sciences New York University New York March 27, 2013 2 Interest Rates & FX Models Contents 1 Introduction

More information

Establishing a Mobile Conference Call Under Delay and Bandwidth Constraints

Establishing a Mobile Conference Call Under Delay and Bandwidth Constraints Establishing a Mobile Conference Call Under Delay and Bandwidth Constraints Amotz Bar-Noy Computer and Information Science Department Brooklyn College, CUNY, New York Email: amotz@sci.brooklyn.cuny.edu

More information

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical

More information

INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS

INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS STEVEN P. LALLEY AND ANDREW NOBEL Abstract. It is shown that there are no consistent decision rules for the hypothesis testing problem

More information

DEA implementation and clustering analysis using the K-Means algorithm

DEA implementation and clustering analysis using the K-Means algorithm Data Mining VI 321 DEA implementation and clustering analysis using the K-Means algorithm C. A. A. Lemos, M. P. E. Lins & N. F. F. Ebecken COPPE/Universidade Federal do Rio de Janeiro, Brazil Abstract

More information

Persuasion by Cheap Talk - Online Appendix

Persuasion by Cheap Talk - Online Appendix Persuasion by Cheap Talk - Online Appendix By ARCHISHMAN CHAKRABORTY AND RICK HARBAUGH Online appendix to Persuasion by Cheap Talk, American Economic Review Our results in the main text concern the case

More information

Fixed Point Theorems

Fixed Point Theorems Fixed Point Theorems Definition: Let X be a set and let T : X X be a function that maps X into itself. (Such a function is often called an operator, a transformation, or a transform on X, and the notation

More information

A Complete Gradient Clustering Algorithm for Features Analysis of X-ray Images

A Complete Gradient Clustering Algorithm for Features Analysis of X-ray Images A Complete Gradient Clustering Algorithm for Features Analysis of X-ray Images Małgorzata Charytanowicz, Jerzy Niewczas, Piotr A. Kowalski, Piotr Kulczycki, Szymon Łukasik, and Sławomir Żak Abstract Methods

More information

Numerical Analysis Lecture Notes

Numerical Analysis Lecture Notes Numerical Analysis Lecture Notes Peter J. Olver 5. Inner Products and Norms The norm of a vector is a measure of its size. Besides the familiar Euclidean norm based on the dot product, there are a number

More information

Linear Programming Notes V Problem Transformations

Linear Programming Notes V Problem Transformations Linear Programming Notes V Problem Transformations 1 Introduction Any linear programming problem can be rewritten in either of two standard forms. In the first form, the objective is to maximize, the material

More information

Optimal Scheduling for Dependent Details Processing Using MS Excel Solver

Optimal Scheduling for Dependent Details Processing Using MS Excel Solver BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 8, No 2 Sofia 2008 Optimal Scheduling for Dependent Details Processing Using MS Excel Solver Daniela Borissova Institute of

More information

Continued Fractions and the Euclidean Algorithm

Continued Fractions and the Euclidean Algorithm Continued Fractions and the Euclidean Algorithm Lecture notes prepared for MATH 326, Spring 997 Department of Mathematics and Statistics University at Albany William F Hammond Table of Contents Introduction

More information

A three point formula for finding roots of equations by the method of least squares

A three point formula for finding roots of equations by the method of least squares A three point formula for finding roots of equations by the method of least squares Ababu Teklemariam Tiruneh 1 ; William N. Ndlela 1 ; Stanley J. Nkambule 1 1 Lecturer, Department of Environmental Health

More information

(Quasi-)Newton methods

(Quasi-)Newton methods (Quasi-)Newton methods 1 Introduction 1.1 Newton method Newton method is a method to find the zeros of a differentiable non-linear function g, x such that g(x) = 0, where g : R n R n. Given a starting

More information

Randomization Approaches for Network Revenue Management with Customer Choice Behavior

Randomization Approaches for Network Revenue Management with Customer Choice Behavior Randomization Approaches for Network Revenue Management with Customer Choice Behavior Sumit Kunnumkal Indian School of Business, Gachibowli, Hyderabad, 500032, India sumit kunnumkal@isb.edu March 9, 2011

More information

Methods for Finding Bases

Methods for Finding Bases Methods for Finding Bases Bases for the subspaces of a matrix Row-reduction methods can be used to find bases. Let us now look at an example illustrating how to obtain bases for the row space, null space,

More information

Unsupervised and supervised data classification via nonsmooth and global optimization 1

Unsupervised and supervised data classification via nonsmooth and global optimization 1 Unsupervised and supervised data classification via nonsmooth and global optimization 1 A. M. Bagirov, A. M. Rubinov, N.V. Soukhoroukova and J. Yearwood School of Information Technology and Mathematical

More information

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic

More information

A Note on the Bertsimas & Sim Algorithm for Robust Combinatorial Optimization Problems

A Note on the Bertsimas & Sim Algorithm for Robust Combinatorial Optimization Problems myjournal manuscript No. (will be inserted by the editor) A Note on the Bertsimas & Sim Algorithm for Robust Combinatorial Optimization Problems Eduardo Álvarez-Miranda Ivana Ljubić Paolo Toth Received:

More information

Optimal shift scheduling with a global service level constraint

Optimal shift scheduling with a global service level constraint Optimal shift scheduling with a global service level constraint Ger Koole & Erik van der Sluis Vrije Universiteit Division of Mathematics and Computer Science De Boelelaan 1081a, 1081 HV Amsterdam The

More information

USING THE AGGLOMERATIVE METHOD OF HIERARCHICAL CLUSTERING AS A DATA MINING TOOL IN CAPITAL MARKET 1. Vera Marinova Boncheva

USING THE AGGLOMERATIVE METHOD OF HIERARCHICAL CLUSTERING AS A DATA MINING TOOL IN CAPITAL MARKET 1. Vera Marinova Boncheva 382 [7] Reznik, A, Kussul, N., Sokolov, A.: Identification of user activity using neural networks. Cybernetics and computer techniques, vol. 123 (1999) 70 79. (in Russian) [8] Kussul, N., et al. : Multi-Agent

More information

EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set

EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set Amhmed A. Bhih School of Electrical and Electronic Engineering Princy Johnson School of Electrical and Electronic Engineering Martin

More information

Applied Algorithm Design Lecture 5

Applied Algorithm Design Lecture 5 Applied Algorithm Design Lecture 5 Pietro Michiardi Eurecom Pietro Michiardi (Eurecom) Applied Algorithm Design Lecture 5 1 / 86 Approximation Algorithms Pietro Michiardi (Eurecom) Applied Algorithm Design

More information

1 VECTOR SPACES AND SUBSPACES

1 VECTOR SPACES AND SUBSPACES 1 VECTOR SPACES AND SUBSPACES What is a vector? Many are familiar with the concept of a vector as: Something which has magnitude and direction. an ordered pair or triple. a description for quantities such

More information

GLOBAL OPTIMIZATION METHOD FOR SOLVING MATHEMATICAL PROGRAMS WITH LINEAR COMPLEMENTARITY CONSTRAINTS. 1. Introduction

GLOBAL OPTIMIZATION METHOD FOR SOLVING MATHEMATICAL PROGRAMS WITH LINEAR COMPLEMENTARITY CONSTRAINTS. 1. Introduction GLOBAL OPTIMIZATION METHOD FOR SOLVING MATHEMATICAL PROGRAMS WITH LINEAR COMPLEMENTARITY CONSTRAINTS N.V. THOAI, Y. YAMAMOTO, AND A. YOSHISE Abstract. We propose a method for finding a global optimal solution

More information

1 Teaching notes on GMM 1.

1 Teaching notes on GMM 1. Bent E. Sørensen January 23, 2007 1 Teaching notes on GMM 1. Generalized Method of Moment (GMM) estimation is one of two developments in econometrics in the 80ies that revolutionized empirical work in

More information

Recovery of primal solutions from dual subgradient methods for mixed binary linear programming; a branch-and-bound approach

Recovery of primal solutions from dual subgradient methods for mixed binary linear programming; a branch-and-bound approach MASTER S THESIS Recovery of primal solutions from dual subgradient methods for mixed binary linear programming; a branch-and-bound approach PAULINE ALDENVIK MIRJAM SCHIERSCHER Department of Mathematical

More information

FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT MINING SYSTEM

FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT MINING SYSTEM International Journal of Innovative Computing, Information and Control ICIC International c 0 ISSN 34-48 Volume 8, Number 8, August 0 pp. 4 FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT

More information

A RIGOROUS AND COMPLETED STATEMENT ON HELMHOLTZ THEOREM

A RIGOROUS AND COMPLETED STATEMENT ON HELMHOLTZ THEOREM Progress In Electromagnetics Research, PIER 69, 287 304, 2007 A RIGOROU AND COMPLETED TATEMENT ON HELMHOLTZ THEOREM Y. F. Gui and W. B. Dou tate Key Lab of Millimeter Waves outheast University Nanjing,

More information

Scalar Valued Functions of Several Variables; the Gradient Vector

Scalar Valued Functions of Several Variables; the Gradient Vector Scalar Valued Functions of Several Variables; the Gradient Vector Scalar Valued Functions vector valued function of n variables: Let us consider a scalar (i.e., numerical, rather than y = φ(x = φ(x 1,

More information

A Novel Fuzzy Clustering Method for Outlier Detection in Data Mining

A Novel Fuzzy Clustering Method for Outlier Detection in Data Mining A Novel Fuzzy Clustering Method for Outlier Detection in Data Mining Binu Thomas and Rau G 2, Research Scholar, Mahatma Gandhi University,Kerala, India. binumarian@rediffmail.com 2 SCMS School of Technology

More information

Types of Degrees in Bipolar Fuzzy Graphs

Types of Degrees in Bipolar Fuzzy Graphs pplied Mathematical Sciences, Vol. 7, 2013, no. 98, 4857-4866 HIKRI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2013.37389 Types of Degrees in Bipolar Fuzzy Graphs Basheer hamed Mohideen Department

More information

Torgerson s Classical MDS derivation: 1: Determining Coordinates from Euclidean Distances

Torgerson s Classical MDS derivation: 1: Determining Coordinates from Euclidean Distances Torgerson s Classical MDS derivation: 1: Determining Coordinates from Euclidean Distances It is possible to construct a matrix X of Cartesian coordinates of points in Euclidean space when we know the Euclidean

More information

Research Article Batch Scheduling on Two-Machine Flowshop with Machine-Dependent Setup Times

Research Article Batch Scheduling on Two-Machine Flowshop with Machine-Dependent Setup Times Hindawi Publishing Corporation Advances in Operations Research Volume 2009, Article ID 153910, 10 pages doi:10.1155/2009/153910 Research Article Batch Scheduling on Two-Machine Flowshop with Machine-Dependent

More information

Data Mining Cluster Analysis: Basic Concepts and Algorithms. Lecture Notes for Chapter 8. Introduction to Data Mining

Data Mining Cluster Analysis: Basic Concepts and Algorithms. Lecture Notes for Chapter 8. Introduction to Data Mining Data Mining Cluster Analysis: Basic Concepts and Algorithms Lecture Notes for Chapter 8 by Tan, Steinbach, Kumar 1 What is Cluster Analysis? Finding groups of objects such that the objects in a group will

More information

Chapter ML:XI (continued)

Chapter ML:XI (continued) Chapter ML:XI (continued) XI. Cluster Analysis Data Mining Overview Cluster Analysis Basics Hierarchical Cluster Analysis Iterative Cluster Analysis Density-Based Cluster Analysis Cluster Evaluation Constrained

More information

Data Mining Project Report. Document Clustering. Meryem Uzun-Per

Data Mining Project Report. Document Clustering. Meryem Uzun-Per Data Mining Project Report Document Clustering Meryem Uzun-Per 504112506 Table of Content Table of Content... 2 1. Project Definition... 3 2. Literature Survey... 3 3. Methods... 4 3.1. K-means algorithm...

More information

Strategic Online Advertising: Modeling Internet User Behavior with

Strategic Online Advertising: Modeling Internet User Behavior with 2 Strategic Online Advertising: Modeling Internet User Behavior with Patrick Johnston, Nicholas Kristoff, Heather McGinness, Phuong Vu, Nathaniel Wong, Jason Wright with William T. Scherer and Matthew

More information

Mechanics 1: Conservation of Energy and Momentum

Mechanics 1: Conservation of Energy and Momentum Mechanics : Conservation of Energy and Momentum If a certain quantity associated with a system does not change in time. We say that it is conserved, and the system possesses a conservation law. Conservation

More information

Practical Guide to the Simplex Method of Linear Programming

Practical Guide to the Simplex Method of Linear Programming Practical Guide to the Simplex Method of Linear Programming Marcel Oliver Revised: April, 0 The basic steps of the simplex algorithm Step : Write the linear programming problem in standard form Linear

More information

The Advantages and Disadvantages of Online Linear Optimization

The Advantages and Disadvantages of Online Linear Optimization LINEAR PROGRAMMING WITH ONLINE LEARNING TATSIANA LEVINA, YURI LEVIN, JEFF MCGILL, AND MIKHAIL NEDIAK SCHOOL OF BUSINESS, QUEEN S UNIVERSITY, 143 UNION ST., KINGSTON, ON, K7L 3N6, CANADA E-MAIL:{TLEVIN,YLEVIN,JMCGILL,MNEDIAK}@BUSINESS.QUEENSU.CA

More information

BIG DATA PROBLEMS AND LARGE-SCALE OPTIMIZATION: A DISTRIBUTED ALGORITHM FOR MATRIX FACTORIZATION

BIG DATA PROBLEMS AND LARGE-SCALE OPTIMIZATION: A DISTRIBUTED ALGORITHM FOR MATRIX FACTORIZATION BIG DATA PROBLEMS AND LARGE-SCALE OPTIMIZATION: A DISTRIBUTED ALGORITHM FOR MATRIX FACTORIZATION Ş. İlker Birbil Sabancı University Ali Taylan Cemgil 1, Hazal Koptagel 1, Figen Öztoprak 2, Umut Şimşekli

More information