A Novel Hyperbolic Smoothing Algorithm for Clustering Large Data Sets
|
|
- Ashlee Cooper
- 7 years ago
- Views:
Transcription
1 A Novel Hyperbolic Smoothing Algorithm for Clustering Large Data Sets Abstract Adilson Elias Xavier Vinicius Layter Xavier Dept. of Systems Engineering and Computer Science Graduate School of Engineering (COPPE) Federal University of Rio de Janeiro P.O. Box Rio de Janeiro,RJ , BRAZIL The minimum sum-of-squares clustering problem is considered. The mathematical modeling of this problem leads to a min sum min formulation which, in addition to its intrinsic bi-level nature, has the significant characteristic of being strongly nondifferentiable. To overcome these difficulties, the proposed resolution method, called Hyperbolic Smoothing, adopts a smoothing strategy using a special C differentiable class function. The final solution is obtained by solving a sequence of low dimension differentiable unconstrained optimization This paper presents an extended method based upon the partition of the set of observations in two non overlapping parts. This last approach engenders a drastic simplification of the computational tasks. Keywords: Cluster Analysis, Pattern Recognition, Min-Sum- Min Problems, Nondifferentiable Programming, Smoothing 1 XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1535
2 1 Introduction Cluster analysis deals with the problems of classification of a set of patterns or observations, in general represented as points in a multidimensional space, into clusters, following two basic and simultaneous objectives: patterns in the same clusters must be similar to another (homogeneity objective) and different from patterns of other clusters (separation objective), see Hartingan (1973) and Späth (1980). A particular clustering problem formulation is considered. Among many criteria used in cluster analysis, the most natural, intuitive and frequently adopted criterion is the minimum sum-of-squares clustering (MSSC). It is a criterion for both the homogeneity and the separation objectives. The minimum sum-of-squares clustering (MSSC) formulation produces a mathematical problem of global optimization. It is both a nondifferentiable and a nonconvex mathematical problem, with a large number of local minimizers. In the cluster analysis scope, algorithms use, traditionally, two main strategies: hierarchical clustering methods and partition clustering methods (Hansen and Jaumard (1997) and Jain et alli (1999)). This paper adopts a different strategy, a smooth approach, which was originally presented in Sousa(2005) and Xavier(2008). For the sake of completeness, we present first the smoothing of the min sum min problem engendered by the modeling of the clustering problem. In a sense, the process whereby this is achieved is an extension of a smoothing scheme, called Hyperbolic Smoothing. This technique was developed through an adaptation of the hyperbolic penalty method originally introduced by Xavier (1982). By smoothing, we fundamentally mean the substitution of an intrinsically nondifferentiable two-level problem by a C unconstrained differentiable single-level alternative. Additionally, the paper presents a extended faster procedure. The basic idea is the partition of the set of observations in two non overlapping parts. By using a conceptual presentation, the first set corresponds to the observation points relatively close to two or more centroids. This set of observations, named boundary band points, can be managed by using the previously presented smoothing approach. The second set corresponds to observation points significantly closer to a single centroid in comparison with others. The set of observations, named gravitational points, is managed in a direct and simple way, offering much faster performance. For the purpose of illustrating both the reliability and the efficiency of the method, a set of computational experiments was performed, making use of traditional test problems described in the literature. XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1536
3 2 The Clustering Problem Transformation Let S = {s 1,..., s m } denote a set of m patterns or observations from an Euclidean n-space, to be clustered into a given number q of disjoint clusters. To formulate the original clustering problem as a min sum min problem, we proceed as follows. Let x i, i = 1,..., q be the centroids of the clusters, where each x i R n. The set of these centroid coordinates will be represented by X R nq. Given a point s j of S, we initially calculate the distance from s j to the center in X that is nearest. This is given by z j = min xi X s j x i 2. The most frequent measurement of the quality of a clustering associated to a specific position of q centroids is provided by the sum of the squares of these distances, which determines the following problem: subject to minimize zj 2 (1) z j = min,...,q s j x i 2, j = 1,..., m. Considering its definition, each z j must necessarily satisfy the following set of inequalities: z j s j x i 2 0, i = 1,..., q. Substituting these inequalities for the equality constraints of problem (1), a relaxed problem would be obtained. Since the variables z j are not bounded from below, the optimum solution of the relaxed problem will be z j = 0, j = 1,..., m. In order to obtain a bounded formulation, we do so by first letting ϕ(y) denote max{0, y} and then observing that, from the set of constraints in (1), it follows that ϕ(z j s j x i 2 ) = 0, j = 1,..., m. (2) By using (2) in place of the set of constraints in (1) and by including a perturbation ɛ > 0 we would obtain the bounded problem: subject to minimize zj 2 (3) ϕ(z j s j x i 2 ) ε, j = 1,..., m. Since the feasible set of problem (1) is the limit of that of (3) when ɛ 0 +, we can then consider solving (1) by solving a sequence of problems like (3) for a sequence of decreasing values for ε that approaches 0. XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1537
4 Analyzing the problem (3), the definition of function ϕ endows it with an extremely rigid nondifferentiable structure, which makes its computational solution very hard. From this perspective, for y R and τ > 0, let us define the function: φ(y, τ) = (y + y 2 + τ 2 ) /2. (4) Function φ constitutes an approximation of function ϕ. To obtain a completely differentiable problem, it is yet necessary to smooth the Euclidean distance s j x i 2. For this purpose, let us use the function θ( s j, x i, γ ) = n (s l j xl i )2 + γ 2 (5) l=1 By using function θ in place of the distance the completely differentiable problem: s j x i 2, it is obtained subject to minimize zj 2 (6) φ(z j θ(s j, x i, γ), τ) ε, j = 1,..., m. So, the properties of functions φ and θ allow us to seek a solution to problem (3) by solving a sequence of subproblems like problem (6), produced by the decreasing of the parameters γ 0, τ 0, and ε 0. On the other hand, given any set of centroids x i, i = 1,..., q, due to the increasing convex property of the hyperbolic smoothing function φ, the constraints of problem (6) will certainly be active. So, it will at last be equivalent to problem: minimize subject to h j (z j, x) = zj 2 (7) φ(z j θ(s j, x i, γ), τ) ε = 0, j = 1,..., m. The dimension of variable domain space of problem (7) is (nq + m). However, it has a separable structure, because each variable z j appears only in one equality constraint. Therefore, as the partial derivative of h(z j, x) with respect to z j, j = 1,..., m is not equal to zero, it is possible to use the XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1538
5 Implicit Function Theorem to calculate each component z j, j = 1,..., m as a function of the centroid variables x i, i = 1,..., q. In this way, the unconstrained problem minimize f(x) = z j (x) 2 (8) is obtained, where each z j (x) results from the calculation of the unique zero of each equation h j (z j, x) = φ(z j θ(s j, x i, γ), τ) ε = 0, j = 1,..., m. (9) Again, due to the Implicit Function Theorem, the functions z j (x) have all derivatives with respect to the variables x i, i = 1,..., q, and therefore it is possible to calculate the gradient of the objective function of problem (8), where f(x) = 2 z j (x) z j (x), (10) z j (x) = h j (z j, x) / h j(z j, x) z j, (11) while h j (z j, x) and h j (z j, x)/ z j are obtained from equations (4), (5) and (9). In this way, it is easy to solve problem (8) by making use of any method based on first order derivative information. At last, it must be emphasized that problem (8) is defined on a (nq) dimensional space, so it is a small problem, since the number of clusters, q, is, in general, very small for real applications. 3 The Partition Procedure The calculation of the objective function of the problem (8) demands the determination of the zeros of m equations (9), one equation for each observation point. Now, it will be presented a faster procedure. The basic idea is the partition of the set of observations in two non overlapping parts. By using a conceptual presentation, the first set corresponds to the observation points that are relatively close to two or more centroids. The second set corresponds to the observation points that are significantly close to a unique centroid in comparison with the other ones. XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1539
6 So, the first part J B is the set of boundary observations and the second is the set J G of gravitational observations. Considering this partition, equation (8) can be expressed as a sum of two independent components. minimize f(x) = z j (x) 2 = z j (x) 2 + z j (x) 2. (12) j J B j J G minimize f(x) = f 1 (x) + f 2 (x). (13) The first part of expression (13), associated with the boundary observations, can be calculated by using the previous presented smoothing approach, see (8) and (9). The second part of expression (13) can be calculated by using a faster procedure, as we will show right away. Let us define the two parts in a more rigorous form. Let be x i, i = 1,..., q be a referential position of centroids of the clusters taken in the iterative process. The boundary concept in relation to the referential point x can be easily specified by defining a δ band zone between neighboring centroids. For a generic observation s R n, we define the first and second nearest distances from s to the centroids: d 1 (s, x) = s x i1 = min i s x i and d 2 (s, x) = s x i2 = min i i1 s x i, where i 1 and i 2 are the labeling indices of these two nearest centroids. By using the above definitions, let us define precisely the δ boundary band zone: Z δ (x) = {s R n d 2 (s, x) d 1 (s, x) < δ } (14) and the gravitational region, this is the complementary space: G δ (x) = {s R n Z δ (x) }. (15) Figure 1 illustrates in R 2 the Z δ (x) and G δ (x) partitions. The central lines form the Voronoy polygon associated with the referential centroids x i, i = 1,..., q. The region between two parallel lines to Voronoy lines constitutes the boundary band zone Z δ (x). Now, the two sets can be defined in a precise form: J B (x) = {j s j Z δ (x)} and J G (x) = {j s j G δ (x)}. Proposition 1: Let s be a generic point belonging to the gravitational region G δ (x), with nearest centroid i 1. Let x be the current position of the centroids. Let x = max i x i x i be the maximum displacement of the centroids. If x < δ then s will continue to be nearest to centroid x i1 than to any other. Since δ x, Proposition 1 makes it possible to calculate exactly expression (12) in a very fast way. First, let us define the subsets of gravitational observations associated with each referential centroid: J i (x) = { } j J G min s j x l = s j x i l=1,...,q (16) XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1540
7 Figure 1: The Z δ (x) and G δ (x) partitions. The gravity centers of the observations in each non-empty subset are given by v i = 1 J i s j J i s j, i = 1,..., q. (17) Let us consider the second sum in expression (12). It will be computed by taking into account the gravity centers defined above. minimize f 2 (x) = j J G z j (x) 2 = j J i s j x i 2 = j J i s j v i + v i x i 2 = s j v i 2 + J i x i v i 2 (18) j J i When the position of centroids x i, i = 1,..., q moves during iterative process, the value of the first sum in (18) assumes a constant value, since the vectors s and v are fixed. On the other hand, for the calculation of the second sum, it is only necessary to calculate q distances, namely: v i x i, i = 1,..., q. The gradient of the second part of the objective function is easily calculated by: f 2 (x) = 2 J i ( x i v i ). (19) So, if it is observed that δ x within the iterative process, j J G z j (x) 2 and its gradient can be computed exactly by very fast procedures. XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1541
8 Simplified Extended Hyperbolic Smoothing Clustering Algorithm Initialization Step: Choose start point: x 0 ; Choose parameter values: γ 1, τ 1, ε 1 ; Choose reduction factors: 0 < ρ 1 < 1, 0 < ρ 2 < 1, 0 < ρ 3 < 1; Specify the boundary band width: δ 1 ; Let k = 1. Main Step: Repeat until a stopping rule is attained For determining the Z δ (x) and G δ (x) partitions, given by (14) and (15), use x = x k 1 and δ = δ k. Calculate the gravity centers of the observations v i, i = 1,..., q of each non-empty subset by using (17). Solve problem (13) starting at the initial point x k 1 and let x k be the solution obtained: For solving the equations (9), associated to the first part given by (8), take the smoothing parameters: γ = γ k, τ = τ k and ε = ε k ; For solving the second part, given by (18), use the above calculated gravity centers of the observations. Updating procedure: Let γ k+1 = ρ 1 γ k, τ k+1 = ρ 2 τ k, ε k+1 = ρ 3 ε k Redefine the boundary value: δ k+1. Let k := k + 1. Just as in other smoothing methods, the solution to the clustering problem is obtained, in theory, by solving an infinite sequence of optimization problems. In the Extended Hyperbolic Smoothing Clustering (XHSC) algorithm, each problem to be minimized is unconstrained and of low dimension. Notice that the algorithm causes τ and γ to approach 0, so the constraints of the subproblems as given in (6) tend to those of (3). In addition, the algorithm causes ε to approach 0, so, in a simultaneous movement, the solved problem (3) gradually approaches the original MSSC problem (1). The efficiency of the XHSC Algorithm depends strongly on the choice of boundary band width δ. A small value choice will imply an improper definition of the set G δ (x) and the frequent violation of the basic condition, x < δ, for the valid of Proposition 1. Otherwise, a big value choice will imply in a decrease of the number of observations belonging to the gravitational region and, therefore, the computational advantages given by the formulation (18) will be reduced. XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1542
9 4 Computational Results The computational results presented below were obtained from a preliminary implementation of the Extended Hyperbolic Smoothing Clustering algorithm. The numerical experiments have been carried out on a PC Intel Celeron with 2.7GHz CPU and 512MB RAM. The programs are coded with Compac Visual FORTRAN, Version 6.1. The unconstrained minimization tasks were carried out by means of a Quasi-Newton algorithm employing the BFGS updating formula from the Harwell Library, obtained in the site: ( In order to show the distinct performance of the XHSC algorithm, results obtained by solving a set of the largest problems of the TSP collection are shown below ( Table 1 presents the results for the TSPLIB-3038 data set. It exhibits the results produced by XHSC algorithm and, for comparison, those of two algorithms presented in Bagirov(2008). The first two columns show the number of clusters (q) and the best known value for the global optimum (f opt ) taken from Bagirov(2008). The next columns show the error (E) for the best solution produced (f Best ) and the mean CPU time (T ime) given in seconds associated to three algorithms: multi-start k-means (MS k-means), modified global k-means (MGKM) and the proposed XHSM. The errors are calculated in the following way: E = 100 (f Best f opt ) / f opt. The multi-start k-means algorithm is the traditional k-means algorithm with multiple initial starting points. In this experiment, to find q clusters, 100 times q starting points were randomly chosen in the MS k-means algorithm. The global k-means (GKM) algorithm, introduced by Likas et alli (2003), is a significant improvement of the k-means algorithm. The MGKS is an improved version of the GKM algorithm proposed by Bagirov(2008). The XHSM solutions were produced by using 10 starting points in all cases, except q = 40 and q = 50, where 20 and 40 starting points were taken, respectively. q f opt MS k-means MGKM XHSC Algorithm E T ime E T ime E T ime E E E E E E E E E Table 1: Results for the TSPLIB-3038 Instance XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1543
10 It is possible to observe in each row of Table 1 that the best solution produced by the new XHSC Algorithm becomes significantly smaller than that by MS k-means when the number of clusters q increases. In fact, this algorithm do not perform well for big instances, in despite being one of most used algorithms. Wu et alli (2008) present the top 10 data mining algorithms identified by the IEEE International Conference on Data Mining in December The k-means assumes the second place in this list. The comparison between XHSC and MGKM solutions demonstrates similar superiority of proposed algorithm. In the same way, the comparison of time columns shows a consistent speed advantage of proposed algorithm over the older ones. On the other hand, the best solution produced by the XHSC Algorithm is very close to the putative global minimum, the best known solution of the TSPLIB-3038 instance. Moreover, in this preliminary experiment, by using a relatively small number of initial starting points, four new putative global minimum results (q = 40, q = 60, q = 80 and q = 100) have been established. Instance q = 5 q = 10 f XHSCBest E Mean T ime Mean f XHSCBest E Mean T ime Mean FL E E FNL E E RL E E RL E E Pla E E RL E E USA E E BRD E E D E E BRD E E Pla E E Pla E E Table 2: Results for larger instances of the TSPLIB collection Table 2 presents the computational results produced by the XHSC Algorithm for the largest instances of Symmetric Traveling Salesman Problem (TSP) collection: FL3795, FNL4461, RL5915, RL5934, Pla7397, RL11849, USA13509, BRD14051, D15112, BRD18512, Pla33810 and Pla For each instance, it is presented two cases: q = 5 and q = 10. Ten different randomly chosen starting points were used. For each case, the table presents: the best objective function value produced by the XHSC Algorithm (f XHSCBest ), the average error of the 10 solutions in relation to the best solution obtained (E Mean ) and CPU mean time given in seconds (T ime Mean ). It was impossible to perform a comparison on given the lack of records of solutions of these instances. Indeed, the clustering literature seldom considers instances with such number of observations. Only a possible remark, the low XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1544
11 values presented in columns (E Mean ) show a consistent performance of the proposed algorithm. 5 Conclusions In this paper, the Extended HSC algorithm, a new method for the solution of the minimum sum-of-squares clustering problem, has been proposed. The robustness performance of the XHSC Algorithm can be attributed to the complete differentiability of the approach. The speed performance of the XHSC Algorithm can be attributed to the partition of the set of observations in two non overlapping parts. This last approach engenders a drastic simplification of computational tasks. From the results presented in Tables 1 and 2, as well other results obtained for solving classical problems of the literature (see Sousa(2005)), it was possible to observe that the implementation of the XHSC algorithm reaches the best known solution, in most of the cases, otherwise calculates a solution which is close to the best. Moreover, some new best results in the cluster literature have been established, as these shown in the present paper for the TSPLIB-3038 instance. Finally, it must be remembered that the MSSC problem is a global optimization problem with a lot of local minima, so the algorithm only can produce local minima. However, the obtained computational results presented a deep local minima property, perfectly adequate to the necessities and demands of real applications. References BAGIROV, A. M. and YEARWOOD, J. (2006) A New Nonsmooth Optimization Algorithm for Minimum Sum-of-Squares Clustering Problems, European Journal of Operational Research, 170 pp BAGIROV, A. M. (2008) Modified Global k-means Algorithm for Minimum Sum-of-Squares Clustering Problems, Pattern Recognition, Vol 41 Issue 10 pp DUBES, R. C. and JAIN, A. K. (1976) Cluster Techniques: The User s Dilemma, Pattern Recognition No. 8 pp HANSEN, P. and JAUMARD, B. (1997) Cluster Analysis and Mathematical Programming, Mathematical Programming No. 79 pp HARTINGAN, J. A. (1975) Clustering Algorithms, John Wiley and Sons, Inc., New York, NY. JAIN, A. K. and DUBES, R. C. (1988) Algorithms for Clustering Data, Prentice-Hall Inc., Upper Saddle River, NJ. XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1545
12 JAIN, A. K.; MURTY, M. N. and FLYNN, P. J. (1999) Data Clustering: A Review, ACM Computing Surveys, Vol. 31, No. 3, Sept LIKAS, A.; VLASSIS, M. and VERBEEK, J. (2003) The Global k-means Clustering Algorithm, Pattern Recognition, Vol 36 pp SOUSA, L.C.F. (2005) Desempenho Computacional do Método de Agrupamento Via Suavização Hiperbólica, M.Sc. thesis - COPPE - UFRJ, Rio de Janeiro. SPÄTH, H. (1980) Cluster Analysis Algorithms for Data Reduction and Classification, Ellis Horwood, Upper Saddle River, NJ. WU, X. et alli (2008) Top 10 Algorithms in Data Mining, Knowledge and Information Systems, Vol 14 No. 1, pp 1-37 XAVIER, A.E. (1982) Penalização Hiperbólica: Um Novo Método para Resolução de Problemas de Otimização, M.Sc. Thesis - COPPE - UFRJ, Rio de Janeiro. XAVIER, A.E. (2008) The Hyperbolic Smothing Clustering Method, submited to Pattern Recognition. XLI SBPO Pesquisa Operacional na Gestão do Conhecimento Pág. 1546
Introduction to Support Vector Machines. Colin Campbell, Bristol University
Introduction to Support Vector Machines Colin Campbell, Bristol University 1 Outline of talk. Part 1. An Introduction to SVMs 1.1. SVMs for binary classification. 1.2. Soft margins and multi-class classification.
More information24. The Branch and Bound Method
24. The Branch and Bound Method It has serious practical consequences if it is known that a combinatorial problem is NP-complete. Then one can conclude according to the present state of science that no
More informationStandardization and Its Effects on K-Means Clustering Algorithm
Research Journal of Applied Sciences, Engineering and Technology 6(7): 399-3303, 03 ISSN: 040-7459; e-issn: 040-7467 Maxwell Scientific Organization, 03 Submitted: January 3, 03 Accepted: February 5, 03
More informationSupport Vector Machines with Clustering for Training with Very Large Datasets
Support Vector Machines with Clustering for Training with Very Large Datasets Theodoros Evgeniou Technology Management INSEAD Bd de Constance, Fontainebleau 77300, France theodoros.evgeniou@insead.fr Massimiliano
More informationARTIFICIAL INTELLIGENCE (CSCU9YE) LECTURE 6: MACHINE LEARNING 2: UNSUPERVISED LEARNING (CLUSTERING)
ARTIFICIAL INTELLIGENCE (CSCU9YE) LECTURE 6: MACHINE LEARNING 2: UNSUPERVISED LEARNING (CLUSTERING) Gabriela Ochoa http://www.cs.stir.ac.uk/~goc/ OUTLINE Preliminaries Classification and Clustering Applications
More informationAdaptive Online Gradient Descent
Adaptive Online Gradient Descent Peter L Bartlett Division of Computer Science Department of Statistics UC Berkeley Berkeley, CA 94709 bartlett@csberkeleyedu Elad Hazan IBM Almaden Research Center 650
More informationNo: 10 04. Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics
No: 10 04 Bilkent University Monotonic Extension Farhad Husseinov Discussion Papers Department of Economics The Discussion Papers of the Department of Economics are intended to make the initial results
More informationBig Data - Lecture 1 Optimization reminders
Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Schedule Introduction Major issues Examples Mathematics
More informationSolving polynomial least squares problems via semidefinite programming relaxations
Solving polynomial least squares problems via semidefinite programming relaxations Sunyoung Kim and Masakazu Kojima August 2007, revised in November, 2007 Abstract. A polynomial optimization problem whose
More informationLevel Set Framework, Signed Distance Function, and Various Tools
Level Set Framework Geometry and Calculus Tools Level Set Framework,, and Various Tools Spencer Department of Mathematics Brigham Young University Image Processing Seminar (Week 3), 2010 Level Set Framework
More informationSeveral Views of Support Vector Machines
Several Views of Support Vector Machines Ryan M. Rifkin Honda Research Institute USA, Inc. Human Intention Understanding Group 2007 Tikhonov Regularization We are considering algorithms of the form min
More informationNumerisches Rechnen. (für Informatiker) M. Grepl J. Berger & J.T. Frings. Institut für Geometrie und Praktische Mathematik RWTH Aachen
(für Informatiker) M. Grepl J. Berger & J.T. Frings Institut für Geometrie und Praktische Mathematik RWTH Aachen Wintersemester 2010/11 Problem Statement Unconstrained Optimality Conditions Constrained
More informationSolving Geometric Problems with the Rotating Calipers *
Solving Geometric Problems with the Rotating Calipers * Godfried Toussaint School of Computer Science McGill University Montreal, Quebec, Canada ABSTRACT Shamos [1] recently showed that the diameter of
More informationLinear Threshold Units
Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear
More informationIntegrating Benders decomposition within Constraint Programming
Integrating Benders decomposition within Constraint Programming Hadrien Cambazard, Narendra Jussien email: {hcambaza,jussien}@emn.fr École des Mines de Nantes, LINA CNRS FRE 2729 4 rue Alfred Kastler BP
More informationA NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION
1 A NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION Dimitri Bertsekas M.I.T. FEBRUARY 2003 2 OUTLINE Convexity issues in optimization Historical remarks Our treatment of the subject Three unifying lines of
More informationNonlinear Optimization: Algorithms 3: Interior-point methods
Nonlinear Optimization: Algorithms 3: Interior-point methods INSEAD, Spring 2006 Jean-Philippe Vert Ecole des Mines de Paris Jean-Philippe.Vert@mines.org Nonlinear optimization c 2006 Jean-Philippe Vert,
More informationWalrasian Demand. u(x) where B(p, w) = {x R n + : p x w}.
Walrasian Demand Econ 2100 Fall 2015 Lecture 5, September 16 Outline 1 Walrasian Demand 2 Properties of Walrasian Demand 3 An Optimization Recipe 4 First and Second Order Conditions Definition Walrasian
More informationModern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem
More informationGenOpt (R) Generic Optimization Program User Manual Version 3.0.0β1
(R) User Manual Environmental Energy Technologies Division Berkeley, CA 94720 http://simulationresearch.lbl.gov Michael Wetter MWetter@lbl.gov February 20, 2009 Notice: This work was supported by the U.S.
More informationA Study of Web Log Analysis Using Clustering Techniques
A Study of Web Log Analysis Using Clustering Techniques Hemanshu Rana 1, Mayank Patel 2 Assistant Professor, Dept of CSE, M.G Institute of Technical Education, Gujarat India 1 Assistant Professor, Dept
More informationBranch-and-Price Approach to the Vehicle Routing Problem with Time Windows
TECHNISCHE UNIVERSITEIT EINDHOVEN Branch-and-Price Approach to the Vehicle Routing Problem with Time Windows Lloyd A. Fasting May 2014 Supervisors: dr. M. Firat dr.ir. M.A.A. Boon J. van Twist MSc. Contents
More informationSPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING
AAS 07-228 SPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING INTRODUCTION James G. Miller * Two historical uncorrelated track (UCT) processing approaches have been employed using general perturbations
More informationDuality of linear conic problems
Duality of linear conic problems Alexander Shapiro and Arkadi Nemirovski Abstract It is well known that the optimal values of a linear programming problem and its dual are equal to each other if at least
More informationThe Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method
The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method Robert M. Freund February, 004 004 Massachusetts Institute of Technology. 1 1 The Algorithm The problem
More information15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
More information2.3 Convex Constrained Optimization Problems
42 CHAPTER 2. FUNDAMENTAL CONCEPTS IN CONVEX OPTIMIZATION Theorem 15 Let f : R n R and h : R R. Consider g(x) = h(f(x)) for all x R n. The function g is convex if either of the following two conditions
More informationDate: April 12, 2001. Contents
2 Lagrange Multipliers Date: April 12, 2001 Contents 2.1. Introduction to Lagrange Multipliers......... p. 2 2.2. Enhanced Fritz John Optimality Conditions...... p. 12 2.3. Informative Lagrange Multipliers...........
More informationDerivative Free Optimization
Department of Mathematics Derivative Free Optimization M.J.D. Powell LiTH-MAT-R--2014/02--SE Department of Mathematics Linköping University S-581 83 Linköping, Sweden. Three lectures 1 on Derivative Free
More informationt := maxγ ν subject to ν {0,1,2,...} and f(x c +γ ν d) f(x c )+cγ ν f (x c ;d).
1. Line Search Methods Let f : R n R be given and suppose that x c is our current best estimate of a solution to P min x R nf(x). A standard method for improving the estimate x c is to choose a direction
More informationDATA MINING CLUSTER ANALYSIS: BASIC CONCEPTS
DATA MINING CLUSTER ANALYSIS: BASIC CONCEPTS 1 AND ALGORITHMS Chiara Renso KDD-LAB ISTI- CNR, Pisa, Italy WHAT IS CLUSTER ANALYSIS? Finding groups of objects such that the objects in a group will be similar
More informationMethodology for Emulating Self Organizing Maps for Visualization of Large Datasets
Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets Macario O. Cordel II and Arnulfo P. Azcarraga College of Computer Studies *Corresponding Author: macario.cordel@dlsu.edu.ph
More informationALMOST COMMON PRIORS 1. INTRODUCTION
ALMOST COMMON PRIORS ZIV HELLMAN ABSTRACT. What happens when priors are not common? We introduce a measure for how far a type space is from having a common prior, which we term prior distance. If a type
More informationRobust Outlier Detection Technique in Data Mining: A Univariate Approach
Robust Outlier Detection Technique in Data Mining: A Univariate Approach Singh Vijendra and Pathak Shivani Faculty of Engineering and Technology Mody Institute of Technology and Science Lakshmangarh, Sikar,
More informationClustering. Danilo Croce Web Mining & Retrieval a.a. 2015/201 16/03/2016
Clustering Danilo Croce Web Mining & Retrieval a.a. 2015/201 16/03/2016 1 Supervised learning vs. unsupervised learning Supervised learning: discover patterns in the data that relate data attributes with
More informationSolving Method for a Class of Bilevel Linear Programming based on Genetic Algorithms
Solving Method for a Class of Bilevel Linear Programming based on Genetic Algorithms G. Wang, Z. Wan and X. Wang Abstract The paper studies and designs an genetic algorithm (GA) of the bilevel linear programming
More informationA simplified implementation of the least squares solution for pairwise comparisons matrices
A simplified implementation of the least squares solution for pairwise comparisons matrices Marcin Anholcer Poznań University of Economics Al. Niepodleg lości 10, 61-875 Poznań, Poland V. Babiy McMaster
More informationElasticity Theory Basics
G22.3033-002: Topics in Computer Graphics: Lecture #7 Geometric Modeling New York University Elasticity Theory Basics Lecture #7: 20 October 2003 Lecturer: Denis Zorin Scribe: Adrian Secord, Yotam Gingold
More informationBuilding an Advanced Invariant Real-Time Human Tracking System
UDC 004.41 Building an Advanced Invariant Real-Time Human Tracking System Fayez Idris 1, Mazen Abu_Zaher 2, Rashad J. Rasras 3, and Ibrahiem M. M. El Emary 4 1 School of Informatics and Computing, German-Jordanian
More informationPacific Journal of Mathematics
Pacific Journal of Mathematics GLOBAL EXISTENCE AND DECREASING PROPERTY OF BOUNDARY VALUES OF SOLUTIONS TO PARABOLIC EQUATIONS WITH NONLOCAL BOUNDARY CONDITIONS Sangwon Seo Volume 193 No. 1 March 2000
More informationIncreasing for all. Convex for all. ( ) Increasing for all (remember that the log function is only defined for ). ( ) Concave for all.
1. Differentiation The first derivative of a function measures by how much changes in reaction to an infinitesimal shift in its argument. The largest the derivative (in absolute value), the faster is evolving.
More informationA hierarchical multicriteria routing model with traffic splitting for MPLS networks
A hierarchical multicriteria routing model with traffic splitting for MPLS networks João Clímaco, José Craveirinha, Marta Pascoal jclimaco@inesccpt, jcrav@deecucpt, marta@matucpt University of Coimbra
More informationK-Means Clustering Tutorial
K-Means Clustering Tutorial By Kardi Teknomo,PhD Preferable reference for this tutorial is Teknomo, Kardi. K-Means Clustering Tutorials. http:\\people.revoledu.com\kardi\ tutorial\kmean\ Last Update: July
More informationCLUSTERING LARGE DATA SETS WITH MIXED NUMERIC AND CATEGORICAL VALUES *
CLUSTERING LARGE DATA SETS WITH MIED NUMERIC AND CATEGORICAL VALUES * ZHEUE HUANG CSIRO Mathematical and Information Sciences GPO Box Canberra ACT, AUSTRALIA huang@cmis.csiro.au Efficient partitioning
More informationThe Heat Equation. Lectures INF2320 p. 1/88
The Heat Equation Lectures INF232 p. 1/88 Lectures INF232 p. 2/88 The Heat Equation We study the heat equation: u t = u xx for x (,1), t >, (1) u(,t) = u(1,t) = for t >, (2) u(x,) = f(x) for x (,1), (3)
More informationGENERALIZED INTEGER PROGRAMMING
Professor S. S. CHADHA, PhD University of Wisconsin, Eau Claire, USA E-mail: schadha@uwec.edu Professor Veena CHADHA University of Wisconsin, Eau Claire, USA E-mail: chadhav@uwec.edu GENERALIZED INTEGER
More informationClustering through Decision Tree Construction in Geology
Nonlinear Analysis: Modelling and Control, 2001, v. 6, No. 2, 29-41 Clustering through Decision Tree Construction in Geology Received: 22.10.2001 Accepted: 31.10.2001 A. Juozapavičius, V. Rapševičius Faculty
More informationLeast-Squares Intersection of Lines
Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a
More informationClassification of Fingerprints. Sarat C. Dass Department of Statistics & Probability
Classification of Fingerprints Sarat C. Dass Department of Statistics & Probability Fingerprint Classification Fingerprint classification is a coarse level partitioning of a fingerprint database into smaller
More informationSEARCH ENGINE WITH PARALLEL PROCESSING AND INCREMENTAL K-MEANS FOR FAST SEARCH AND RETRIEVAL
SEARCH ENGINE WITH PARALLEL PROCESSING AND INCREMENTAL K-MEANS FOR FAST SEARCH AND RETRIEVAL Krishna Kiran Kattamuri 1 and Rupa Chiramdasu 2 Department of Computer Science Engineering, VVIT, Guntur, India
More informationMathematical finance and linear programming (optimization)
Mathematical finance and linear programming (optimization) Geir Dahl September 15, 2009 1 Introduction The purpose of this short note is to explain how linear programming (LP) (=linear optimization) may
More informationE3: PROBABILITY AND STATISTICS lecture notes
E3: PROBABILITY AND STATISTICS lecture notes 2 Contents 1 PROBABILITY THEORY 7 1.1 Experiments and random events............................ 7 1.2 Certain event. Impossible event............................
More informationSimultaneous Gamma Correction and Registration in the Frequency Domain
Simultaneous Gamma Correction and Registration in the Frequency Domain Alexander Wong a28wong@uwaterloo.ca William Bishop wdbishop@uwaterloo.ca Department of Electrical and Computer Engineering University
More informationFairness in Routing and Load Balancing
Fairness in Routing and Load Balancing Jon Kleinberg Yuval Rabani Éva Tardos Abstract We consider the issue of network routing subject to explicit fairness conditions. The optimization of fairness criteria
More informationTHE BANACH CONTRACTION PRINCIPLE. Contents
THE BANACH CONTRACTION PRINCIPLE ALEX PONIECKI Abstract. This paper will study contractions of metric spaces. To do this, we will mainly use tools from topology. We will give some examples of contractions,
More informationTHREE DIMENSIONAL REPRESENTATION OF AMINO ACID CHARAC- TERISTICS
THREE DIMENSIONAL REPRESENTATION OF AMINO ACID CHARAC- TERISTICS O.U. Sezerman 1, R. Islamaj 2, E. Alpaydin 2 1 Laborotory of Computational Biology, Sabancı University, Istanbul, Turkey. 2 Computer Engineering
More informationA QUICK GUIDE TO THE FORMULAS OF MULTIVARIABLE CALCULUS
A QUIK GUIDE TO THE FOMULAS OF MULTIVAIABLE ALULUS ontents 1. Analytic Geometry 2 1.1. Definition of a Vector 2 1.2. Scalar Product 2 1.3. Properties of the Scalar Product 2 1.4. Length and Unit Vectors
More informationAn Overview of Knowledge Discovery Database and Data mining Techniques
An Overview of Knowledge Discovery Database and Data mining Techniques Priyadharsini.C 1, Dr. Antony Selvadoss Thanamani 2 M.Phil, Department of Computer Science, NGM College, Pollachi, Coimbatore, Tamilnadu,
More informationINTEREST RATES AND FX MODELS
INTEREST RATES AND FX MODELS 8. Portfolio greeks Andrew Lesniewski Courant Institute of Mathematical Sciences New York University New York March 27, 2013 2 Interest Rates & FX Models Contents 1 Introduction
More informationEstablishing a Mobile Conference Call Under Delay and Bandwidth Constraints
Establishing a Mobile Conference Call Under Delay and Bandwidth Constraints Amotz Bar-Noy Computer and Information Science Department Brooklyn College, CUNY, New York Email: amotz@sci.brooklyn.cuny.edu
More informationPATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical
More informationINDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS
INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS STEVEN P. LALLEY AND ANDREW NOBEL Abstract. It is shown that there are no consistent decision rules for the hypothesis testing problem
More informationDEA implementation and clustering analysis using the K-Means algorithm
Data Mining VI 321 DEA implementation and clustering analysis using the K-Means algorithm C. A. A. Lemos, M. P. E. Lins & N. F. F. Ebecken COPPE/Universidade Federal do Rio de Janeiro, Brazil Abstract
More informationPersuasion by Cheap Talk - Online Appendix
Persuasion by Cheap Talk - Online Appendix By ARCHISHMAN CHAKRABORTY AND RICK HARBAUGH Online appendix to Persuasion by Cheap Talk, American Economic Review Our results in the main text concern the case
More informationFixed Point Theorems
Fixed Point Theorems Definition: Let X be a set and let T : X X be a function that maps X into itself. (Such a function is often called an operator, a transformation, or a transform on X, and the notation
More informationA Complete Gradient Clustering Algorithm for Features Analysis of X-ray Images
A Complete Gradient Clustering Algorithm for Features Analysis of X-ray Images Małgorzata Charytanowicz, Jerzy Niewczas, Piotr A. Kowalski, Piotr Kulczycki, Szymon Łukasik, and Sławomir Żak Abstract Methods
More informationNumerical Analysis Lecture Notes
Numerical Analysis Lecture Notes Peter J. Olver 5. Inner Products and Norms The norm of a vector is a measure of its size. Besides the familiar Euclidean norm based on the dot product, there are a number
More informationLinear Programming Notes V Problem Transformations
Linear Programming Notes V Problem Transformations 1 Introduction Any linear programming problem can be rewritten in either of two standard forms. In the first form, the objective is to maximize, the material
More informationOptimal Scheduling for Dependent Details Processing Using MS Excel Solver
BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 8, No 2 Sofia 2008 Optimal Scheduling for Dependent Details Processing Using MS Excel Solver Daniela Borissova Institute of
More informationContinued Fractions and the Euclidean Algorithm
Continued Fractions and the Euclidean Algorithm Lecture notes prepared for MATH 326, Spring 997 Department of Mathematics and Statistics University at Albany William F Hammond Table of Contents Introduction
More informationA three point formula for finding roots of equations by the method of least squares
A three point formula for finding roots of equations by the method of least squares Ababu Teklemariam Tiruneh 1 ; William N. Ndlela 1 ; Stanley J. Nkambule 1 1 Lecturer, Department of Environmental Health
More information(Quasi-)Newton methods
(Quasi-)Newton methods 1 Introduction 1.1 Newton method Newton method is a method to find the zeros of a differentiable non-linear function g, x such that g(x) = 0, where g : R n R n. Given a starting
More informationRandomization Approaches for Network Revenue Management with Customer Choice Behavior
Randomization Approaches for Network Revenue Management with Customer Choice Behavior Sumit Kunnumkal Indian School of Business, Gachibowli, Hyderabad, 500032, India sumit kunnumkal@isb.edu March 9, 2011
More informationMethods for Finding Bases
Methods for Finding Bases Bases for the subspaces of a matrix Row-reduction methods can be used to find bases. Let us now look at an example illustrating how to obtain bases for the row space, null space,
More informationUnsupervised and supervised data classification via nonsmooth and global optimization 1
Unsupervised and supervised data classification via nonsmooth and global optimization 1 A. M. Bagirov, A. M. Rubinov, N.V. Soukhoroukova and J. Yearwood School of Information Technology and Mathematical
More informationModelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic
More informationA Note on the Bertsimas & Sim Algorithm for Robust Combinatorial Optimization Problems
myjournal manuscript No. (will be inserted by the editor) A Note on the Bertsimas & Sim Algorithm for Robust Combinatorial Optimization Problems Eduardo Álvarez-Miranda Ivana Ljubić Paolo Toth Received:
More informationOptimal shift scheduling with a global service level constraint
Optimal shift scheduling with a global service level constraint Ger Koole & Erik van der Sluis Vrije Universiteit Division of Mathematics and Computer Science De Boelelaan 1081a, 1081 HV Amsterdam The
More informationUSING THE AGGLOMERATIVE METHOD OF HIERARCHICAL CLUSTERING AS A DATA MINING TOOL IN CAPITAL MARKET 1. Vera Marinova Boncheva
382 [7] Reznik, A, Kussul, N., Sokolov, A.: Identification of user activity using neural networks. Cybernetics and computer techniques, vol. 123 (1999) 70 79. (in Russian) [8] Kussul, N., et al. : Multi-Agent
More informationEM Clustering Approach for Multi-Dimensional Analysis of Big Data Set
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set Amhmed A. Bhih School of Electrical and Electronic Engineering Princy Johnson School of Electrical and Electronic Engineering Martin
More informationApplied Algorithm Design Lecture 5
Applied Algorithm Design Lecture 5 Pietro Michiardi Eurecom Pietro Michiardi (Eurecom) Applied Algorithm Design Lecture 5 1 / 86 Approximation Algorithms Pietro Michiardi (Eurecom) Applied Algorithm Design
More information1 VECTOR SPACES AND SUBSPACES
1 VECTOR SPACES AND SUBSPACES What is a vector? Many are familiar with the concept of a vector as: Something which has magnitude and direction. an ordered pair or triple. a description for quantities such
More informationGLOBAL OPTIMIZATION METHOD FOR SOLVING MATHEMATICAL PROGRAMS WITH LINEAR COMPLEMENTARITY CONSTRAINTS. 1. Introduction
GLOBAL OPTIMIZATION METHOD FOR SOLVING MATHEMATICAL PROGRAMS WITH LINEAR COMPLEMENTARITY CONSTRAINTS N.V. THOAI, Y. YAMAMOTO, AND A. YOSHISE Abstract. We propose a method for finding a global optimal solution
More information1 Teaching notes on GMM 1.
Bent E. Sørensen January 23, 2007 1 Teaching notes on GMM 1. Generalized Method of Moment (GMM) estimation is one of two developments in econometrics in the 80ies that revolutionized empirical work in
More informationRecovery of primal solutions from dual subgradient methods for mixed binary linear programming; a branch-and-bound approach
MASTER S THESIS Recovery of primal solutions from dual subgradient methods for mixed binary linear programming; a branch-and-bound approach PAULINE ALDENVIK MIRJAM SCHIERSCHER Department of Mathematical
More informationFUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT MINING SYSTEM
International Journal of Innovative Computing, Information and Control ICIC International c 0 ISSN 34-48 Volume 8, Number 8, August 0 pp. 4 FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT
More informationA RIGOROUS AND COMPLETED STATEMENT ON HELMHOLTZ THEOREM
Progress In Electromagnetics Research, PIER 69, 287 304, 2007 A RIGOROU AND COMPLETED TATEMENT ON HELMHOLTZ THEOREM Y. F. Gui and W. B. Dou tate Key Lab of Millimeter Waves outheast University Nanjing,
More informationScalar Valued Functions of Several Variables; the Gradient Vector
Scalar Valued Functions of Several Variables; the Gradient Vector Scalar Valued Functions vector valued function of n variables: Let us consider a scalar (i.e., numerical, rather than y = φ(x = φ(x 1,
More informationA Novel Fuzzy Clustering Method for Outlier Detection in Data Mining
A Novel Fuzzy Clustering Method for Outlier Detection in Data Mining Binu Thomas and Rau G 2, Research Scholar, Mahatma Gandhi University,Kerala, India. binumarian@rediffmail.com 2 SCMS School of Technology
More informationTypes of Degrees in Bipolar Fuzzy Graphs
pplied Mathematical Sciences, Vol. 7, 2013, no. 98, 4857-4866 HIKRI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2013.37389 Types of Degrees in Bipolar Fuzzy Graphs Basheer hamed Mohideen Department
More informationTorgerson s Classical MDS derivation: 1: Determining Coordinates from Euclidean Distances
Torgerson s Classical MDS derivation: 1: Determining Coordinates from Euclidean Distances It is possible to construct a matrix X of Cartesian coordinates of points in Euclidean space when we know the Euclidean
More informationResearch Article Batch Scheduling on Two-Machine Flowshop with Machine-Dependent Setup Times
Hindawi Publishing Corporation Advances in Operations Research Volume 2009, Article ID 153910, 10 pages doi:10.1155/2009/153910 Research Article Batch Scheduling on Two-Machine Flowshop with Machine-Dependent
More informationData Mining Cluster Analysis: Basic Concepts and Algorithms. Lecture Notes for Chapter 8. Introduction to Data Mining
Data Mining Cluster Analysis: Basic Concepts and Algorithms Lecture Notes for Chapter 8 by Tan, Steinbach, Kumar 1 What is Cluster Analysis? Finding groups of objects such that the objects in a group will
More informationChapter ML:XI (continued)
Chapter ML:XI (continued) XI. Cluster Analysis Data Mining Overview Cluster Analysis Basics Hierarchical Cluster Analysis Iterative Cluster Analysis Density-Based Cluster Analysis Cluster Evaluation Constrained
More informationData Mining Project Report. Document Clustering. Meryem Uzun-Per
Data Mining Project Report Document Clustering Meryem Uzun-Per 504112506 Table of Content Table of Content... 2 1. Project Definition... 3 2. Literature Survey... 3 3. Methods... 4 3.1. K-means algorithm...
More informationStrategic Online Advertising: Modeling Internet User Behavior with
2 Strategic Online Advertising: Modeling Internet User Behavior with Patrick Johnston, Nicholas Kristoff, Heather McGinness, Phuong Vu, Nathaniel Wong, Jason Wright with William T. Scherer and Matthew
More informationMechanics 1: Conservation of Energy and Momentum
Mechanics : Conservation of Energy and Momentum If a certain quantity associated with a system does not change in time. We say that it is conserved, and the system possesses a conservation law. Conservation
More informationPractical Guide to the Simplex Method of Linear Programming
Practical Guide to the Simplex Method of Linear Programming Marcel Oliver Revised: April, 0 The basic steps of the simplex algorithm Step : Write the linear programming problem in standard form Linear
More informationThe Advantages and Disadvantages of Online Linear Optimization
LINEAR PROGRAMMING WITH ONLINE LEARNING TATSIANA LEVINA, YURI LEVIN, JEFF MCGILL, AND MIKHAIL NEDIAK SCHOOL OF BUSINESS, QUEEN S UNIVERSITY, 143 UNION ST., KINGSTON, ON, K7L 3N6, CANADA E-MAIL:{TLEVIN,YLEVIN,JMCGILL,MNEDIAK}@BUSINESS.QUEENSU.CA
More informationBIG DATA PROBLEMS AND LARGE-SCALE OPTIMIZATION: A DISTRIBUTED ALGORITHM FOR MATRIX FACTORIZATION
BIG DATA PROBLEMS AND LARGE-SCALE OPTIMIZATION: A DISTRIBUTED ALGORITHM FOR MATRIX FACTORIZATION Ş. İlker Birbil Sabancı University Ali Taylan Cemgil 1, Hazal Koptagel 1, Figen Öztoprak 2, Umut Şimşekli
More information