A Novel Hyperbolic Smoothing Algorithm for Clustering Large Data Sets



Similar documents
Introduction to Support Vector Machines. Colin Campbell, Bristol University

24. The Branch and Bound Method

Standardization and Its Effects on K-Means Clustering Algorithm

Support Vector Machines with Clustering for Training with Very Large Datasets

ARTIFICIAL INTELLIGENCE (CSCU9YE) LECTURE 6: MACHINE LEARNING 2: UNSUPERVISED LEARNING (CLUSTERING)

Adaptive Online Gradient Descent

No: Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics

Big Data - Lecture 1 Optimization reminders

Solving polynomial least squares problems via semidefinite programming relaxations

Level Set Framework, Signed Distance Function, and Various Tools

Several Views of Support Vector Machines

Numerisches Rechnen. (für Informatiker) M. Grepl J. Berger & J.T. Frings. Institut für Geometrie und Praktische Mathematik RWTH Aachen

Solving Geometric Problems with the Rotating Calipers *

Linear Threshold Units

Integrating Benders decomposition within Constraint Programming

A NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION

Nonlinear Optimization: Algorithms 3: Interior-point methods

Walrasian Demand. u(x) where B(p, w) = {x R n + : p x w}.

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh

GenOpt (R) Generic Optimization Program User Manual Version 3.0.0β1

A Study of Web Log Analysis Using Clustering Techniques

Branch-and-Price Approach to the Vehicle Routing Problem with Time Windows

SPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING

Duality of linear conic problems

The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method

Data Mining: Algorithms and Applications Matrix Math Review

2.3 Convex Constrained Optimization Problems

Date: April 12, Contents

Derivative Free Optimization

t := maxγ ν subject to ν {0,1,2,...} and f(x c +γ ν d) f(x c )+cγ ν f (x c ;d).

DATA MINING CLUSTER ANALYSIS: BASIC CONCEPTS

Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets

ALMOST COMMON PRIORS 1. INTRODUCTION

Robust Outlier Detection Technique in Data Mining: A Univariate Approach

Clustering. Danilo Croce Web Mining & Retrieval a.a. 2015/201 16/03/2016

Solving Method for a Class of Bilevel Linear Programming based on Genetic Algorithms

A simplified implementation of the least squares solution for pairwise comparisons matrices

Elasticity Theory Basics

Building an Advanced Invariant Real-Time Human Tracking System

Pacific Journal of Mathematics

Increasing for all. Convex for all. ( ) Increasing for all (remember that the log function is only defined for ). ( ) Concave for all.

A hierarchical multicriteria routing model with traffic splitting for MPLS networks

K-Means Clustering Tutorial

CLUSTERING LARGE DATA SETS WITH MIXED NUMERIC AND CATEGORICAL VALUES *

The Heat Equation. Lectures INF2320 p. 1/88

GENERALIZED INTEGER PROGRAMMING

Clustering through Decision Tree Construction in Geology

Least-Squares Intersection of Lines

Classification of Fingerprints. Sarat C. Dass Department of Statistics & Probability

SEARCH ENGINE WITH PARALLEL PROCESSING AND INCREMENTAL K-MEANS FOR FAST SEARCH AND RETRIEVAL

Mathematical finance and linear programming (optimization)

E3: PROBABILITY AND STATISTICS lecture notes

Simultaneous Gamma Correction and Registration in the Frequency Domain

Fairness in Routing and Load Balancing

THE BANACH CONTRACTION PRINCIPLE. Contents

THREE DIMENSIONAL REPRESENTATION OF AMINO ACID CHARAC- TERISTICS

A QUICK GUIDE TO THE FORMULAS OF MULTIVARIABLE CALCULUS

An Overview of Knowledge Discovery Database and Data mining Techniques

INTEREST RATES AND FX MODELS

Establishing a Mobile Conference Call Under Delay and Bandwidth Constraints

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION

INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS

DEA implementation and clustering analysis using the K-Means algorithm

Persuasion by Cheap Talk - Online Appendix

Fixed Point Theorems

A Complete Gradient Clustering Algorithm for Features Analysis of X-ray Images

Numerical Analysis Lecture Notes

Linear Programming Notes V Problem Transformations

Optimal Scheduling for Dependent Details Processing Using MS Excel Solver

Continued Fractions and the Euclidean Algorithm

A three point formula for finding roots of equations by the method of least squares

(Quasi-)Newton methods

Methods for Finding Bases

Unsupervised and supervised data classification via nonsmooth and global optimization 1

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches

A Note on the Bertsimas & Sim Algorithm for Robust Combinatorial Optimization Problems

Optimal shift scheduling with a global service level constraint

USING THE AGGLOMERATIVE METHOD OF HIERARCHICAL CLUSTERING AS A DATA MINING TOOL IN CAPITAL MARKET 1. Vera Marinova Boncheva

EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set

Applied Algorithm Design Lecture 5

1 VECTOR SPACES AND SUBSPACES

1 Teaching notes on GMM 1.

Recovery of primal solutions from dual subgradient methods for mixed binary linear programming; a branch-and-bound approach

FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT MINING SYSTEM

A RIGOROUS AND COMPLETED STATEMENT ON HELMHOLTZ THEOREM

Scalar Valued Functions of Several Variables; the Gradient Vector

A Novel Fuzzy Clustering Method for Outlier Detection in Data Mining

Types of Degrees in Bipolar Fuzzy Graphs

Torgerson s Classical MDS derivation: 1: Determining Coordinates from Euclidean Distances

Data Mining Cluster Analysis: Basic Concepts and Algorithms. Lecture Notes for Chapter 8. Introduction to Data Mining

Chapter ML:XI (continued)

Data Mining Project Report. Document Clustering. Meryem Uzun-Per

Strategic Online Advertising: Modeling Internet User Behavior with

Mechanics 1: Conservation of Energy and Momentum

Practical Guide to the Simplex Method of Linear Programming

The Advantages and Disadvantages of Online Linear Optimization

BIG DATA PROBLEMS AND LARGE-SCALE OPTIMIZATION: A DISTRIBUTED ALGORITHM FOR MATRIX FACTORIZATION

Transcription:

A Novel Hyperbolic Smoothing Algorithm for Clustering Large Data Sets Abstract Adilson Elias Xavier Vinicius Layter Xavier Dept. of Systems Engineering and Computer Science Graduate School of Engineering (COPPE) Federal University of Rio de Janeiro P.O. Box 68511 Rio de Janeiro,RJ 21941-972, BRAZIL e-mail: {adilson,vinicius}@cos.ufrj.br The minimum sum-of-squares clustering problem is considered. The mathematical modeling of this problem leads to a min sum min formulation which, in addition to its intrinsic bi-level nature, has the significant characteristic of being strongly nondifferentiable. To overcome these difficulties, the proposed resolution method, called Hyperbolic Smoothing, adopts a smoothing strategy using a special C differentiable class function. The final solution is obtained by solving a sequence of low dimension differentiable unconstrained optimization This paper presents an extended method based upon the partition of the set of observations in two non overlapping parts. This last approach engenders a drastic simplification of the computational tasks. Keywords: Cluster Analysis, Pattern Recognition, Min-Sum- Min Problems, Nondifferentiable Programming, Smoothing 1 XLI SBPO 2009 - Pesquisa Operacional na Gestão do Conhecimento Pág. 1535

1 Introduction Cluster analysis deals with the problems of classification of a set of patterns or observations, in general represented as points in a multidimensional space, into clusters, following two basic and simultaneous objectives: patterns in the same clusters must be similar to another (homogeneity objective) and different from patterns of other clusters (separation objective), see Hartingan (1973) and Späth (1980). A particular clustering problem formulation is considered. Among many criteria used in cluster analysis, the most natural, intuitive and frequently adopted criterion is the minimum sum-of-squares clustering (MSSC). It is a criterion for both the homogeneity and the separation objectives. The minimum sum-of-squares clustering (MSSC) formulation produces a mathematical problem of global optimization. It is both a nondifferentiable and a nonconvex mathematical problem, with a large number of local minimizers. In the cluster analysis scope, algorithms use, traditionally, two main strategies: hierarchical clustering methods and partition clustering methods (Hansen and Jaumard (1997) and Jain et alli (1999)). This paper adopts a different strategy, a smooth approach, which was originally presented in Sousa(2005) and Xavier(2008). For the sake of completeness, we present first the smoothing of the min sum min problem engendered by the modeling of the clustering problem. In a sense, the process whereby this is achieved is an extension of a smoothing scheme, called Hyperbolic Smoothing. This technique was developed through an adaptation of the hyperbolic penalty method originally introduced by Xavier (1982). By smoothing, we fundamentally mean the substitution of an intrinsically nondifferentiable two-level problem by a C unconstrained differentiable single-level alternative. Additionally, the paper presents a extended faster procedure. The basic idea is the partition of the set of observations in two non overlapping parts. By using a conceptual presentation, the first set corresponds to the observation points relatively close to two or more centroids. This set of observations, named boundary band points, can be managed by using the previously presented smoothing approach. The second set corresponds to observation points significantly closer to a single centroid in comparison with others. The set of observations, named gravitational points, is managed in a direct and simple way, offering much faster performance. For the purpose of illustrating both the reliability and the efficiency of the method, a set of computational experiments was performed, making use of traditional test problems described in the literature. XLI SBPO 2009 - Pesquisa Operacional na Gestão do Conhecimento Pág. 1536

2 The Clustering Problem Transformation Let S = {s 1,..., s m } denote a set of m patterns or observations from an Euclidean n-space, to be clustered into a given number q of disjoint clusters. To formulate the original clustering problem as a min sum min problem, we proceed as follows. Let x i, i = 1,..., q be the centroids of the clusters, where each x i R n. The set of these centroid coordinates will be represented by X R nq. Given a point s j of S, we initially calculate the distance from s j to the center in X that is nearest. This is given by z j = min xi X s j x i 2. The most frequent measurement of the quality of a clustering associated to a specific position of q centroids is provided by the sum of the squares of these distances, which determines the following problem: subject to minimize zj 2 (1) z j = min,...,q s j x i 2, j = 1,..., m. Considering its definition, each z j must necessarily satisfy the following set of inequalities: z j s j x i 2 0, i = 1,..., q. Substituting these inequalities for the equality constraints of problem (1), a relaxed problem would be obtained. Since the variables z j are not bounded from below, the optimum solution of the relaxed problem will be z j = 0, j = 1,..., m. In order to obtain a bounded formulation, we do so by first letting ϕ(y) denote max{0, y} and then observing that, from the set of constraints in (1), it follows that ϕ(z j s j x i 2 ) = 0, j = 1,..., m. (2) By using (2) in place of the set of constraints in (1) and by including a perturbation ɛ > 0 we would obtain the bounded problem: subject to minimize zj 2 (3) ϕ(z j s j x i 2 ) ε, j = 1,..., m. Since the feasible set of problem (1) is the limit of that of (3) when ɛ 0 +, we can then consider solving (1) by solving a sequence of problems like (3) for a sequence of decreasing values for ε that approaches 0. XLI SBPO 2009 - Pesquisa Operacional na Gestão do Conhecimento Pág. 1537

Analyzing the problem (3), the definition of function ϕ endows it with an extremely rigid nondifferentiable structure, which makes its computational solution very hard. From this perspective, for y R and τ > 0, let us define the function: φ(y, τ) = (y + y 2 + τ 2 ) /2. (4) Function φ constitutes an approximation of function ϕ. To obtain a completely differentiable problem, it is yet necessary to smooth the Euclidean distance s j x i 2. For this purpose, let us use the function θ( s j, x i, γ ) = n (s l j xl i )2 + γ 2 (5) l=1 By using function θ in place of the distance the completely differentiable problem: s j x i 2, it is obtained subject to minimize zj 2 (6) φ(z j θ(s j, x i, γ), τ) ε, j = 1,..., m. So, the properties of functions φ and θ allow us to seek a solution to problem (3) by solving a sequence of subproblems like problem (6), produced by the decreasing of the parameters γ 0, τ 0, and ε 0. On the other hand, given any set of centroids x i, i = 1,..., q, due to the increasing convex property of the hyperbolic smoothing function φ, the constraints of problem (6) will certainly be active. So, it will at last be equivalent to problem: minimize subject to h j (z j, x) = zj 2 (7) φ(z j θ(s j, x i, γ), τ) ε = 0, j = 1,..., m. The dimension of variable domain space of problem (7) is (nq + m). However, it has a separable structure, because each variable z j appears only in one equality constraint. Therefore, as the partial derivative of h(z j, x) with respect to z j, j = 1,..., m is not equal to zero, it is possible to use the XLI SBPO 2009 - Pesquisa Operacional na Gestão do Conhecimento Pág. 1538

Implicit Function Theorem to calculate each component z j, j = 1,..., m as a function of the centroid variables x i, i = 1,..., q. In this way, the unconstrained problem minimize f(x) = z j (x) 2 (8) is obtained, where each z j (x) results from the calculation of the unique zero of each equation h j (z j, x) = φ(z j θ(s j, x i, γ), τ) ε = 0, j = 1,..., m. (9) Again, due to the Implicit Function Theorem, the functions z j (x) have all derivatives with respect to the variables x i, i = 1,..., q, and therefore it is possible to calculate the gradient of the objective function of problem (8), where f(x) = 2 z j (x) z j (x), (10) z j (x) = h j (z j, x) / h j(z j, x) z j, (11) while h j (z j, x) and h j (z j, x)/ z j are obtained from equations (4), (5) and (9). In this way, it is easy to solve problem (8) by making use of any method based on first order derivative information. At last, it must be emphasized that problem (8) is defined on a (nq) dimensional space, so it is a small problem, since the number of clusters, q, is, in general, very small for real applications. 3 The Partition Procedure The calculation of the objective function of the problem (8) demands the determination of the zeros of m equations (9), one equation for each observation point. Now, it will be presented a faster procedure. The basic idea is the partition of the set of observations in two non overlapping parts. By using a conceptual presentation, the first set corresponds to the observation points that are relatively close to two or more centroids. The second set corresponds to the observation points that are significantly close to a unique centroid in comparison with the other ones. XLI SBPO 2009 - Pesquisa Operacional na Gestão do Conhecimento Pág. 1539

So, the first part J B is the set of boundary observations and the second is the set J G of gravitational observations. Considering this partition, equation (8) can be expressed as a sum of two independent components. minimize f(x) = z j (x) 2 = z j (x) 2 + z j (x) 2. (12) j J B j J G minimize f(x) = f 1 (x) + f 2 (x). (13) The first part of expression (13), associated with the boundary observations, can be calculated by using the previous presented smoothing approach, see (8) and (9). The second part of expression (13) can be calculated by using a faster procedure, as we will show right away. Let us define the two parts in a more rigorous form. Let be x i, i = 1,..., q be a referential position of centroids of the clusters taken in the iterative process. The boundary concept in relation to the referential point x can be easily specified by defining a δ band zone between neighboring centroids. For a generic observation s R n, we define the first and second nearest distances from s to the centroids: d 1 (s, x) = s x i1 = min i s x i and d 2 (s, x) = s x i2 = min i i1 s x i, where i 1 and i 2 are the labeling indices of these two nearest centroids. By using the above definitions, let us define precisely the δ boundary band zone: Z δ (x) = {s R n d 2 (s, x) d 1 (s, x) < δ } (14) and the gravitational region, this is the complementary space: G δ (x) = {s R n Z δ (x) }. (15) Figure 1 illustrates in R 2 the Z δ (x) and G δ (x) partitions. The central lines form the Voronoy polygon associated with the referential centroids x i, i = 1,..., q. The region between two parallel lines to Voronoy lines constitutes the boundary band zone Z δ (x). Now, the two sets can be defined in a precise form: J B (x) = {j s j Z δ (x)} and J G (x) = {j s j G δ (x)}. Proposition 1: Let s be a generic point belonging to the gravitational region G δ (x), with nearest centroid i 1. Let x be the current position of the centroids. Let x = max i x i x i be the maximum displacement of the centroids. If x < δ then s will continue to be nearest to centroid x i1 than to any other. Since δ x, Proposition 1 makes it possible to calculate exactly expression (12) in a very fast way. First, let us define the subsets of gravitational observations associated with each referential centroid: J i (x) = { } j J G min s j x l = s j x i l=1,...,q (16) XLI SBPO 2009 - Pesquisa Operacional na Gestão do Conhecimento Pág. 1540

Figure 1: The Z δ (x) and G δ (x) partitions. The gravity centers of the observations in each non-empty subset are given by v i = 1 J i s j J i s j, i = 1,..., q. (17) Let us consider the second sum in expression (12). It will be computed by taking into account the gravity centers defined above. minimize f 2 (x) = j J G z j (x) 2 = j J i s j x i 2 = j J i s j v i + v i x i 2 = s j v i 2 + J i x i v i 2 (18) j J i When the position of centroids x i, i = 1,..., q moves during iterative process, the value of the first sum in (18) assumes a constant value, since the vectors s and v are fixed. On the other hand, for the calculation of the second sum, it is only necessary to calculate q distances, namely: v i x i, i = 1,..., q. The gradient of the second part of the objective function is easily calculated by: f 2 (x) = 2 J i ( x i v i ). (19) So, if it is observed that δ x within the iterative process, j J G z j (x) 2 and its gradient can be computed exactly by very fast procedures. XLI SBPO 2009 - Pesquisa Operacional na Gestão do Conhecimento Pág. 1541

Simplified Extended Hyperbolic Smoothing Clustering Algorithm Initialization Step: Choose start point: x 0 ; Choose parameter values: γ 1, τ 1, ε 1 ; Choose reduction factors: 0 < ρ 1 < 1, 0 < ρ 2 < 1, 0 < ρ 3 < 1; Specify the boundary band width: δ 1 ; Let k = 1. Main Step: Repeat until a stopping rule is attained For determining the Z δ (x) and G δ (x) partitions, given by (14) and (15), use x = x k 1 and δ = δ k. Calculate the gravity centers of the observations v i, i = 1,..., q of each non-empty subset by using (17). Solve problem (13) starting at the initial point x k 1 and let x k be the solution obtained: For solving the equations (9), associated to the first part given by (8), take the smoothing parameters: γ = γ k, τ = τ k and ε = ε k ; For solving the second part, given by (18), use the above calculated gravity centers of the observations. Updating procedure: Let γ k+1 = ρ 1 γ k, τ k+1 = ρ 2 τ k, ε k+1 = ρ 3 ε k Redefine the boundary value: δ k+1. Let k := k + 1. Just as in other smoothing methods, the solution to the clustering problem is obtained, in theory, by solving an infinite sequence of optimization problems. In the Extended Hyperbolic Smoothing Clustering (XHSC) algorithm, each problem to be minimized is unconstrained and of low dimension. Notice that the algorithm causes τ and γ to approach 0, so the constraints of the subproblems as given in (6) tend to those of (3). In addition, the algorithm causes ε to approach 0, so, in a simultaneous movement, the solved problem (3) gradually approaches the original MSSC problem (1). The efficiency of the XHSC Algorithm depends strongly on the choice of boundary band width δ. A small value choice will imply an improper definition of the set G δ (x) and the frequent violation of the basic condition, x < δ, for the valid of Proposition 1. Otherwise, a big value choice will imply in a decrease of the number of observations belonging to the gravitational region and, therefore, the computational advantages given by the formulation (18) will be reduced. XLI SBPO 2009 - Pesquisa Operacional na Gestão do Conhecimento Pág. 1542

4 Computational Results The computational results presented below were obtained from a preliminary implementation of the Extended Hyperbolic Smoothing Clustering algorithm. The numerical experiments have been carried out on a PC Intel Celeron with 2.7GHz CPU and 512MB RAM. The programs are coded with Compac Visual FORTRAN, Version 6.1. The unconstrained minimization tasks were carried out by means of a Quasi-Newton algorithm employing the BFGS updating formula from the Harwell Library, obtained in the site: (http://www.cse.scitech.ac.uk/nag/hsl/). In order to show the distinct performance of the XHSC algorithm, results obtained by solving a set of the largest problems of the TSP collection are shown below (http://www.iwr.uni-heidelberg.de/groups/comopt/software). Table 1 presents the results for the TSPLIB-3038 data set. It exhibits the results produced by XHSC algorithm and, for comparison, those of two algorithms presented in Bagirov(2008). The first two columns show the number of clusters (q) and the best known value for the global optimum (f opt ) taken from Bagirov(2008). The next columns show the error (E) for the best solution produced (f Best ) and the mean CPU time (T ime) given in seconds associated to three algorithms: multi-start k-means (MS k-means), modified global k-means (MGKM) and the proposed XHSM. The errors are calculated in the following way: E = 100 (f Best f opt ) / f opt. The multi-start k-means algorithm is the traditional k-means algorithm with multiple initial starting points. In this experiment, to find q clusters, 100 times q starting points were randomly chosen in the MS k-means algorithm. The global k-means (GKM) algorithm, introduced by Likas et alli (2003), is a significant improvement of the k-means algorithm. The MGKS is an improved version of the GKM algorithm proposed by Bagirov(2008). The XHSM solutions were produced by using 10 starting points in all cases, except q = 40 and q = 50, where 20 and 40 starting points were taken, respectively. q f opt MS k-means MGKM XHSC Algorithm E T ime E T ime E T ime 2 0.31688E10 0.00 12.97 0.00 0.86 0.05 0.07 10 0.56025E09 0.00 11.52 0.58 3.30 0.01 0.28 20 0.26681E09 0.42 14.53 0.48 5.77 0.05 0.59 30 0.17557E09 1.16 19.09 0.67 8.25 0.31 0.86 40 0.12548E09 2.24 22.28 1.35 10.70-0.11 1.09 50 0.98400E08 2.60 23.55 1.41 13.23 0.44 1.36 60 0.82006E08 5.56 27.64 0.98 15.75-0.80 1.91 80 0.61217E08 4.84 30.02 0.63 20.94-0.73 6.72 100 0.48912E08 5.99 33.59 0.00 26.11-0.60 9.79 Table 1: Results for the TSPLIB-3038 Instance XLI SBPO 2009 - Pesquisa Operacional na Gestão do Conhecimento Pág. 1543

It is possible to observe in each row of Table 1 that the best solution produced by the new XHSC Algorithm becomes significantly smaller than that by MS k-means when the number of clusters q increases. In fact, this algorithm do not perform well for big instances, in despite being one of most used algorithms. Wu et alli (2008) present the top 10 data mining algorithms identified by the IEEE International Conference on Data Mining in December 2006. The k-means assumes the second place in this list. The comparison between XHSC and MGKM solutions demonstrates similar superiority of proposed algorithm. In the same way, the comparison of time columns shows a consistent speed advantage of proposed algorithm over the older ones. On the other hand, the best solution produced by the XHSC Algorithm is very close to the putative global minimum, the best known solution of the TSPLIB-3038 instance. Moreover, in this preliminary experiment, by using a relatively small number of initial starting points, four new putative global minimum results (q = 40, q = 60, q = 80 and q = 100) have been established. Instance q = 5 q = 10 f XHSCBest E Mean T ime Mean f XHSCBest E Mean T ime Mean FL3795 0.368283E09 6.18 0.18 0.106394E09 2.30 0.26 FNL4461 0.181667E10 0.43 0.31 0.853304E09 0.36 0.52 RL5915 0.379585E11 1.01 0.45 0.187794E11 0.41 0.74 RL5934 0.393650E11 1.69 0.39 0.191761E11 2.35 0.76 Pla7397 0.506247E14 1.94 0.34 0.243486E14 2.10 0.80 RL11849 0.809552E11 1.11 0.83 0.369192E11 0.53 1.55 USA13509 0.329511E14 0.01 1.01 0.149816E14 1.39 1.69 BRD14051 0.122288E11 1.20 0.82 0.593928E10 1.17 2.00 D15112 0.132707E12 0.00 0.88 0.644901E11 0.71 2.27 BRD18512 0.233416E11 1.30 1.25 0.105912E11 1.05 2.24 Pla33810 0.335680E15 0.22 3.54 0.164824E15 0.68 5.14 Pla85900 0.133972E16 1.35 7.78 0.682942E15 0.46 14.75 Table 2: Results for larger instances of the TSPLIB collection Table 2 presents the computational results produced by the XHSC Algorithm for the largest instances of Symmetric Traveling Salesman Problem (TSP) collection: FL3795, FNL4461, RL5915, RL5934, Pla7397, RL11849, USA13509, BRD14051, D15112, BRD18512, Pla33810 and Pla85900. For each instance, it is presented two cases: q = 5 and q = 10. Ten different randomly chosen starting points were used. For each case, the table presents: the best objective function value produced by the XHSC Algorithm (f XHSCBest ), the average error of the 10 solutions in relation to the best solution obtained (E Mean ) and CPU mean time given in seconds (T ime Mean ). It was impossible to perform a comparison on given the lack of records of solutions of these instances. Indeed, the clustering literature seldom considers instances with such number of observations. Only a possible remark, the low XLI SBPO 2009 - Pesquisa Operacional na Gestão do Conhecimento Pág. 1544

values presented in columns (E Mean ) show a consistent performance of the proposed algorithm. 5 Conclusions In this paper, the Extended HSC algorithm, a new method for the solution of the minimum sum-of-squares clustering problem, has been proposed. The robustness performance of the XHSC Algorithm can be attributed to the complete differentiability of the approach. The speed performance of the XHSC Algorithm can be attributed to the partition of the set of observations in two non overlapping parts. This last approach engenders a drastic simplification of computational tasks. From the results presented in Tables 1 and 2, as well other results obtained for solving classical problems of the literature (see Sousa(2005)), it was possible to observe that the implementation of the XHSC algorithm reaches the best known solution, in most of the cases, otherwise calculates a solution which is close to the best. Moreover, some new best results in the cluster literature have been established, as these shown in the present paper for the TSPLIB-3038 instance. Finally, it must be remembered that the MSSC problem is a global optimization problem with a lot of local minima, so the algorithm only can produce local minima. However, the obtained computational results presented a deep local minima property, perfectly adequate to the necessities and demands of real applications. References BAGIROV, A. M. and YEARWOOD, J. (2006) A New Nonsmooth Optimization Algorithm for Minimum Sum-of-Squares Clustering Problems, European Journal of Operational Research, 170 pp. 578-596 BAGIROV, A. M. (2008) Modified Global k-means Algorithm for Minimum Sum-of-Squares Clustering Problems, Pattern Recognition, Vol 41 Issue 10 pp 3192-3199. DUBES, R. C. and JAIN, A. K. (1976) Cluster Techniques: The User s Dilemma, Pattern Recognition No. 8 pp. 247-260. HANSEN, P. and JAUMARD, B. (1997) Cluster Analysis and Mathematical Programming, Mathematical Programming No. 79 pp. 191-215. HARTINGAN, J. A. (1975) Clustering Algorithms, John Wiley and Sons, Inc., New York, NY. JAIN, A. K. and DUBES, R. C. (1988) Algorithms for Clustering Data, Prentice-Hall Inc., Upper Saddle River, NJ. XLI SBPO 2009 - Pesquisa Operacional na Gestão do Conhecimento Pág. 1545

JAIN, A. K.; MURTY, M. N. and FLYNN, P. J. (1999) Data Clustering: A Review, ACM Computing Surveys, Vol. 31, No. 3, Sept 1999. LIKAS, A.; VLASSIS, M. and VERBEEK, J. (2003) The Global k-means Clustering Algorithm, Pattern Recognition, Vol 36 pp 451-461. SOUSA, L.C.F. (2005) Desempenho Computacional do Método de Agrupamento Via Suavização Hiperbólica, M.Sc. thesis - COPPE - UFRJ, Rio de Janeiro. SPÄTH, H. (1980) Cluster Analysis Algorithms for Data Reduction and Classification, Ellis Horwood, Upper Saddle River, NJ. WU, X. et alli (2008) Top 10 Algorithms in Data Mining, Knowledge and Information Systems, Vol 14 No. 1, pp 1-37 XAVIER, A.E. (1982) Penalização Hiperbólica: Um Novo Método para Resolução de Problemas de Otimização, M.Sc. Thesis - COPPE - UFRJ, Rio de Janeiro. XAVIER, A.E. (2008) The Hyperbolic Smothing Clustering Method, submited to Pattern Recognition. XLI SBPO 2009 - Pesquisa Operacional na Gestão do Conhecimento Pág. 1546