Clustering DHS , ,
|
|
|
- Sandra Watkins
- 9 years ago
- Views:
Transcription
1 Clustering DHS , ,
2 Clustering Definition A form of unsupervised learning, where we identify groups in feature space for an unlabeled sample set Define class regions in feature space using unlabeled data Note: the classes identified are abstract, in the sense that we obtain cluster 0... cluster n as our classes (e.g. clustering MNIST digits, we may not get 10 clusters) 2
3 Applications Clustering Applications Include: Data reduction: represent samples by their associated cluster Hypothesis generation Discover possible patterns in the data: validate on other data sets Hypothesis testing Test assumed patterns in data Prediction based on groups e.g. selecting medication for a patient using clusters of previous patients and their reactions to medication for a given disease 3
4 Kuncheva: Supervised vs. Unsupervised Classification
5 A Simple Example Assume Class Distributions Known to be Normal Can define clusters by mean and covariance matrix However... We may need more information to cluster well Many different distributions can share a mean and covariance matrix...number of clusters? 5
6 FIGURE These four data sets have identical statistics up to second-order that is, the same mean and covariance. In such cases it is important to include in the model more parameters to represent the structure more completely. From: Richard O. Duda, Peter E. Hart, and David G. Stork, Pattern Classification. Copyright c 2001 by John Wiley & Sons, Inc.
7 Steps for Clustering 1. Feature Selection Ideal: small number of features with little redundancy 2. Similarity (or Proximity) Measure Measure of similarity or dissimilarity 3. Clustering Criterion Red: defining cluster space Determine how distance patterns determine cluster likelihood (e.g. preferring circular to elongated clusters) 4. Clustering Algorithm Search method used with the clustering criterion to identify clusters 5. Validation of Results Using appropriate tests (e.g. statistical) 6. Interpretation of Results Domain expert interprets clusters (clusters are subjective) 7
8 Choosing a Similarity Measure Most Common: Euclidean Distance Roughly speaking, want distance between samples in a cluster to be smaller than the distance between samples in different clusters Example (next slide): define clusters by a maximum distance d0 between a point and a point in a cluster Rescaling features can be useful (transform the space) Unfortunately, normalizing data (e.g. by setting features to zero mean, unit variance) may eliminate subclasses One might also choose to rotate axes so they coincide with eigenvectors of the covariance matrix (i.e. apply PCA) 8
9 1 x 2 x 2 x 2 d 0 = d 0 =.1 d 0 = x 1 0 x x FIGURE The distance threshold affects the number and size of clusters in similarity based clustering methods. For three different values of distance d 0, lines are drawn between points closer than d 0 the smaller the value of d 0, the smaller and more numerous the clusters. From: Richard O. Duda, Peter E. Hart, and David G. Stork, Pattern Classification. Copyright c 2001 by John Wiley & Sons, Inc.
10 FIGURE Scaling axes affects the clusters in a minimum distance cluster method. The original data and minimum-distance clusters are shown in the upper left; points in one cluster are shown in red, while the others are shown in gray. When the vertical axis is expanded by a factor of 2.0 and the horizontal axis shrunk by a factor of 0.5, the clustering is altered (as shown at the right). Alternatively, if the vertical axis is shrunk by a factor of 0.5 and the horizontal axis is expanded by a factor of 2.0, smaller more numerous clusters result (shown at the bottom). In both these scaled cases, the assignment of points to clusters differ from that in the original space. From: Richard O. Duda, Peter 1.6 x x ( 0 2) x ( 0.5) x x 1 x 1
11 x 2 x 2 x 1 x 1 FIGURE If the data fall into well-separated clusters (left), normalization by scaling for unit variance for the full data may reduce the separation, and hence be undesirable (right). Such a normalization may in fact be appropriate if the full data set arises from a single fundamental process (with noise), but inappropriate if there are several different processes, as shown here. From: Richard O. Duda, Peter E. Hart, and David G. Stork, Pattern Classification. Copyright c 2001 by John Wiley & Sons, Inc.
12 Other Similarity Measures Minkowski Metric (Dissimilarity) Change the exponent q: q = 1: Manhattan (city-block) distance q = 2: Euclidean distance (only form invariant to translation and rotation in feature space) Cosine Similarity d(x, x )= d(x, x )= ( d ( d Characterizes similarity by the cosine of the angle between two feature vectors (in [0,1]) Ratio of inner product to vector magnitude product Invariant to rotations and dilation (not translation) 12 k=1 k=1 x k x k q ) 1/q s(x, x )= xt x x x x k x k q ) 1/q
13 More on Cosine Similarity ( d If features binary-valued: Inner product is sum of shared feature values Product of magnitudes is geometric mean of number of attributes in the two vectors Variations d(x, x )= Frequently used for Information Retrieval Ratio of shared attributes (identical lengths): s(x, x )= xt x d Tanimoto distance: ratio of shared attributes to s(x, x attributes in x or x x T x )= ( d ( d ) 1/q d(x, x )= x d(x, k x )= k q s(x, x )= k=1 k=1 x k x k q ) 1/q s(x, x )= xt x x x k=1 x k x k q ) 1/q s(x, x )= xt x s(x, x )= xt x x x x x s(x, x )= xt x x T x x T x + x T x x T x x T x + x T x x T x 13 d
14 system (e.g. we observed a decrease in security A B of 5% = ly an increase in usability of 0.3% relative to matching t only author-supplied tags). In addition, we could not te the security impact of adding title words A B using our = quencies (which are calculated over tag space, not title, and so we decided to not allow title words. they only contain 1 s and 0 s) over the union of the a i b i Thus, tagsthe in cosine similai. both videos. Therefore, the dot product simply reduces to the size intersection of the two tag sets (i.e., A t n R t ) and Adding Related the product of the magnitudes reduces to the square Tags Cosine Similarity: Tag Sets for YouTube (aroot i ) Once 2 of n 4. Return Z (b i ) the related 2 the number of tags in the first tag set times the square root of videos the number of tags in the second tag set (i.e., A Videos (Example by K. Kluever) t ilarity This techn R t ). order, we ground introdu tru Therefore, the cosine similarity between a set of author the ground tags truth. ated nthe b and a set of related Intags our case, can easily A and bebcomputed are binary as: lowed tag occurrences in a YouTube more vectors than ta g Related Videos by Cosinethey Similarity only contain 1 s and 0 s) over tag set thecould union contain of the tag 6 ct tags from Let those A and videos B that be both cos have binary θ videos. = the A t R most vectors t Therefore, similarof the the dot beproduct asame singlesimply character), reduce to the challenge length video, (represent we the performed sizeall intersection At without ex consider t R tags a sort using of t A&B) the twonumber tag sets (i.e., of related Ational t vide Rtags t ) sine similarity of the tags onthe related product videos of theand magnitudes the reduces Therefore, to the adding If square theall next roo the challenge video. The cosine the number similarity of tags metric in the isfirst tag add setup times to 6000 the square new tag ro ard similarity metric used in Information the number of Retrieval tags in the to secondbytag adding set (i.e., up all of these Tag Set Occ. Vector dog puppy funny cat to to A avoid n t bi A t A add R re text documents [20]. R t The cosine Therefore, B similarity the 1 cosine between 1 similarity 0 videos between 1 (sorted a set of inauthor decrea ctors A and B can be easily computed and a set of asrelated follows: tags can easily a challenge be computed Rejecting video as: v, a Table 1. Example of a tag occurrence table. Security a SIM(A, B) = cos θ = A B ber of related tags to g cos θ = A t R generates t up to n relate A B At the three m Consider an example where A t = {dog, puppy, funny} R t tained thro and R t = {dog, puppy, cat}. We can build a simple ta- product whichof corresponds magnitudes to are: the tag occurrence over the union tion). F i erating fu RELATEDTAGS(A, R, n t product andble of both tag sets (see Table Tag Set 1). Reading Occ. Vector row-wise dogfrom 14 n 1. Create puppy this t is a frequ Here SIM(A, B) is 2/3. an empty funny set, cat table, the tag occurrence A vectors for A B = a i b t A t and R t 1are A i 2. Sort 1= ation, after related videos 1 R 0 {1, 1, 1, 0} and B = {1, R 1, 0, 1}, respectively. t B 1 Next, we have been compute the dot product: of their tagquency sets relat gre
15 Additional Similarity Metrics Theodoridis Text Defines a large number of alternative distance metrics, including: Hamming distance: number of locations where two vectors (usually bit vectors) disagree Correlation coefficient Weighted distances... 15
16 Criterion Functions for Clustering Criterion Function Quantifies quality of a set of clusters Clustering task: partition data set D into c disjoint sets D1... Dc Choose partition maximizing the criterion function 16
17 s(x, x )= d s(x, x )= J e = x T x x T x + x T x x T x Criterion: Sum of Squared Error c x D i x µ Di 2 Measures total squared error incurred by choice of cluster centers (cluster means) Optimal Clustering Minimizes this quantity Issues Well suited when clusters compact and well-separated Different # points in each cluster can lead to large clusters being split unnaturally (next slide) Sensitive to outliers 17
18 J e = large J e = small FIGURE When two natural groupings have very different numbers of points, the clusters minimizing a sum-squared-error criterion J e of Eq. 54 may not reveal the true underlying structure. Here the criterion is smaller for the two clusters at the bottom than for the more natural clustering at the top. From: Richard O. Duda, Peter E. Hart, and David G. Stork, Pattern Classification. Copyright c 2001 by John Wiley & Sons, Inc.
19 Related Criteria: J e = Min 1 n i s Variance i 2 c J e = 1 e = 1 n i s i n 2 i s s i = 1 n 2 2 i x D i An Equivalent Formulation for SSE s i = 1 n 2 i c x D i : mean squared distance x between x 2 Jpoints e = 1 c in cluster i (variance) x D i x D i 2 J n i s i e = 1 c n i s i 2 Alternative Criterions: use median, maximum, other s i = 1 descriptive statistic on distance for n 2 J e = 1 c n s i s i = 1 x x 2 i i 2 n 2 s i = 1 x x i x D i x n D 2 i i x D i s i = 1 n 2 i x D i x x 2 s(x, x ) x D i x Variation: Using D i Similarity (e.g. Tanimoto) s s i = min s(x, x i = 1 x x n 2 ) 2 s i = 1 n 2 s i = 1 i x,x x D i D x D i i i x D i n 2 i s(x, x ) s(x, x ) s may be any similarity function (in this case, x D maximize) i x D i s i = 1 n 2 i x D i x D i s(x, x ) c x D i x x 2 x D i x D i s i = min s s(x, x ) x,x D i = min s(x, x ) i x,x D i 19 s i = min s(x, x ) x,x D i
20 Criterion: Scatter Matrix-Based S w = c trace[s w ]= x D i (x µ i )(x µ i ) T c x D i x µ i 2 = J e trace[s Minimize Trace of Sw (within-class) w ]= Equivalent to SSE! S w = c Recall that total c scatter is the sum of within S and between-class w = (x µ scatter i )(x µ (Sm i ) = T x D i Sw + Sb). This means that by c minimizing the trace of Sw, we trace[s also w maximize ]= Sb (as Sm is fixed): trace[s b ]= c x D i x µ i 2 = J e c n i µ i µ 0 2 x D i (x µ i )(x µ i ) T trace[s i ]= c x D i x µ i 2 = 20
21 trace[s w ]= x µ i 2 = J e x D i Scatter-Based trace[s b ]= Criterions, n i µ i µ 0 2 Cont d J d = S w = c Determinant Criterion c x D i (x µ i )(x µ i ) T Roughly measures square of the scattering volume; proportional to product of variances in principal axes (minimize!) Minimum error partition will not change with axis scaling, unlike SSE 21
22 Scatter-Based: Invariant Criteria c Invariant S w Criteria = c (Eigenvalue-based) S w = (x i )(x µ i ) T x D i (x µ i )(x µ i ) T x D i Eigenvalues: measure c ratio of between to withincluster trace[s scatter w ]= c trace[s in direction x of i eigenvectors 2 J e w ]= x µ i 2 = J e x D x D i (maximize!) i trace[s b ]= c Trace trace[s of a matrix b ]= is nsum i µ i of eigenvalues µ 0 2 (here d is length of feature vector) c n i µ i µ 0 2 JEigenvalues J d d = S w = are invariant under non-singular linear (x µ i µ i ) T i )(x µ i ) T transformations x D (rotations, i translations, scaling, etc.) trace[s w 1 S b ]= J f = trace[s 1 m S w ]= d λ i i d 1 1+λ i 22
23 Clustering with a Criterion Choosing Criterion Creates a well-defined problem Define clusters so as to maximize the criterion function A search problem Brute force solution: enumerate partitions of the training set, select the partition with maximum criterion value 23
24 Comparison: Scatter-Based Criteria 24
25 Hierarchical Clustering Motivation Capture similarity/distance relationships between sub-groups and samples within the chosen clusters Common in scientific taxonomies (e.g. biology) Can operate bottom up (individual samples to clusters, or agglomerative clustering) or topdown (single cluster to individual samples, or divisive clustering) 25
26 Agglomerative Hierarchical Clustering Problem: Given n samples, we want c clusters One solution: Create a sequence of partitions (clusterings) First partition, k = 1: n clusters (one cluster per sample) Second partition, k = 2: n-1 clusters Continue reducing the number of clusters by one: merge 2 closest clusters (a cluster may be a single sample) at each step k until... Goal partition: k = n - c + 1: c clusters Done; but if we re curious, we can continue on until the......final partition, k = n: one cluster Result All samples and sub-clusters organized into a tree (a dendrogram) Often show cluster similarity for a dendrogram diagram (Y-axis) If as stated above whenever two samples share a cluster they remain in a cluster at higher levels, we have a hierarchical clustering 26
27 k = 1 k = 2 k = 3 k = 4 k = 5 k = 6 k = 7 k = 8 x 1 x 2 x 3 x 4 x 5 x 6 x 7 x similarity scale FIGURE A dendrogram can represent the results of hierarchical clustering algorithms. The vertical axis shows a generalized measure of similarity among clusters. Here, at level 1 all eight points lie in singleton clusters; each point in a cluster is highly similar to itself, of course. Points x 6 and x 7 happen to be the most similar, and are merged at level 2, and so forth. From: Richard O. Duda, Peter E. Hart, and David G. Stork, Pattern Classification. Copyright c 2001 by John Wiley & Sons, Inc. x 2 4 x 6 x 3 x x x 5 x x 8 5 k = 8 FIGURE A set or Venn diagram representation of two-dimensional data (which was used in the dendrogram of Fig ) reveals the hierarchical structure but not the quantitative distances between clusters. The levels are numbered by k, in red. From: Richard O. Duda, Peter E. Hart, and David G. Stork, Pattern Classification. Copyright c 2001 by John Wiley & Sons, Inc.
28 Distance Measures d min (D i,d j ) = d max (D i,d j ) = min x x x D i,x D j max x x x D i,x D j d avg (D i,d j )= 1 n i n j x D i x D j x x Listed Above: d mean (D i,d j )= m i m j Minimum, maximum and average inter-sample distance (samples for clusters i,j: Di, Dj) Difference in cluster means (mi, mj) 28
29 Nearest-Neighbor Algorithm Also Known as Single-Linkage Algorithm d min (D i,d j ) = min x x x D i,x D j Agglomerative hierarchical clustering using dmin d max (D i,d j ) = max x x x D i,x D j Two nearest neighbors in separate clusters determine clusters merged at each step d avg (D i,d j )= 1 x x n i n j x D If we continue until k = n (c = 1), produce a minimum spanning i x tree D j (similar to Kruskal s alg.) d mean (D i,d j )= m i m j MST: Path exists between all node (sample) pairs, sum of edge costs minimum for all spanning trees Issues Sensitive to noise and slight changes in position of data points (chaining effect) Example: next slide 29
30 FIGURE Two Gaussians were used to generate two-dimensional samples, shown in pink and black. The nearest-neighbor clustering algorithm gives two clusters that well approximate the generating Gaussians (left). If, however, another particular sample is generated (circled red point at the right) and the procedure is restarted, the clusters do not well approximate the Gaussians. This illustrates how the algorithm is sensitive to the details of the samples. From: Richard O. Duda, Peter E. Hart, and David G. Stork, Pattern Classification. Copyright c 2001 by John Wiley & Sons, Inc.
31 d min (D i,d j ) = min x x x D i,x D j Farthest-Neighbor Algorithm d max (D i,d j ) = max x x x D i,x D j Agglomerative hierarchical d avg (D i,d j )= 1 clustering using dmax x x n i n j x D j Clusters with the smallest maximum distance between two x D points are merged at each step i Goal: d mean minimal (D i,d increase j )= m to largest i mcluster j diameter at each iteration (discourages elongated clusters) Known as Complete-Linkage Algorithm if terminated when distance between nearest clusters exceeds a given threshold distance Issues Works well for compact and roughly equal in size; with elongated clusters, result can be meaningless 31
32 d max = large d max = small FIGURE The farthest-neighbor clustering algorithm uses the separation between the most distant points as a criterion for cluster membership. If this distance is set very large, then all points lie in the same cluster. In the case shown at the left, a fairly large d max leads to three clusters; a smaller d max gives four clusters, as shown at the right. From: Richard O. Duda, Peter E. Hart, and David G. Stork, Pattern Classification. Copyright c 2001 by John Wiley & Sons, Inc.
33 d min (D i,d j ) = min x x x D i,x D j Using Mean, Avg Distances d max (D i,d j ) = max x x x D i,x D j d avg (D i,d j )= 1 n i n j x D i x D j x x d mean (D i,d j )= m i m j Reduces Sensitivity to Outliers Mean less expensive to compute than avg, min, max (each require ni * nj distances) 33
34 Stepwise Optimal Hierarchical Clustering Problem None of the agglomerative methods discussed so far directly minimize a specific criterion function Modified Agglomerative Algorithm: For k = 1 to (n - c + 1) d min (D i,d j ) = d max (D i,d j ) = d avg (D i,d j )= 1 n i n j min x x x D i,x D j max x x x D i,x D j Find clusters whose merger changes criterion least, Di and Dj d mean (D i,d j )= m i m j x D i x D j x x Merge Di and Dj Example: Minimal increase in SSE (Je) d e (D i,d j )= ni n j n i + n j m i m j de defines the cluster pair that increases Je as little as possible. May not minimize SSE, but often good starting point prefers merging single elements or small with large clusters vs. merging medium-size clusters 34
35 k-means Clustering k-means Algorithm For a number of clusters k: 1. Choose k data points at random 2. Assign all data points to closest of the k cluster centers 3. Re-compute k cluster centers as the mean vector of each cluster If cluster centers do not change, stop Else, goto 2 Complexity O(ndcT) - T iterations, d features, n points, c clusters, in practice usually T << n (much fewer than n iterations) Note: means tend to move minimizing squared error criterion 35
36 µ µ FIGURE The k-means clustering procedure is a form of stochastic hill climbing in the log-likelihood function. The contours represent equal log-likelihood values for the one-dimensional data in Fig The dots indicate parameter values after different iterations of the k-means algorithm. Six of the starting points shown lead to local maxima, whereas two (i.e., µ 1 (0) = µ 2 (0)) lead to a saddle point near = 0. From: Richard O. Duda, Peter E. Hart, and David G. Stork, Pattern Classification. Copyright c 2001 by John Wiley & Sons, Inc.
37 p(x µ b ) p(x µ a ) source density x l(µ 1, µ 2 ) µ µ a start µ b start µ 1 FIGURE (Above) The source mixture density used to generate sample data, and two maximum-likelihood estimates based on the data in the table. (Bottom) Loglikelihood of a mixture model consisting of two univariate Gaussians as a function of their means, for the data in the table. Trajectories for the iterative maximum-likelihood estimation of the means of a two-gaussian mixture model based on the data are shown as red lines. Two local optima (with log-likelihoods 52.2 and 56.7) correspond to the two density estimates shown above. From: Richard O. Duda, Peter E. Hart, and David G. Stork, Pattern Classification. Copyright c 2001 by John Wiley & Sons, Inc.
38 x FIGURE Trajectories for the means of the k-means clustering procedure applied to two-dimensional data. The final Voronoi tesselation (for classification) is also shown the means correspond to the centers of the Voronoi cells. In this case, convergence is obtained in three iterations. From: Richard O. Duda, Peter E. Hart, and David G. Stork, Pattern Classification. Copyright c 2001 by John Wiley & Sons, Inc. x 1
39 Basic Idea Fuzzy k-means Allow every point to have a probability of membership in every cluster. The criterion (cost function) minimized is: J fuz = c n j=1 [ ˆP (ω i x j, ˆΘ)] b x j µ i 2 Theta is the membership function parameter set. b ( blending ) is a free parameter: b = 0: Sum of squared error criterion (one cluster per data point) b > 1: each pattern may belong to multiple clusters 39
40 J fuz = c Fuzzy k-mean Clustering µ j = Algorithm n j=1 [ ˆP (ω i x j, ˆΘ)] b x j µ i 2 J fuz = [ ˆP (ω i x j, ˆΘ)] b x j µ i 2 j=1 nj=1 [ ˆP (ω i x j )] b x j nj=1 [ ˆP (ω i x j )] b µ j = nj=1 [ ˆP (ω i x j )] b x j nj=1 [ ˆP (ω i x j )] b ˆP (ω i x j )= (1/d ij) 1/(b 1) cr=1 (1/d rj ) 1/(b 1) Algorithm d ij = x j µ i 2 1. Compute probability of each class for every point in the training set (uniform probability: equal likelihood in each cluster) 2. Recompute means using expression at top-left 3. Recompute probability of each class for each point using expression at top right If change in means and membership probabilities for points is small, stop Else goto 2 40
41 x FIGURE At each iteration of the fuzzy k-means clustering algorithm, the probability of category memberships for each point are adjusted according to Eqs. 32 and 33 (here b = 2). While most points have nonnegligible memberships in two or three clusters, we nevertheless draw the boundary of a Voronoi tesselation to illustrate the progress of the algorithm. After four iterations, the algorithm has converged to the red cluster centers and associated Voronoi tesselation. From: Richard O. Duda, Peter E. Hart, and David G. Stork, Pattern Classification. Copyright c 2001 by John Wiley & Sons, Inc. x 1
42 Fuzzy k-means, Cont d Convergence Properties Sometimes fuzzy k-means improves convergence over classical k-means However, probability of cluster membership depends on the number of clusters; can lead to problems if poor choice of k is made 42
43 Cluster Validity So far... We ve assumed that we know the number of clusters When number of clusters isn t known We can try a clustering procedure using c=1, c=2, etc., and making note of sudden decreases in the error criterion (e.g. SSE) More formal: statistical tests, however problem of testing cluster validity is unsolved DHS: Section presents a statistical test centered around testing the null hypothesis of having c clusters, by comparing with c+1 clusters 43
DATA MINING CLUSTER ANALYSIS: BASIC CONCEPTS
DATA MINING CLUSTER ANALYSIS: BASIC CONCEPTS 1 AND ALGORITHMS Chiara Renso KDD-LAB ISTI- CNR, Pisa, Italy WHAT IS CLUSTER ANALYSIS? Finding groups of objects such that the objects in a group will be similar
Neural Networks Lesson 5 - Cluster Analysis
Neural Networks Lesson 5 - Cluster Analysis Prof. Michele Scarpiniti INFOCOM Dpt. - Sapienza University of Rome http://ispac.ing.uniroma1.it/scarpiniti/index.htm [email protected] Rome, 29
Clustering. Danilo Croce Web Mining & Retrieval a.a. 2015/201 16/03/2016
Clustering Danilo Croce Web Mining & Retrieval a.a. 2015/201 16/03/2016 1 Supervised learning vs. unsupervised learning Supervised learning: discover patterns in the data that relate data attributes with
Unsupervised learning: Clustering
Unsupervised learning: Clustering Salissou Moutari Centre for Statistical Science and Operational Research CenSSOR 17 th September 2013 Unsupervised learning: Clustering 1/52 Outline 1 Introduction What
Medical Information Management & Mining. You Chen Jan,15, 2013 [email protected]
Medical Information Management & Mining You Chen Jan,15, 2013 [email protected] 1 Trees Building Materials Trees cannot be used to build a house directly. How can we transform trees to building materials?
Cluster Analysis. Isabel M. Rodrigues. Lisboa, 2014. Instituto Superior Técnico
Instituto Superior Técnico Lisboa, 2014 Introduction: Cluster analysis What is? Finding groups of objects such that the objects in a group will be similar (or related) to one another and different from
SPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING
AAS 07-228 SPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING INTRODUCTION James G. Miller * Two historical uncorrelated track (UCT) processing approaches have been employed using general perturbations
Clustering. Adrian Groza. Department of Computer Science Technical University of Cluj-Napoca
Clustering Adrian Groza Department of Computer Science Technical University of Cluj-Napoca Outline 1 Cluster Analysis What is Datamining? Cluster Analysis 2 K-means 3 Hierarchical Clustering What is Datamining?
How To Cluster
Data Clustering Dec 2nd, 2013 Kyrylo Bessonov Talk outline Introduction to clustering Types of clustering Supervised Unsupervised Similarity measures Main clustering algorithms k-means Hierarchical Main
Clustering UE 141 Spring 2013
Clustering UE 141 Spring 013 Jing Gao SUNY Buffalo 1 Definition of Clustering Finding groups of obects such that the obects in a group will be similar (or related) to one another and different from (or
. Learn the number of classes and the structure of each class using similarity between unlabeled training patterns
Outline Part 1: of data clustering Non-Supervised Learning and Clustering : Problem formulation cluster analysis : Taxonomies of Clustering Techniques : Data types and Proximity Measures : Difficulties
Clustering. Data Mining. Abraham Otero. Data Mining. Agenda
Clustering 1/46 Agenda Introduction Distance K-nearest neighbors Hierarchical clustering Quick reference 2/46 1 Introduction It seems logical that in a new situation we should act in a similar way as in
Chapter ML:XI (continued)
Chapter ML:XI (continued) XI. Cluster Analysis Data Mining Overview Cluster Analysis Basics Hierarchical Cluster Analysis Iterative Cluster Analysis Density-Based Cluster Analysis Cluster Evaluation Constrained
Cluster Analysis: Basic Concepts and Algorithms
8 Cluster Analysis: Basic Concepts and Algorithms Cluster analysis divides data into groups (clusters) that are meaningful, useful, or both. If meaningful groups are the goal, then the clusters should
Chapter 7. Cluster Analysis
Chapter 7. Cluster Analysis. What is Cluster Analysis?. A Categorization of Major Clustering Methods. Partitioning Methods. Hierarchical Methods 5. Density-Based Methods 6. Grid-Based Methods 7. Model-Based
Linear Threshold Units
Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear
Clustering. 15-381 Artificial Intelligence Henry Lin. Organizing data into clusters such that there is
Clustering 15-381 Artificial Intelligence Henry Lin Modified from excellent slides of Eamonn Keogh, Ziv Bar-Joseph, and Andrew Moore What is Clustering? Organizing data into clusters such that there is
Data Mining. Cluster Analysis: Advanced Concepts and Algorithms
Data Mining Cluster Analysis: Advanced Concepts and Algorithms Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 1 More Clustering Methods Prototype-based clustering Density-based clustering Graph-based
Cluster Analysis: Advanced Concepts
Cluster Analysis: Advanced Concepts and dalgorithms Dr. Hui Xiong Rutgers University Introduction to Data Mining 08/06/2006 1 Introduction to Data Mining 08/06/2006 1 Outline Prototype-based Fuzzy c-means
Data Mining Clustering (2) Sheets are based on the those provided by Tan, Steinbach, and Kumar. Introduction to Data Mining
Data Mining Clustering (2) Toon Calders Sheets are based on the those provided by Tan, Steinbach, and Kumar. Introduction to Data Mining Outline Partitional Clustering Distance-based K-means, K-medoids,
Data Mining Cluster Analysis: Basic Concepts and Algorithms. Lecture Notes for Chapter 8. Introduction to Data Mining
Data Mining Cluster Analysis: Basic Concepts and Algorithms Lecture Notes for Chapter 8 Introduction to Data Mining by Tan, Steinbach, Kumar Tan,Steinbach, Kumar Introduction to Data Mining 4/8/2004 Hierarchical
Information Retrieval and Web Search Engines
Information Retrieval and Web Search Engines Lecture 7: Document Clustering December 10 th, 2013 Wolf-Tilo Balke and Kinda El Maarry Institut für Informationssysteme Technische Universität Braunschweig
How To Solve The Cluster Algorithm
Cluster Algorithms Adriano Cruz [email protected] 28 de outubro de 2013 Adriano Cruz [email protected] () Cluster Algorithms 28 de outubro de 2013 1 / 80 Summary 1 K-Means Adriano Cruz [email protected]
Social Media Mining. Data Mining Essentials
Introduction Data production rate has been increased dramatically (Big Data) and we are able store much more data than before E.g., purchase data, social media data, mobile phone data Businesses and customers
Standardization and Its Effects on K-Means Clustering Algorithm
Research Journal of Applied Sciences, Engineering and Technology 6(7): 399-3303, 03 ISSN: 040-7459; e-issn: 040-7467 Maxwell Scientific Organization, 03 Submitted: January 3, 03 Accepted: February 5, 03
An Introduction to Cluster Analysis for Data Mining
An Introduction to Cluster Analysis for Data Mining 10/02/2000 11:42 AM 1. INTRODUCTION... 4 1.1. Scope of This Paper... 4 1.2. What Cluster Analysis Is... 4 1.3. What Cluster Analysis Is Not... 5 2. OVERVIEW...
Data Mining Cluster Analysis: Basic Concepts and Algorithms. Lecture Notes for Chapter 8. Introduction to Data Mining
Data Mining Cluster Analysis: Basic Concepts and Algorithms Lecture Notes for Chapter 8 by Tan, Steinbach, Kumar 1 What is Cluster Analysis? Finding groups of objects such that the objects in a group will
Clustering. Chapter 7. 7.1 Introduction to Clustering Techniques. 7.1.1 Points, Spaces, and Distances
240 Chapter 7 Clustering Clustering is the process of examining a collection of points, and grouping the points into clusters according to some distance measure. The goal is that points in the same cluster
Segmentation & Clustering
EECS 442 Computer vision Segmentation & Clustering Segmentation in human vision K-mean clustering Mean-shift Graph-cut Reading: Chapters 14 [FP] Some slides of this lectures are courtesy of prof F. Li,
Environmental Remote Sensing GEOG 2021
Environmental Remote Sensing GEOG 2021 Lecture 4 Image classification 2 Purpose categorising data data abstraction / simplification data interpretation mapping for land cover mapping use land cover class
Data Mining Cluster Analysis: Basic Concepts and Algorithms. Lecture Notes for Chapter 8. Introduction to Data Mining
Data Mining Cluster Analsis: Basic Concepts and Algorithms Lecture Notes for Chapter 8 Introduction to Data Mining b Tan, Steinbach, Kumar Tan,Steinbach, Kumar Introduction to Data Mining /8/ What is Cluster
Statistical Databases and Registers with some datamining
Unsupervised learning - Statistical Databases and Registers with some datamining a course in Survey Methodology and O cial Statistics Pages in the book: 501-528 Department of Statistics Stockholm University
Data Mining Project Report. Document Clustering. Meryem Uzun-Per
Data Mining Project Report Document Clustering Meryem Uzun-Per 504112506 Table of Content Table of Content... 2 1. Project Definition... 3 2. Literature Survey... 3 3. Methods... 4 3.1. K-means algorithm...
ARTIFICIAL INTELLIGENCE (CSCU9YE) LECTURE 6: MACHINE LEARNING 2: UNSUPERVISED LEARNING (CLUSTERING)
ARTIFICIAL INTELLIGENCE (CSCU9YE) LECTURE 6: MACHINE LEARNING 2: UNSUPERVISED LEARNING (CLUSTERING) Gabriela Ochoa http://www.cs.stir.ac.uk/~goc/ OUTLINE Preliminaries Classification and Clustering Applications
USING SPECTRAL RADIUS RATIO FOR NODE DEGREE TO ANALYZE THE EVOLUTION OF SCALE- FREE NETWORKS AND SMALL-WORLD NETWORKS
USING SPECTRAL RADIUS RATIO FOR NODE DEGREE TO ANALYZE THE EVOLUTION OF SCALE- FREE NETWORKS AND SMALL-WORLD NETWORKS Natarajan Meghanathan Jackson State University, 1400 Lynch St, Jackson, MS, USA [email protected]
Using Data Mining for Mobile Communication Clustering and Characterization
Using Data Mining for Mobile Communication Clustering and Characterization A. Bascacov *, C. Cernazanu ** and M. Marcu ** * Lasting Software, Timisoara, Romania ** Politehnica University of Timisoara/Computer
Statistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
Data Mining Cluster Analysis: Basic Concepts and Algorithms. Clustering Algorithms. Lecture Notes for Chapter 8. Introduction to Data Mining
Data Mining Cluster Analsis: Basic Concepts and Algorithms Lecture Notes for Chapter 8 Introduction to Data Mining b Tan, Steinbach, Kumar Clustering Algorithms K-means and its variants Hierarchical clustering
Distances, Clustering, and Classification. Heatmaps
Distances, Clustering, and Classification Heatmaps 1 Distance Clustering organizes things that are close into groups What does it mean for two genes to be close? What does it mean for two samples to be
K-Means Cluster Analysis. Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 1
K-Means Cluster Analsis Chapter 3 PPDM Class Tan,Steinbach, Kumar Introduction to Data Mining 4/18/4 1 What is Cluster Analsis? Finding groups of objects such that the objects in a group will be similar
Cluster analysis Cosmin Lazar. COMO Lab VUB
Cluster analysis Cosmin Lazar COMO Lab VUB Introduction Cluster analysis foundations rely on one of the most fundamental, simple and very often unnoticed ways (or methods) of understanding and learning,
STATISTICA Formula Guide: Logistic Regression. Table of Contents
: Table of Contents... 1 Overview of Model... 1 Dispersion... 2 Parameterization... 3 Sigma-Restricted Model... 3 Overparameterized Model... 4 Reference Coding... 4 Model Summary (Summary Tab)... 5 Summary
Hierarchical Cluster Analysis Some Basics and Algorithms
Hierarchical Cluster Analysis Some Basics and Algorithms Nethra Sambamoorthi CRMportals Inc., 11 Bartram Road, Englishtown, NJ 07726 (NOTE: Please use always the latest copy of the document. Click on this
Lecture 3: Linear methods for classification
Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,
Solving Simultaneous Equations and Matrices
Solving Simultaneous Equations and Matrices The following represents a systematic investigation for the steps used to solve two simultaneous linear equations in two unknowns. The motivation for considering
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical
The SPSS TwoStep Cluster Component
White paper technical report The SPSS TwoStep Cluster Component A scalable component enabling more efficient customer segmentation Introduction The SPSS TwoStep Clustering Component is a scalable cluster
STATISTICA. Clustering Techniques. Case Study: Defining Clusters of Shopping Center Patrons. and
Clustering Techniques and STATISTICA Case Study: Defining Clusters of Shopping Center Patrons STATISTICA Solutions for Business Intelligence, Data Mining, Quality Control, and Web-based Analytics Table
15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
Performance Metrics for Graph Mining Tasks
Performance Metrics for Graph Mining Tasks 1 Outline Introduction to Performance Metrics Supervised Learning Performance Metrics Unsupervised Learning Performance Metrics Optimizing Metrics Statistical
SoSe 2014: M-TANI: Big Data Analytics
SoSe 2014: M-TANI: Big Data Analytics Lecture 4 21/05/2014 Sead Izberovic Dr. Nikolaos Korfiatis Agenda Recap from the previous session Clustering Introduction Distance mesures Hierarchical Clustering
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set Amhmed A. Bhih School of Electrical and Electronic Engineering Princy Johnson School of Electrical and Electronic Engineering Martin
Example: Document Clustering. Clustering: Definition. Notion of a Cluster can be Ambiguous. Types of Clusterings. Hierarchical Clustering
Overview Prognostic Models and Data Mining in Medicine, part I Cluster Analsis What is Cluster Analsis? K-Means Clustering Hierarchical Clustering Cluster Validit Eample: Microarra data analsis 6 Summar
Unsupervised Learning and Data Mining. Unsupervised Learning and Data Mining. Clustering. Supervised Learning. Supervised Learning
Unsupervised Learning and Data Mining Unsupervised Learning and Data Mining Clustering Decision trees Artificial neural nets K-nearest neighbor Support vectors Linear regression Logistic regression...
Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.
Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C
Nonlinear Iterative Partial Least Squares Method
Numerical Methods for Determining Principal Component Analysis Abstract Factors Béchu, S., Richard-Plouet, M., Fernandez, V., Walton, J., and Fairley, N. (2016) Developments in numerical treatments for
Algebra 1 2008. Academic Content Standards Grade Eight and Grade Nine Ohio. Grade Eight. Number, Number Sense and Operations Standard
Academic Content Standards Grade Eight and Grade Nine Ohio Algebra 1 2008 Grade Eight STANDARDS Number, Number Sense and Operations Standard Number and Number Systems 1. Use scientific notation to express
Session 7 Bivariate Data and Analysis
Session 7 Bivariate Data and Analysis Key Terms for This Session Previously Introduced mean standard deviation New in This Session association bivariate analysis contingency table co-variation least squares
Solving Geometric Problems with the Rotating Calipers *
Solving Geometric Problems with the Rotating Calipers * Godfried Toussaint School of Computer Science McGill University Montreal, Quebec, Canada ABSTRACT Shamos [1] recently showed that the diameter of
Cluster Analysis. Alison Merikangas Data Analysis Seminar 18 November 2009
Cluster Analysis Alison Merikangas Data Analysis Seminar 18 November 2009 Overview What is cluster analysis? Types of cluster Distance functions Clustering methods Agglomerative K-means Density-based Interpretation
A Study of Web Log Analysis Using Clustering Techniques
A Study of Web Log Analysis Using Clustering Techniques Hemanshu Rana 1, Mayank Patel 2 Assistant Professor, Dept of CSE, M.G Institute of Technical Education, Gujarat India 1 Assistant Professor, Dept
Classifying Large Data Sets Using SVMs with Hierarchical Clusters. Presented by :Limou Wang
Classifying Large Data Sets Using SVMs with Hierarchical Clusters Presented by :Limou Wang Overview SVM Overview Motivation Hierarchical micro-clustering algorithm Clustering-Based SVM (CB-SVM) Experimental
Discovering Local Subgroups, with an Application to Fraud Detection
Discovering Local Subgroups, with an Application to Fraud Detection Abstract. In Subgroup Discovery, one is interested in finding subgroups that behave differently from the average behavior of the entire
Object Recognition and Template Matching
Object Recognition and Template Matching Template Matching A template is a small image (sub-image) The goal is to find occurrences of this template in a larger image That is, you want to find matches of
0.1 What is Cluster Analysis?
Cluster Analysis 1 2 0.1 What is Cluster Analysis? Cluster analysis is concerned with forming groups of similar objects based on several measurements of different kinds made on the objects. The key idea
Common Core Unit Summary Grades 6 to 8
Common Core Unit Summary Grades 6 to 8 Grade 8: Unit 1: Congruence and Similarity- 8G1-8G5 rotations reflections and translations,( RRT=congruence) understand congruence of 2 d figures after RRT Dilations
203.4770: Introduction to Machine Learning Dr. Rita Osadchy
203.4770: Introduction to Machine Learning Dr. Rita Osadchy 1 Outline 1. About the Course 2. What is Machine Learning? 3. Types of problems and Situations 4. ML Example 2 About the course Course Homepage:
The Science and Art of Market Segmentation Using PROC FASTCLUS Mark E. Thompson, Forefront Economics Inc, Beaverton, Oregon
The Science and Art of Market Segmentation Using PROC FASTCLUS Mark E. Thompson, Forefront Economics Inc, Beaverton, Oregon ABSTRACT Effective business development strategies often begin with market segmentation,
High-dimensional labeled data analysis with Gabriel graphs
High-dimensional labeled data analysis with Gabriel graphs Michaël Aupetit CEA - DAM Département Analyse Surveillance Environnement BP 12-91680 - Bruyères-Le-Châtel, France Abstract. We propose the use
LABEL PROPAGATION ON GRAPHS. SEMI-SUPERVISED LEARNING. ----Changsheng Liu 10-30-2014
LABEL PROPAGATION ON GRAPHS. SEMI-SUPERVISED LEARNING ----Changsheng Liu 10-30-2014 Agenda Semi Supervised Learning Topics in Semi Supervised Learning Label Propagation Local and global consistency Graph
Steven M. Ho!and. Department of Geology, University of Georgia, Athens, GA 30602-2501
CLUSTER ANALYSIS Steven M. Ho!and Department of Geology, University of Georgia, Athens, GA 30602-2501 January 2006 Introduction Cluster analysis includes a broad suite of techniques designed to find groups
Linear Classification. Volker Tresp Summer 2015
Linear Classification Volker Tresp Summer 2015 1 Classification Classification is the central task of pattern recognition Sensors supply information about an object: to which class do the object belong
1816 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 15, NO. 7, JULY 2006. Principal Components Null Space Analysis for Image and Video Classification
1816 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 15, NO. 7, JULY 2006 Principal Components Null Space Analysis for Image and Video Classification Namrata Vaswani, Member, IEEE, and Rama Chellappa, Fellow,
Machine Learning using MapReduce
Machine Learning using MapReduce What is Machine Learning Machine learning is a subfield of artificial intelligence concerned with techniques that allow computers to improve their outputs based on previous
How To Understand Multivariate Models
Neil H. Timm Applied Multivariate Analysis With 42 Figures Springer Contents Preface Acknowledgments List of Tables List of Figures vii ix xix xxiii 1 Introduction 1 1.1 Overview 1 1.2 Multivariate Models
Classification Techniques for Remote Sensing
Classification Techniques for Remote Sensing Selim Aksoy Department of Computer Engineering Bilkent University Bilkent, 06800, Ankara [email protected] http://www.cs.bilkent.edu.tr/ saksoy/courses/cs551
Multivariate Analysis of Ecological Data
Multivariate Analysis of Ecological Data MICHAEL GREENACRE Professor of Statistics at the Pompeu Fabra University in Barcelona, Spain RAUL PRIMICERIO Associate Professor of Ecology, Evolutionary Biology
Binary Image Scanning Algorithm for Cane Segmentation
Binary Image Scanning Algorithm for Cane Segmentation Ricardo D. C. Marin Department of Computer Science University Of Canterbury Canterbury, Christchurch [email protected] Tom
VISUALIZING HIERARCHICAL DATA. Graham Wills SPSS Inc., http://willsfamily.org/gwills
VISUALIZING HIERARCHICAL DATA Graham Wills SPSS Inc., http://willsfamily.org/gwills SYNONYMS Hierarchical Graph Layout, Visualizing Trees, Tree Drawing, Information Visualization on Hierarchies; Hierarchical
NCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( )
Chapter 340 Principal Components Regression Introduction is a technique for analyzing multiple regression data that suffer from multicollinearity. When multicollinearity occurs, least squares estimates
Cluster Analysis: Basic Concepts and Algorithms
Cluster Analsis: Basic Concepts and Algorithms What does it mean clustering? Applications Tpes of clustering K-means Intuition Algorithm Choosing initial centroids Bisecting K-means Post-processing Strengths
Character Image Patterns as Big Data
22 International Conference on Frontiers in Handwriting Recognition Character Image Patterns as Big Data Seiichi Uchida, Ryosuke Ishida, Akira Yoshida, Wenjie Cai, Yaokai Feng Kyushu University, Fukuoka,
An Introduction to Machine Learning
An Introduction to Machine Learning L5: Novelty Detection and Regression Alexander J. Smola Statistical Machine Learning Program Canberra, ACT 0200 Australia [email protected] Tata Institute, Pune,
Logistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression
Logistic Regression Department of Statistics The Pennsylvania State University Email: [email protected] Logistic Regression Preserve linear classification boundaries. By the Bayes rule: Ĝ(x) = arg max
Manifold Learning Examples PCA, LLE and ISOMAP
Manifold Learning Examples PCA, LLE and ISOMAP Dan Ventura October 14, 28 Abstract We try to give a helpful concrete example that demonstrates how to use PCA, LLE and Isomap, attempts to provide some intuition
There are a number of different methods that can be used to carry out a cluster analysis; these methods can be classified as follows:
Statistics: Rosie Cornish. 2007. 3.1 Cluster Analysis 1 Introduction This handout is designed to provide only a brief introduction to cluster analysis and how it is done. Books giving further details are
UNSUPERVISED MACHINE LEARNING TECHNIQUES IN GENOMICS
UNSUPERVISED MACHINE LEARNING TECHNIQUES IN GENOMICS Dwijesh C. Mishra I.A.S.R.I., Library Avenue, New Delhi-110 012 [email protected] What is Learning? "Learning denotes changes in a system that enable
Analysis of kiva.com Microlending Service! Hoda Eydgahi Julia Ma Andy Bardagjy December 9, 2010 MAS.622j
Analysis of kiva.com Microlending Service! Hoda Eydgahi Julia Ma Andy Bardagjy December 9, 2010 MAS.622j What is Kiva? An organization that allows people to lend small amounts of money via the Internet
USING THE AGGLOMERATIVE METHOD OF HIERARCHICAL CLUSTERING AS A DATA MINING TOOL IN CAPITAL MARKET 1. Vera Marinova Boncheva
382 [7] Reznik, A, Kussul, N., Sokolov, A.: Identification of user activity using neural networks. Cybernetics and computer techniques, vol. 123 (1999) 70 79. (in Russian) [8] Kussul, N., et al. : Multi-Agent
Classification of Fingerprints. Sarat C. Dass Department of Statistics & Probability
Classification of Fingerprints Sarat C. Dass Department of Statistics & Probability Fingerprint Classification Fingerprint classification is a coarse level partitioning of a fingerprint database into smaller
A Demonstration of Hierarchical Clustering
Recitation Supplement: Hierarchical Clustering and Principal Component Analysis in SAS November 18, 2002 The Methods In addition to K-means clustering, SAS provides several other types of unsupervised
Vector Quantization and Clustering
Vector Quantization and Clustering Introduction K-means clustering Clustering issues Hierarchical clustering Divisive (top-down) clustering Agglomerative (bottom-up) clustering Applications to speech recognition
DATA ANALYSIS II. Matrix Algorithms
DATA ANALYSIS II Matrix Algorithms Similarity Matrix Given a dataset D = {x i }, i=1,..,n consisting of n points in R d, let A denote the n n symmetric similarity matrix between the points, given as where
CS 2750 Machine Learning. Lecture 1. Machine Learning. http://www.cs.pitt.edu/~milos/courses/cs2750/ CS 2750 Machine Learning.
Lecture Machine Learning Milos Hauskrecht [email protected] 539 Sennott Square, x5 http://www.cs.pitt.edu/~milos/courses/cs75/ Administration Instructor: Milos Hauskrecht [email protected] 539 Sennott
Introduction to General and Generalized Linear Models
Introduction to General and Generalized Linear Models General Linear Models - part I Henrik Madsen Poul Thyregod Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby
Spazi vettoriali e misure di similaritá
Spazi vettoriali e misure di similaritá R. Basili Corso di Web Mining e Retrieval a.a. 2008-9 April 10, 2009 Outline Outline Spazi vettoriali a valori reali Operazioni tra vettori Indipendenza Lineare
2. Simple Linear Regression
Research methods - II 3 2. Simple Linear Regression Simple linear regression is a technique in parametric statistics that is commonly used for analyzing mean response of a variable Y which changes according
Decision Support System Methodology Using a Visual Approach for Cluster Analysis Problems
Decision Support System Methodology Using a Visual Approach for Cluster Analysis Problems Ran M. Bittmann School of Business Administration Ph.D. Thesis Submitted to the Senate of Bar-Ilan University Ramat-Gan,
Visualization by Linear Projections as Information Retrieval
Visualization by Linear Projections as Information Retrieval Jaakko Peltonen Helsinki University of Technology, Department of Information and Computer Science, P. O. Box 5400, FI-0015 TKK, Finland [email protected]
