# A FUZZY BASED APPROACH TO TEXT MINING AND DOCUMENT CLUSTERING

Save this PDF as:

Size: px
Start display at page:

## Transcription

1 A FUZZY BASED APPROACH TO TEXT MINING AND DOCUMENT CLUSTERING Sumit Goswami 1 and Mayank Singh Shishodia 2 1 Indian Institute of Technology-Kharagpur, Kharagpur, India 2 School of Computer Science and Engineering, VIT University, Vellore, India ABSTRACT Fuzzy logic deals with degrees of truth. In this paper, we have shown how to apply fuzzy logic in text mining in order to perform document clustering. We took an example of document clustering where the documents had to be clustered into two categories. The method involved cleaning up the text and stemming of words. Then, we chose m features which differ significantly in their word frequencies (WF), normalized by document length, between documents belonging to these two clusters. The documents to be clustered were represented as a collection of m normalized WF values. Fuzzy c-means (FCM) algorithm was used to cluster these documents into two clusters. After the FCM execution finished, the documents in the two clusters were analysed for the values of their respective m features. It was known that documents belonging to a document type X tend to have higher WF values for some particular features. If the documents belonging to a cluster had higher WF values for those same features, then that cluster was said to represent X. By fuzzy logic, we not only get the cluster name, but also the degree to which a document belongs to a cluster. KEYWORDS Fuzzy Logic, Text Mining, Fuzzy c-means, Document Clustering 1. INTRODUCTION There is an increasing amount of textual data available every day. Text Mining is the automated process of extracting some previously unknown information from analysis of these unstructured texts [1]. Although Text Mining is similar to Data Mining, yet they are quite different. Data mining is related to pattern-extraction from databases while for text-mining, it is natural language. It is also different from search, web search in particular, where the information being searched already exists. In Text Mining, the computer is trying to extract something new from the already available resources. In this review paper, we have applied fuzzy logic to text-mining to perform document clustering into a number of pre-specified clusters. As an example, we have shown how to classify given documents into two categories sports and politics. First, the documents are cleaned by removal of advertisements and tags. The hyphens are dealt with and the stop words are removed. Word stemming is carried out to represent each word by its root [2]. Each document is then represented as a bag of words, a method which assigns weights to words on basis of their significance in the document which we considered to be their frequency of occurrence. We calculated word frequency as

2 WF = (Word Count/ (Total Words in the Document)) x10000 (1) m number of words are chosen [3][4][5] and each document is represented as a set of m values which are the WF values of those corresponding m words in that document. We then perform FCM on this representation to cluster the documents into required number of categories. Pre-known knowledge is used to name these clusters. In the forthcoming sections, we explain what fuzzy logic is, how it differs from probability, the FCM algorithm and how it can be used to perform document clustering with an example. We tell why fuzzy logic should be used for this purpose. 2. FUZZY LOGIC Fuzzy logic is a mathematical logic model in which truth can be partial i.e. it can have value between 0 and 1, that is completely false and completely true [6]. It is based on approximate reasoning instead of exact reasoning. 2.1 Fuzzy Logic vs Probability Probability can be divided into two broad categories Evidential and Physical. Consider an event E with probability 80%. Evidential Probability says that given the current available evidence, the likelihood of E occurring is 80%. On the other hand, Physical Probability says that if the experiment involving E as an outcome was repeated an infinite number of times, then 80% times, E will occur [7]. For example, consider a car A that goes 150 Kilometres per hour. Now this car would definitely be considered fast. Now consider another car B that goes 180 Kilometres per hour. In a repeatable experiment, it is not that the car B will be fast more often than A [Physical Probability] or that given the current information; it is more likely that Car B is faster than Car A [Evidential Probability]. Both the cars possess the property fast, but in different degrees. If we use Fuzzy Logic for this problem and there is a class called Fast Cars that contains the cars with the property fast, then the respective membership values of both Car A and Car B can be compared. The higher membership value means that the corresponding car possesses the property fast to a higher degree. 2.2 Degrees of Truth The degrees of truth are obtained from the membership function. An example is shown in Figure1: Figure 1: Membership functions for word frequency of "ball"

3 Suppose, we are given a statement that high frequency of occurrence of the word Ball in a document means that the document is related to Sports. In order to associate numerical values with these high or similarly low, medium frequency of words, we use a membership function. Linguistic hedging can be used to create more categories by adding adjectives like Very, Fairly, Not so, Less, More etc. Figure1 shows a membership function which decides when the frequency of occurrence of the word ball in a document is said to be low, medium or high. But it should be noted that the membership function depends on the situation and problem as perceived by the designer. For example, someone else may design the membership function where the word frequency of even 500 is considered low. This can be the case for a document containing the ball-by-ball commentary of a cricket match. A sample portion of this document could be like The captain has handed over the ball to Player A for the final over. He runs in to bowl the first ball. It s a short ball outside the off stump. It s a no-ball! The batsman smashes the ball and it goes for a sixer. Here, the word ball occurs 5 times in total of 44 words. This is a 1136 word frequency (per words) value. 3. CLUSTERING Using Fuzzy Logic and text-mining we can cluster similar documents together. Document Clustering is used by the computer to group documents into meaningful groups. We will use c-means algorithm. This algorithm is classified into two types- Hard c-means and FCM [6]. Hard c-means algorithm is used to cluster m observations into c clusters. Each cluster has its cluster centre. The observation belongs to the cluster which has the least distance from it. The clustering is crisp, which means that each observation is clustered into one and only one cluster. FCM is a variation of the hard c-means clustering algorithm. Every observation here has a membership value associated with each of the clusters which is related inversely to the distance of that observation from the centre of the cluster. 3.1 The FCM Algorithm The main aim of FCM algorithm is to minimize the objective function given by [6] :, (2) Where m denotes any real number greater than 1 uij denotes membership degree of xi in Jth cluster x i is ith dimension of the measured data c i is the ith dimension of the cluster centre. The formula for updating the cluster centre c i is:

4 (3) The Partition Matrix values are updated using the formula: The FCM algorithm s iteration stops when the maximum change in the values of Fuzzy c-partition matrix is less than E, where E is a termination criterion with value between 0 and 1. The steps followed in FCM Algorithm are given below: 1. Initialize the fuzzy partition matrix, U=[u ij ] matrix, U (0) 2. At k-step: calculate the centre s vectors C (k) = [c j ] with U (k) using equation (3). 3. Update U (k),u (k+1) using equation (4). 4. If U (k+1) - U (k) < then STOP; otherwise return to 2. (4) The documents are clustered based on a representation called bag of words. 4. Bag of Words Representation of a Document Words are assigned weights according to significance in the document which can be the number of occurrences of that word in that document. Consider a paragraph: It is a variation of the Hard C-Means clustering algorithm. Each observation here has a membership value associated with each of the clusters which is related inversely to the distance of that observation from the centre of the cluster. The Bag of Words representation for the above paragraph will look similar to Table1.

5 Table 1: Bag of Words representation for the paragraph Word Occurrences Variation 1 Hard 1 Cluster 2 Observation 2 Centre 1 Apart from single words, we can also use phrases better results. Few examples of these are engineering student, computer games, Tourist attraction, gold medal etc. The same Bag of Words technique can be modified to treat the phrases as single words. In our method, we assigned the weights as the number of occurrences of the word in the document normalized by the document length and multiplied by 10, STEPS INVOLVED IN APPLYING FUZZY LOGIC TO TEXT MINING The following steps are followed in order to apply fuzzy logic to text mining. 5.1 Pre-processing of Text This phase involves cleaning up the text like removing advertisements etc. Also we have to deal with hyphens, remove tags from documents like HTML, XML etc. Removal of this unwanted text helps improve the efficiency of our algorithm. 5.2 Feature Generation The documents are represented in the Bag of Words method. Stop words like the, a, an etc. are removed. Word Stemming [2] is carried out. Word Stemming refers to representing the word by its root, example the words - driving, drove and driven should be represented by drive. This is important because not all the documents are in the same active or passive voice, or the same tense. 5.3 Feature Selection Here, the features that we will use are selected. This selection of features can either be done before use or based on use. Number of features that will be used for our process is further reduced. We choose only the features which will help us in our process [3][4][5]. In order to cluster separately, the documents relating to sports and politics, presence of names of people in a document will not help much. There can be names in both of those. However, presence of words like stadium, ball, cheerleaders etc. will help relate the documents to the field of sports. Similarly, words like election, democracy, parliament will relate the documents to politics category.

6 In some situations, however, it will be hard to manually point out the features which distinguish documents that should belong to different types. In such situations, we need to find out what features we should use for our clustering. To do this, we can take sample documents which are already known to belong to our required categories. Then the information gained from the analysis of these documents can be applied to point out the differences among documents belonging to different clusters. The word frequency, for various words (normalized by document length), can be calculated for these documents. We choose the words with a significant word frequency variation between different document types. 5.4 Clustering The method used here is FCM clustering algorithm. Each document to be clustered has already been represented as a bag of words, and from that bag only the essential words (features) have been kept. The aim now is to cluster the documents into categories (called classes in FCM).Using FCM, the documents having similar frequencies (normalized by document length) of various selected features are clustered together. The number of clusters to be made can be specified at an early stage but this might result in inaccurate results later if this number is not appropriate. This problem is known as Cluster Validity. Once the FCM execution finishes, either due to accomplishing the pre-specified number of iterations or because of max-change in the Fuzzy Partition Matrix being lesser than the threshold, we obtain the required number of clusters and all the documents have been clustered among them. 5.5 Evaluation and Interpretation of Results From the Fuzzy Partition Matrix, one can find to what degree a given document belongs to each cluster. If the membership value of a document for a given cluster is high, the document can be said to strongly belong to that cluster. However, if all the membership values corresponding to a specific document are almost same (Example.35,.35,.30 among three clusters), it would imply that the document does not quite strongly belong to any of those clusters. 6. EXAMPLE Suppose we are given some documents which we have to cluster into two categories - sports and politics. We will not present the first two steps that we followed as they are pretty simple. We will start from the third feature selection. We had no idea about which words (features) we could use to cluster our documents. Hence, we had to find such features. We took some documents that we knew were related to Sports and Politics respectively. We applied step1 and step 2 on them and then counted the word frequency of various words using eq(1). Then we analysed the differences in these WF values for same words between documents relating to Sports category and the documents relating to Politics category. The partial result of our analysis is presented in Table 2.

7 Table 2: Word Frequencies for Sports and Politics-related documents Word WF (Sports) WF (Politics) Win Stadium Democracy Ball Team Candidate Campaign By looking at the Table 2, it was apparent that Words like Win, Campaign are used in similar amounts in documents belonging to both groups. However, words like Stadium (203 v/s 7), Ball (501 v/s 30), Team (250 v/s 80) and Democracy (1 v/s 140) could be used to differentiate among documents belonging to Sports and Politics respectively. For the sake of simplicity and an easy example, we chose only 4 words Stadium, Ball, Team and Democracy. On basis of these 4 words, we clustered the given documents into Politics and Sports categories. Now that we had the documents to be clustered and had selected the features we wanted to use for the clustering process; we could start our clustering process. Let D= {d 1, d 2, d 3 d n } represent our n documents to be clustered. Now each of these documents, d i, is defined by the m selected features i.e. d i = {d i1, d i2, d i3..d im } Here, each d i in the universe D is a m-dimensional vector representing the m selected features which will be normalised in the same way we did for feature selecting by using eq(1). An alternate way to look at this problem is considering each d i as a point in m-dimensional space representing the features. D, the universe, is a set of points with n elements (documents) in the same sample space. Now we have to initialise the fuzzy partition matrix. Keeping in mind the constraints that all entries should be in the interval [0, 1] and the columns should add to a total of 1 and considering that we are given 4 documents to cluster:

8 Table 3: Fuzzy Partition Matrix Doc1 Doc2 Doc3 Doc4 Doc5 Doc6 Doc7 Doc8 Cluster Cluster Our initial assumption is that Doc1, Doc2, Doc5 and Doc6 belong to cluster 1 and Doc3, Doc4, Doc7 and Doc8 belong to cluster 2. Now we have to calculate the initial cluster centres. For c=1, that is cluster1, the centre (V 1 ) can be calculated using eq (3). Since, the membership value of Doc3, Doc4, Doc7 and Doc8 in Cluster 1 is 0, we have V 1j =((1)x 1j +(1)x 2j +(1)x 5j +(1)x 6j ) / ( )= (x 1j +x 2j + x 5j + x 6j )/4 Now, say our word frequencies were like Table 4: Word frequencies in given documents (to be clustered) Doc1 Doc2 Doc3 Doc4 Doc5 Doc6 Doc7 Doc8 Stadium Ball Team Democracy V 11 = ( )/4= Similarly, Calculating all V 1j we get V 1 = {149.25, 300, 162.5, 7.5} is the centre of cluster 1. Similarly, V 2 = {50,110.75, 67.75, 33.25} is the centre of cluster 2. Now calculating the Euclidian distances of each Document vector from both centre clusters: D 11 = (( ) 2 + ( ) 2 + ( ) 2 + (1-7.75) 2 ) 1/2 = Similarly, D 12 = 149.5, D 13 = 339.5, D 14 = 352.5, D 15 = 102.2, D 16 = 353.4, D 17 = 109.2, D 18 = Similarly, from the cluster two:

9 D 21 = ((180-50) 2 + ( ) 2 + ( ) 2 + ( ) 2 ) 1/2 = D 22 = 382.2, D 23 = 105.1, D 24 = 118.3, D 25 = 334.4, D 26 = 119.6, D 27 = 339.7, D 28 = Now that we have distance of each vector from both cluster centres, we will now update the Fuzzy Partition Matrix, by using eq(4). Note that we have already decided the value of the m in this formula, earlier in the example. U 11 = [1+ ( / ) 2 ] -1 =.90, U 12 =.867, U 23 =.913, U 24 =.898, U 15 =.9146, U 26 =.897, U 17 =.906, U 28 =.901 The remaining values of the fuzzy partition matrix are updated such that each column adds to 1. We get the updated Fuzzy Partition Matrix as Table 5: Updated fuzzy partition matrix after one iteration Doc1 Doc2 Doc3 Doc4 Doc5 Doc6 Doc7 Doc8 Cluster Cluster Say our threshold change (stopping condition) was Our max change here is which is greater than Hence we will continue. Now we will repeat the process by again calculating the new cluster centres. But now we will use the updated fuzzy partition matrix values. So, now our centre for cluster 1 will be calculated using V 1j = (0.900 x 1j x 2j x 3j x 4j x 5j x 6j x 7j x 8j )/ ( ) 7. INTERPRETING THE RESULTS Suppose that our final fuzzy partition matrix for 8 documents looks something like this: Table 6: Sample Final Fuzzy Partition Matrix Doc1 Doc2 Doc3 Doc4 Doc5 Doc6 Doc7 Doc8 Cluster Cluster

10 We see that Doc1, Doc2, Doc5, Doc7 belong to cluster 1 whereas Doc3, Doc4, Doc6, Doc8 belong to cluster 2 on basis of high membership values in those respective clusters. In the example we took, the documents Doc1, Doc2, Doc5, and Doc7 had higher frequency of words like stadium, ball and team (which are related to sports). Since these documents have been clustered together, one can say that Cluster 1 is Sports. At the same time, Cluster2 will contain documents with higher frequency of words like democracy (which are related to politics). Hence Cluster2 represents Politics. Here, Doc1 relates to Sports to a large degree (due to its very high membership value). If we set the criteria that membership values greater than 0.85 with respect to a given cluster can be called strong membership, then Doc1 and Doc5 can be said to strongly belong to Cluster1 which is sports. Moreover, we can say that Doc1 relates to Sports more strongly than Doc5 does. This interpretation of results in linguistic form [8] is what gives advantage to usage of Fuzzy Logic over Probability models like Bayesian in Text Mining. 8. CONCLUSION In this paper, we showed how one can use fuzzy logic in text mining to cluster documents by taking an example where the documents were clustered into two topics :- sports and politics. The advantage of using fuzzy logic over probability was that in the former, we could calculate the degree to which a given document belonged to either categories- sports as well as politics. This was not possible in the probability model. In other words, instead of simply saying whether the document belonged to sports or politics, we could now tell the degree to which the document characteristics resembled that of a sports-related document as well as a politics-related document. By doing this for all documents in the data-set, we could also compare two documents and tell which one belongs more to which topic. Also, because every document will have some membership values in each of the clusters, no useful document will ever excluded from search results [1][5]. 9. REFERENCES 1. Vishal Gupta, Gurpreet S. Lehal; A Survey of Text Mining Techniques and Applications ; Journal of Emerging Technologies in Web Intelligence, Vol.1, No.1, August K.Sathiyakumari, V.Preamsudha, G.Manimekalai; Unsupervised Approach for Document Clustering Using Modified Fuzzy C mean Algorithm ; International Journal of Computer & Organization Trends Volume 11 Issue R. Rajendra Prasath, Sudeshna Sarkar: Unsupervised Feature Generation using Knowledge Repositories for Effective Text Categorization. ECAI 2010: Sumit Goswami, Sudeshna Sarkar, Mayur Rustagi: Stylometric Analysis of Bloggers' Age and Gender. ICWSM Sumit Goswami, Mayank Singh Shishodia; A fuzzy based approach to stylometric analysis of blogger s age and gender ; HIS 2012: Ross, T. J. (2010); Fuzzy Logic with Engineering Applications, Third Edition, John Wiley & Sons, Ltd, Chichester, UK 7. Fuzzy logic vs Proobability (Good Math, Bad Math); /goodmath/ 2011 /02 /02 / fuzzy-logic-vs-probability/, last checked on 28th July Nogueira, T.M. ; On The Use of Fuzzy Rules to Text Document Classification ; th International Conference on Hybrid Intelligent Systems (HIS),; Aug 2010 Atlanta, US.

### FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT MINING SYSTEM

International Journal of Innovative Computing, Information and Control ICIC International c 0 ISSN 34-48 Volume 8, Number 8, August 0 pp. 4 FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT

### Use of Data Mining Techniques to Improve the Effectiveness of Sales and Marketing

Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 4, Issue. 4, April 2015,

### dm106 TEXT MINING FOR CUSTOMER RELATIONSHIP MANAGEMENT: AN APPROACH BASED ON LATENT SEMANTIC ANALYSIS AND FUZZY CLUSTERING

dm106 TEXT MINING FOR CUSTOMER RELATIONSHIP MANAGEMENT: AN APPROACH BASED ON LATENT SEMANTIC ANALYSIS AND FUZZY CLUSTERING ABSTRACT In most CRM (Customer Relationship Management) systems, information on

### Performance Analysis of Naive Bayes and J48 Classification Algorithm for Data Classification

Performance Analysis of Naive Bayes and J48 Classification Algorithm for Data Classification Tina R. Patil, Mrs. S. S. Sherekar Sant Gadgebaba Amravati University, Amravati tnpatil2@gmail.com, ss_sherekar@rediffmail.com

### Joint models for classification and comparison of mortality in different countries.

Joint models for classification and comparison of mortality in different countries. Viani D. Biatat 1 and Iain D. Currie 1 1 Department of Actuarial Mathematics and Statistics, and the Maxwell Institute

### Clustering. Danilo Croce Web Mining & Retrieval a.a. 2015/201 16/03/2016

Clustering Danilo Croce Web Mining & Retrieval a.a. 2015/201 16/03/2016 1 Supervised learning vs. unsupervised learning Supervised learning: discover patterns in the data that relate data attributes with

### Knowledge Discovery and Data Mining. Structured vs. Non-Structured Data

Knowledge Discovery and Data Mining Unit # 2 1 Structured vs. Non-Structured Data Most business databases contain structured data consisting of well-defined fields with numeric or alphanumeric values.

### Web Mining using Artificial Ant Colonies : A Survey

Web Mining using Artificial Ant Colonies : A Survey Richa Gupta Department of Computer Science University of Delhi ABSTRACT : Web mining has been very crucial to any organization as it provides useful

### Evolutionary Detection of Rules for Text Categorization. Application to Spam Filtering

Advances in Intelligent Systems and Technologies Proceedings ECIT2004 - Third European Conference on Intelligent Systems and Technologies Iasi, Romania, July 21-23, 2004 Evolutionary Detection of Rules

### A Study of Web Log Analysis Using Clustering Techniques

A Study of Web Log Analysis Using Clustering Techniques Hemanshu Rana 1, Mayank Patel 2 Assistant Professor, Dept of CSE, M.G Institute of Technical Education, Gujarat India 1 Assistant Professor, Dept

### QOS Based Web Service Ranking Using Fuzzy C-means Clusters

Research Journal of Applied Sciences, Engineering and Technology 10(9): 1045-1050, 2015 ISSN: 2040-7459; e-issn: 2040-7467 Maxwell Scientific Organization, 2015 Submitted: March 19, 2015 Accepted: April

### Using Data Mining for Mobile Communication Clustering and Characterization

Using Data Mining for Mobile Communication Clustering and Characterization A. Bascacov *, C. Cernazanu ** and M. Marcu ** * Lasting Software, Timisoara, Romania ** Politehnica University of Timisoara/Computer

### SPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING

AAS 07-228 SPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING INTRODUCTION James G. Miller * Two historical uncorrelated track (UCT) processing approaches have been employed using general perturbations

### Data quality in Accounting Information Systems

Data quality in Accounting Information Systems Comparing Several Data Mining Techniques Erjon Zoto Department of Statistics and Applied Informatics Faculty of Economy, University of Tirana Tirana, Albania

### In this presentation, you will be introduced to data mining and the relationship with meaningful use.

In this presentation, you will be introduced to data mining and the relationship with meaningful use. Data mining refers to the art and science of intelligent data analysis. It is the application of machine

### Standardization and Its Effects on K-Means Clustering Algorithm

Research Journal of Applied Sciences, Engineering and Technology 6(7): 399-3303, 03 ISSN: 040-7459; e-issn: 040-7467 Maxwell Scientific Organization, 03 Submitted: January 3, 03 Accepted: February 5, 03

### STATISTICA. Clustering Techniques. Case Study: Defining Clusters of Shopping Center Patrons. and

Clustering Techniques and STATISTICA Case Study: Defining Clusters of Shopping Center Patrons STATISTICA Solutions for Business Intelligence, Data Mining, Quality Control, and Web-based Analytics Table

### Categorical Data Visualization and Clustering Using Subjective Factors

Categorical Data Visualization and Clustering Using Subjective Factors Chia-Hui Chang and Zhi-Kai Ding Department of Computer Science and Information Engineering, National Central University, Chung-Li,

### Machine Learning using MapReduce

Machine Learning using MapReduce What is Machine Learning Machine learning is a subfield of artificial intelligence concerned with techniques that allow computers to improve their outputs based on previous

### DECISION TREE INDUCTION FOR FINANCIAL FRAUD DETECTION USING ENSEMBLE LEARNING TECHNIQUES

DECISION TREE INDUCTION FOR FINANCIAL FRAUD DETECTION USING ENSEMBLE LEARNING TECHNIQUES Vijayalakshmi Mahanra Rao 1, Yashwant Prasad Singh 2 Multimedia University, Cyberjaya, MALAYSIA 1 lakshmi.mahanra@gmail.com

### International Journal of Computer Science Trends and Technology (IJCST) Volume 2 Issue 3, May-Jun 2014

RESEARCH ARTICLE OPEN ACCESS A Survey of Data Mining: Concepts with Applications and its Future Scope Dr. Zubair Khan 1, Ashish Kumar 2, Sunny Kumar 3 M.Tech Research Scholar 2. Department of Computer

### Effective Data Retrieval Mechanism Using AML within the Web Based Join Framework

Effective Data Retrieval Mechanism Using AML within the Web Based Join Framework Usha Nandini D 1, Anish Gracias J 2 1 ushaduraisamy@yahoo.co.in 2 anishgracias@gmail.com Abstract A vast amount of assorted

### Social Media Mining. Data Mining Essentials

Introduction Data production rate has been increased dramatically (Big Data) and we are able store much more data than before E.g., purchase data, social media data, mobile phone data Businesses and customers

### Prototype-based classification by fuzzification of cases

Prototype-based classification by fuzzification of cases Parisa KordJamshidi Dep.Telecommunications and Information Processing Ghent university pkord@telin.ugent.be Bernard De Baets Dep. Applied Mathematics

### Detection of DDoS Attack Scheme

Chapter 4 Detection of DDoS Attac Scheme In IEEE 802.15.4 low rate wireless personal area networ, a distributed denial of service attac can be launched by one of three adversary types, namely, jamming

### DATA ANALYTICS USING R

DATA ANALYTICS USING R Duration: 90 Hours Intended audience and scope: The course is targeted at fresh engineers, practicing engineers and scientists who are interested in learning and understanding data

### International Journal of Software and Web Sciences (IJSWS) Web Log Mining Based on Improved FCM Algorithm using Multiobjective

International Association of Scientific Innovation and Research (IASIR) (An Association Unifying the Sciences, Engineering, and Applied Research) ISSN (Print): 2279-0063 ISSN (Online): 2279-0071 International

### Data Mining Part 5. Prediction

Data Mining Part 5. Prediction 5.1 Spring 2010 Instructor: Dr. Masoud Yaghini Outline Classification vs. Numeric Prediction Prediction Process Data Preparation Comparing Prediction Methods References Classification

### W. Heath Rushing Adsurgo LLC. Harness the Power of Text Analytics: Unstructured Data Analysis for Healthcare. Session H-1 JTCC: October 23, 2015

W. Heath Rushing Adsurgo LLC Harness the Power of Text Analytics: Unstructured Data Analysis for Healthcare Session H-1 JTCC: October 23, 2015 Outline Demonstration: Recent article on cnn.com Introduction

### Fortgeschrittene Computerintensive Methoden: Fuzzy Clustering Steffen Unkel

Fortgeschrittene Computerintensive Methoden: Fuzzy Clustering Steffen Unkel Institut für Statistik LMU München Sommersemester 2013 Outline 1 Setting the scene 2 Methods for fuzzy clustering 3 The assessment

### BIDM Project. Predicting the contract type for IT/ITES outsourcing contracts

BIDM Project Predicting the contract type for IT/ITES outsourcing contracts N a n d i n i G o v i n d a r a j a n ( 6 1 2 1 0 5 5 6 ) The authors believe that data modelling can be used to predict if an

### Data Mining Approach For Subscription-Fraud. Detection in Telecommunication Sector

Contemporary Engineering Sciences, Vol. 7, 2014, no. 11, 515-522 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ces.2014.4431 Data Mining Approach For Subscription-Fraud Detection in Telecommunication

### Credit Card Fraud Detection Using Self Organised Map

International Journal of Information & Computation Technology. ISSN 0974-2239 Volume 4, Number 13 (2014), pp. 1343-1348 International Research Publications House http://www. irphouse.com Credit Card Fraud

### Improved Fuzzy C-means Clustering Algorithm Based on Cluster Density

Journal of Computational Information Systems 8: 2 (2012) 727 737 Available at http://www.jofcis.com Improved Fuzzy C-means Clustering Algorithm Based on Cluster Density Xiaojun LOU, Junying LI, Haitao

### 15.062 Data Mining: Algorithms and Applications Matrix Math Review

.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop

### BEHAVIOR BASED CREDIT CARD FRAUD DETECTION USING SUPPORT VECTOR MACHINES

BEHAVIOR BASED CREDIT CARD FRAUD DETECTION USING SUPPORT VECTOR MACHINES 123 CHAPTER 7 BEHAVIOR BASED CREDIT CARD FRAUD DETECTION USING SUPPORT VECTOR MACHINES 7.1 Introduction Even though using SVM presents

### Comparison of K-means and Backpropagation Data Mining Algorithms

Comparison of K-means and Backpropagation Data Mining Algorithms Nitu Mathuriya, Dr. Ashish Bansal Abstract Data mining has got more and more mature as a field of basic research in computer science and

### Environmental Remote Sensing GEOG 2021

Environmental Remote Sensing GEOG 2021 Lecture 4 Image classification 2 Purpose categorising data data abstraction / simplification data interpretation mapping for land cover mapping use land cover class

### Least Squares Estimation

Least Squares Estimation SARA A VAN DE GEER Volume 2, pp 1041 1045 in Encyclopedia of Statistics in Behavioral Science ISBN-13: 978-0-470-86080-9 ISBN-10: 0-470-86080-4 Editors Brian S Everitt & David

: Table of Contents... 1 Overview of Model... 1 Dispersion... 2 Parameterization... 3 Sigma-Restricted Model... 3 Overparameterized Model... 4 Reference Coding... 4 Model Summary (Summary Tab)... 5 Summary

### Clustering Technique in Data Mining for Text Documents

Clustering Technique in Data Mining for Text Documents Ms.J.Sathya Priya Assistant Professor Dept Of Information Technology. Velammal Engineering College. Chennai. Ms.S.Priyadharshini Assistant Professor

### Cryptography and Network Security Department of Computer Science and Engineering Indian Institute of Technology Kharagpur

Cryptography and Network Security Department of Computer Science and Engineering Indian Institute of Technology Kharagpur Module No. # 01 Lecture No. # 05 Classic Cryptosystems (Refer Slide Time: 00:42)

### A Survey on Intrusion Detection System with Data Mining Techniques

A Survey on Intrusion Detection System with Data Mining Techniques Ms. Ruth D 1, Mrs. Lovelin Ponn Felciah M 2 1 M.Phil Scholar, Department of Computer Science, Bishop Heber College (Autonomous), Trichirappalli,

Building A Smart Academic Advising System Using Association Rule Mining Raed Shatnawi +962795285056 raedamin@just.edu.jo Qutaibah Althebyan +962796536277 qaalthebyan@just.edu.jo Baraq Ghalib & Mohammed

### Forecasting stock markets with Twitter

Forecasting stock markets with Twitter Argimiro Arratia argimiro@lsi.upc.edu Joint work with Marta Arias and Ramón Xuriguera To appear in: ACM Transactions on Intelligent Systems and Technology, 2013,

### Statistical Validation and Data Analytics in ediscovery. Jesse Kornblum

Statistical Validation and Data Analytics in ediscovery Jesse Kornblum Administrivia Silence your mobile Interactive talk Please ask questions 2 Outline Introduction Big Questions What Makes Things Similar?

### Data Mining Classification: Decision Trees

Data Mining Classification: Decision Trees Classification Decision Trees: what they are and how they work Hunt s (TDIDT) algorithm How to select the best split How to handle Inconsistent data Continuous

### CLUSTERING FOR FORENSIC ANALYSIS

IMPACT: International Journal of Research in Engineering & Technology (IMPACT: IJRET) ISSN(E): 2321-8843; ISSN(P): 2347-4599 Vol. 2, Issue 4, Apr 2014, 129-136 Impact Journals CLUSTERING FOR FORENSIC ANALYSIS

### Movie Classification Using k-means and Hierarchical Clustering

Movie Classification Using k-means and Hierarchical Clustering An analysis of clustering algorithms on movie scripts Dharak Shah DA-IICT, Gandhinagar Gujarat, India dharak_shah@daiict.ac.in Saheb Motiani

### 1. LINEAR EQUATIONS. A linear equation in n unknowns x 1, x 2,, x n is an equation of the form

1. LINEAR EQUATIONS A linear equation in n unknowns x 1, x 2,, x n is an equation of the form a 1 x 1 + a 2 x 2 + + a n x n = b, where a 1, a 2,..., a n, b are given real numbers. For example, with x and

### Fig. 1 A typical Knowledge Discovery process [2]

Volume 4, Issue 7, July 2014 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com A Review on Clustering

### Search Result Optimization using Annotators

Search Result Optimization using Annotators Vishal A. Kamble 1, Amit B. Chougule 2 1 Department of Computer Science and Engineering, D Y Patil College of engineering, Kolhapur, Maharashtra, India 2 Professor,

### Manifold Learning Examples PCA, LLE and ISOMAP

Manifold Learning Examples PCA, LLE and ISOMAP Dan Ventura October 14, 28 Abstract We try to give a helpful concrete example that demonstrates how to use PCA, LLE and Isomap, attempts to provide some intuition

### Inner Classification of Clusters for Online News

Inner Classification of Clusters for Online News Harmandeep Kaur 1, Sheenam Malhotra 2 1 (Computer Science and Engineering Department, Shri Guru Granth Sahib World University Fatehgarh Sahib) 2 (Assistant

### Search and Information Retrieval

Search and Information Retrieval Search on the Web 1 is a daily activity for many people throughout the world Search and communication are most popular uses of the computer Applications involving search

### DATA MINING TECHNIQUES AND APPLICATIONS

DATA MINING TECHNIQUES AND APPLICATIONS Mrs. Bharati M. Ramageri, Lecturer Modern Institute of Information Technology and Research, Department of Computer Application, Yamunanagar, Nigdi Pune, Maharashtra,

### Intrusion Detection System using Log Files and Reinforcement Learning

Intrusion Detection System using Log Files and Reinforcement Learning Bhagyashree Deokar, Ambarish Hazarnis Department of Computer Engineering K. J. Somaiya College of Engineering, Mumbai, India ABSTRACT

### Data Mining for Business Analytics

Data Mining for Business Analytics Lecture 2: Introduction to Predictive Modeling Stern School of Business New York University Spring 2014 MegaTelCo: Predicting Customer Churn You just landed a great analytical

### Prediction of Heart Disease Using Naïve Bayes Algorithm

Prediction of Heart Disease Using Naïve Bayes Algorithm R.Karthiyayini 1, S.Chithaara 2 Assistant Professor, Department of computer Applications, Anna University, BIT campus, Tiruchirapalli, Tamilnadu,

### Support Vector Machines with Clustering for Training with Very Large Datasets

Support Vector Machines with Clustering for Training with Very Large Datasets Theodoros Evgeniou Technology Management INSEAD Bd de Constance, Fontainebleau 77300, France theodoros.evgeniou@insead.fr Massimiliano

### Strategic Online Advertising: Modeling Internet User Behavior with

2 Strategic Online Advertising: Modeling Internet User Behavior with Patrick Johnston, Nicholas Kristoff, Heather McGinness, Phuong Vu, Nathaniel Wong, Jason Wright with William T. Scherer and Matthew

### Grid Density Clustering Algorithm

Grid Density Clustering Algorithm Amandeep Kaur Mann 1, Navneet Kaur 2, Scholar, M.Tech (CSE), RIMT, Mandi Gobindgarh, Punjab, India 1 Assistant Professor (CSE), RIMT, Mandi Gobindgarh, Punjab, India 2

### Operations and Supply Chain Management Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras

Operations and Supply Chain Management Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras Lecture - 36 Location Problems In this lecture, we continue the discussion

### Impelling Heart Attack Prediction System using Data Mining and Artificial Neural Network

General Article International Journal of Current Engineering and Technology E-ISSN 2277 4106, P-ISSN 2347-5161 2014 INPRESSCO, All Rights Reserved Available at http://inpressco.com/category/ijcet Impelling

### Predicting the Stock Market with News Articles

Predicting the Stock Market with News Articles Kari Lee and Ryan Timmons CS224N Final Project Introduction Stock market prediction is an area of extreme importance to an entire industry. Stock price is

### Text Mining for Business Intelligence

Project Proposal (Draft) Text Mining for Business Intelligence By Abhinut Srimasorn (5322793399) Advisor Dr. Thanaruk Theeramunkong School of Information, Computer and Communication Technology, Sirindhorn

### FUZZY C - MEANS CLUSTERING IN MATLAB

FUZZY C - MEANS CLUSTERING IN MATLAB Makhalova Elena Abstract Paper is a survey of fuzzy logic theory applied in cluster analysis. Fuzzy logic becomes more and more important in modern science. It is widely

### Data Mining. Practical Machine Learning Tools and Techniques. Classification, association, clustering, numeric prediction

Data Mining Practical Machine Learning Tools and Techniques Slides for Chapter 2 of Data Mining by I. H. Witten and E. Frank Input: Concepts, instances, attributes Terminology What s a concept? Classification,

### Employee Survey Analysis

Employee Survey Analysis Josh Froelich, Megaputer Intelligence Sergei Ananyan, Megaputer Intelligence www.megaputer.com Megaputer Intelligence, Inc. 120 West Seventh Street, Suite 310 Bloomington, IN 47404

### A Survey on Outlier Detection Techniques for Credit Card Fraud Detection

IOSR Journal of Computer Engineering (IOSR-JCE) e-issn: 2278-0661, p- ISSN: 2278-8727Volume 16, Issue 2, Ver. VI (Mar-Apr. 2014), PP 44-48 A Survey on Outlier Detection Techniques for Credit Card Fraud

### A New Operation on Triangular Fuzzy Number for. Solving Fuzzy Linear Programming Problem

Applied Mathematical Sciences, Vol. 6, 2012, no. 11, 525 532 A New Operation on Triangular Fuzzy Number for Solving Fuzzy Linear Programming Problem A. Nagoor Gani PG and Research Department of Mathematics

### Clustering in Machine Learning. By: Ibrar Hussain Student ID:

Clustering in Machine Learning By: Ibrar Hussain Student ID: 11021083 Presentation An Overview Introduction Definition Types of Learning Clustering in Machine Learning K-means Clustering Example of k-means

### Chapter 6: The Information Function 129. CHAPTER 7 Test Calibration

Chapter 6: The Information Function 129 CHAPTER 7 Test Calibration 130 Chapter 7: Test Calibration CHAPTER 7 Test Calibration For didactic purposes, all of the preceding chapters have assumed that the

### Web Usage Mining for Predicting Users Browsing Behaviors by using FPCM Clustering

IACSIT International Journal of Engineering and Technology Vol 3 No 5 October 20 Web Usage Mining for Predicting Users Browsing Behaviors by using FPCM Clustering RKhanchana and M Punithavalli Abstract--World

### Machine Learning and Data Mining. Fundamentals, robotics, recognition

Machine Learning and Data Mining Fundamentals, robotics, recognition Machine Learning, Data Mining, Knowledge Discovery in Data Bases Their mutual relations Data Mining, Knowledge Discovery in Databases,

### An Overview of Knowledge Discovery Database and Data mining Techniques

An Overview of Knowledge Discovery Database and Data mining Techniques Priyadharsini.C 1, Dr. Antony Selvadoss Thanamani 2 M.Phil, Department of Computer Science, NGM College, Pollachi, Coimbatore, Tamilnadu,

### Parallel Algorithm for Dense Matrix Multiplication

Parallel Algorithm for Dense Matrix Multiplication CSE633 Parallel Algorithms Fall 2012 Ortega, Patricia Outline Problem definition Assumptions Implementation Test Results Future work Conclusions Problem

### Using multiple models: Bagging, Boosting, Ensembles, Forests

Using multiple models: Bagging, Boosting, Ensembles, Forests Bagging Combining predictions from multiple models Different models obtained from bootstrap samples of training data Average predictions or

### Chapter 20: Data Analysis

Chapter 20: Data Analysis Database System Concepts, 6 th Ed. See www.db-book.com for conditions on re-use Chapter 20: Data Analysis Decision Support Systems Data Warehousing Data Mining Classification

### Recommendation Tool Using Collaborative Filtering

Recommendation Tool Using Collaborative Filtering Aditya Mandhare 1, Soniya Nemade 2, M.Kiruthika 3 Student, Computer Engineering Department, FCRIT, Vashi, India 1 Student, Computer Engineering Department,

### Multivariate Analysis of Variance (MANOVA): I. Theory

Gregory Carey, 1998 MANOVA: I - 1 Multivariate Analysis of Variance (MANOVA): I. Theory Introduction The purpose of a t test is to assess the likelihood that the means for two groups are sampled from the

### ROBERTO BATTITI, MAURO BRUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Apr 2015

ROBERTO BATTITI, MAURO BRUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Apr 2015 http://intelligentoptimization.org/lionbook Roberto Battiti

### Mobile Customer Clustering Analysis Based on Call Detail Records

Mobile Customer Clustering Analysis Based on Call Detail Records Qining Economics and Management School Beijing University of Posts and Telecommunications ABSTRACT Competition in the mobile telecommunications

### Nine Common Types of Data Mining Techniques Used in Predictive Analytics

1 Nine Common Types of Data Mining Techniques Used in Predictive Analytics By Laura Patterson, President, VisionEdge Marketing Predictive analytics enable you to develop mathematical models to help better

### A Risk Management System Framework for New Product Development (NPD)

2011 International Conference on Economics and Finance Research IPEDR vol.4 (2011) (2011) IACSIT Press, Singapore A Risk Management System Framework for New Product Development (NPD) Seonmuk Park, Jongseong

### Behavioral Segmentation

Behavioral Segmentation TM Contents 1. The Importance of Segmentation in Contemporary Marketing... 2 2. Traditional Methods of Segmentation and their Limitations... 2 2.1 Lack of Homogeneity... 3 2.2 Determining

### Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets

Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets Macario O. Cordel II and Arnulfo P. Azcarraga College of Computer Studies *Corresponding Author: macario.cordel@dlsu.edu.ph

### TRANSPORT PLANNING IN CONDITIONS OF DIFFERENT TRANSPORT TARIFFS APPLICATION OF INTEGER PROGRAMMING

TOTAL LOGISTIC MANAGEMENT No. 1 2008 PP. 25 31 Paweł HANCZAR TRANSPORT PLANNING IN CONDITIONS OF DIFFERENT TRANSPORT TARIFFS APPLICATION OF INTEGER PROGRAMMING Abstract: This paper presents the application

### AUTOMATION OF ENERGY DEMAND FORECASTING. Sanzad Siddique, B.S.

AUTOMATION OF ENERGY DEMAND FORECASTING by Sanzad Siddique, B.S. A Thesis submitted to the Faculty of the Graduate School, Marquette University, in Partial Fulfillment of the Requirements for the Degree

### Using LSI for Implementing Document Management Systems Turning unstructured data from a liability to an asset.

White Paper Using LSI for Implementing Document Management Systems Turning unstructured data from a liability to an asset. Using LSI for Implementing Document Management Systems By Mike Harrison, Director,

### Mining Airline Data for CRM Strategies

Proceedings of the 7th WSEAS International Conference on Simulation, Modelling and Optimization, Beijing, China, September 15-17, 2007 345 Mining Airline Data for CRM Strategies LENA MAALOUF, NASHAT MANSOUR

### Factoring Algorithms

Factoring Algorithms The p 1 Method and Quadratic Sieve November 17, 2008 () Factoring Algorithms November 17, 2008 1 / 12 Fermat s factoring method Fermat made the observation that if n has two factors

### Data Clustering. Dec 2nd, 2013 Kyrylo Bessonov

Data Clustering Dec 2nd, 2013 Kyrylo Bessonov Talk outline Introduction to clustering Types of clustering Supervised Unsupervised Similarity measures Main clustering algorithms k-means Hierarchical Main

### Using Artificial Intelligence to Manage Big Data for Litigation

FEBRUARY 3 5, 2015 / THE HILTON NEW YORK Using Artificial Intelligence to Manage Big Data for Litigation Understanding Artificial Intelligence to Make better decisions Improve the process Allay the fear

### Data Mining Techniques in CRM

Data Mining Techniques in CRM Inside Customer Segmentation Konstantinos Tsiptsis CRM 6- Customer Intelligence Expert, Athens, Greece Antonios Chorianopoulos Data Mining Expert, Athens, Greece WILEY A John

### Unsupervised learning: Clustering

Unsupervised learning: Clustering Salissou Moutari Centre for Statistical Science and Operational Research CenSSOR 17 th September 2013 Unsupervised learning: Clustering 1/52 Outline 1 Introduction What

### University of Lille I PC first year list of exercises n 7. Review

University of Lille I PC first year list of exercises n 7 Review Exercise Solve the following systems in 4 different ways (by substitution, by the Gauss method, by inverting the matrix of coefficients