CLUSTERING FOR FORENSIC ANALYSIS



Similar documents
An Approach to Improve Computer Forensic Analysis via Document Clustering Algorithms

IMPROVISATION OF STUDYING COMPUTER BY CLUSTER STRATEGIES

Clustering of Documents for Forensic Analysis

ELEVATING FORENSIC INVESTIGATION SYSTEM FOR FILE CLUSTERING

Data Mining Project Report. Document Clustering. Meryem Uzun-Per

Data Mining Clustering (2) Sheets are based on the those provided by Tan, Steinbach, and Kumar. Introduction to Data Mining

Clustering UE 141 Spring 2013

Clustering. Adrian Groza. Department of Computer Science Technical University of Cluj-Napoca

Chapter ML:XI (continued)

Clustering Technique in Data Mining for Text Documents

DATA MINING TECHNIQUES AND APPLICATIONS

Movie Classification Using k-means and Hierarchical Clustering

Information Retrieval and Web Search Engines

Hierarchical Cluster Analysis Some Basics and Algorithms

Social Media Mining. Data Mining Essentials

Use of Data Mining Techniques to Improve the Effectiveness of Sales and Marketing

Data Mining Cluster Analysis: Basic Concepts and Algorithms. Lecture Notes for Chapter 8. Introduction to Data Mining

Chapter 7. Cluster Analysis

Knowledge Discovery and Data Mining. Structured vs. Non-Structured Data

An Overview of Knowledge Discovery Database and Data mining Techniques

Dynamical Clustering of Personalized Web Search Results

A Complete Gradient Clustering Algorithm for Features Analysis of X-ray Images

Bisecting K-Means for Clustering Web Log data

Hadoop SNS. renren.com. Saturday, December 3, 11

The Data Mining Process

SEARCH ENGINE WITH PARALLEL PROCESSING AND INCREMENTAL K-MEANS FOR FAST SEARCH AND RETRIEVAL

Cluster Analysis. Alison Merikangas Data Analysis Seminar 18 November 2009

Standardization and Its Effects on K-Means Clustering Algorithm

How To Cluster

SPATIAL DATA CLASSIFICATION AND DATA MINING

Comparision of k-means and k-medoids Clustering Algorithms for Big Data Using MapReduce Techniques

Keywords cosine similarity, correlation, standard deviation, page count, Enron dataset

The SPSS TwoStep Cluster Component

Personalized Hierarchical Clustering

SoSe 2014: M-TANI: Big Data Analytics

Data Mining Cluster Analysis: Basic Concepts and Algorithms. Lecture Notes for Chapter 8. Introduction to Data Mining

Machine Learning using MapReduce

FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT MINING SYSTEM

Comparison of K-means and Backpropagation Data Mining Algorithms

Unsupervised learning: Clustering

Using Data Mining for Mobile Communication Clustering and Characterization

DATA MINING CLUSTER ANALYSIS: BASIC CONCEPTS

A Study of Web Log Analysis Using Clustering Techniques

International Journal of Advance Research in Computer Science and Management Studies

PERFORMANCE ANALYSIS OF CLUSTERING ALGORITHMS IN DATA MINING IN WEKA

A FUZZY BASED APPROACH TO TEXT MINING AND DOCUMENT CLUSTERING

There are a number of different methods that can be used to carry out a cluster analysis; these methods can be classified as follows:

Unsupervised Data Mining (Clustering)

Cluster Analysis. Isabel M. Rodrigues. Lisboa, Instituto Superior Técnico

Text Classification Using Symbolic Data Analysis

A comparison of various clustering methods and algorithms in data mining

. Learn the number of classes and the structure of each class using similarity between unlabeled training patterns

Research on Clustering Analysis of Big Data Yuan Yuanming 1, 2, a, Wu Chanle 1, 2

Text Mining in JMP with R Andrew T. Karl, Senior Management Consultant, Adsurgo LLC Heath Rushing, Principal Consultant and Co-Founder, Adsurgo LLC

Big Data & Scripting Part II Streaming Algorithms

Available online at Available online at Advanced in Control Engineering and Information Science

IT services for analyses of various data samples

EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set

COMP3420: Advanced Databases and Data Mining. Classification and prediction: Introduction and Decision Tree Induction

KNIME TUTORIAL. Anna Monreale KDD-Lab, University of Pisa

Neural Networks Lesson 5 - Cluster Analysis

Case study on asset tracing

A Relevant Document Information Clustering Algorithm for Web Search Engine

Applying Data Analysis to Big Data Benchmarks. Jazmine Olinger

How To Cluster Of Complex Systems

How To Solve The Cluster Algorithm

STOCK MARKET TRENDS USING CLUSTER ANALYSIS AND ARIMA MODEL

QOS Based Web Service Ranking Using Fuzzy C-means Clusters

Large-Scale Data Sets Clustering Based on MapReduce and Hadoop

Homework 4 Statistics W4240: Data Mining Columbia University Due Tuesday, October 29 in Class

Unsupervised Learning and Data Mining. Unsupervised Learning and Data Mining. Clustering. Supervised Learning. Supervised Learning

Effective Data Retrieval Mechanism Using AML within the Web Based Join Framework

K-Means Cluster Analysis. Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 1

Clustering. Danilo Croce Web Mining & Retrieval a.a. 2015/201 16/03/2016

Web Document Clustering

A Survey of Classification Techniques in the Area of Big Data.

An Analysis on Density Based Clustering of Multi Dimensional Spatial Data

Statistical Databases and Registers with some datamining

Comparison and Analysis of Various Clustering Methods in Data mining On Education data set Using the weak tool

A Comparative Study of the Pickup Method and its Variations Using a Simulated Hotel Reservation Data

DESKTOP BASED RECOMMENDATION SYSTEM FOR CAMPUS RECRUITMENT USING MAHOUT

Clustering Data Streams

A Regression Approach for Forecasting Vendor Revenue in Telecommunication Industries

Analytics on Big Data

Clustering. Data Mining. Abraham Otero. Data Mining. Agenda

SPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING

Protein Protein Interaction Networks

MIRACLE at VideoCLEF 2008: Classification of Multilingual Speech Transcripts

Example: Document Clustering. Clustering: Definition. Notion of a Cluster can be Ambiguous. Types of Clusterings. Hierarchical Clustering

Hadoop Technology for Flow Analysis of the Internet Traffic

IBM SPSS Direct Marketing 23

STATISTICA. Clustering Techniques. Case Study: Defining Clusters of Shopping Center Patrons. and

Mining Social-Network Graphs

Steven M. Ho!and. Department of Geology, University of Georgia, Athens, GA

Transcription:

IMPACT: International Journal of Research in Engineering & Technology (IMPACT: IJRET) ISSN(E): 2321-8843; ISSN(P): 2347-4599 Vol. 2, Issue 4, Apr 2014, 129-136 Impact Journals CLUSTERING FOR FORENSIC ANALYSIS SAHIL KHENAT, PRATIK KOLHATKAR, SARANG PARIT & SHARDUL JOSHI Department of Computer Engineering, UCOER, Pune University, Pune, Maharashtra, India ABSTRACT In today s digital world, information in computers has great importance and this information is very essential in context for future references and studies irrespective of various fields. So surveying of such information is critical and important task. In computer forensic analysis, a lot of information present in the digital devices (computers in context of our paper) is examined to extract information. And computer consists of hundreds of thousands of files which contain unstructured text or information, so clustering algorithms are of great interest. Clustering helps to improve analysis of documents under consideration. This document clustering analysis is very useful for crime investigations to analyze the information from seized digital devices like computers, laptops, hard disks, tablets. There are total six algorithms used for clustering of documents like K-means, K-medoids, single link, complete link, Average Link, CSPA. These six algorithms are very efficiently used to cluster the digital documents. These algorithms make the analysis very fast by clustering very close documents in one cluster. Also two validity index are used to find out how many clusters are formed. KEYWORDS: Clustering, CSPA, K-Mean, K-Medoids INTRODUCTION Estimates that are proposed by IT IS are digital data density increased 18 times in latest 5 to 6 years. Here in year 2006, data density that is volume which is 161 hexabytes raised to 988 hexabyte in year 2011. And is growth is going exponentially. So as digital world contains very important, complex and unstructured data, clustering algorithms play important role in forensic analysis of such digital documents In particular domain related to paper, hundreds of thousands of documents are examined. And this surveillance which exceeds capabilities of expertise which monitors or analyze such documents. So it is very prime requirement to make data simple and to use some techniques which boosts the analysis of complex, unstructured documents. And here Data mining techniques as well as pattern recognizing techniques are of great importance. Because clusters are going to form according to their pattern also Clustering algorithms are used to explore data without knowledge of data. The main purpose behind why to perform clustering is to group the documents which having some sort of similar data. It is like grouping of similar things together so searching or finding can perform efficiently. The same concept is carried out here in clustering. So that particular cluster Ci consist of documents Di contains some sort of similar content. So that now experts can focus on particular cluster for any document rather than analyze all documents. Also there are in all 6 clustering algorithms namely K-means, K-medoids, Single link, complete link, Average Link, CSPA. With different parameters, when algorithms were executed gives sixteen new instantiations. Silhouette and its simplified version are two relative indexes which are used to finding out number of clusters that will be made after algorithm implementation. Impact Factor(JCC): 1.3268 - This article can be downloaded from www.impactjournals.us

130 Sahil Khenat, Pratik Kolhatkar, Sarang Parit & Shardul Joshi In any digital forensic analysis, number of clusters is very important and critical factor which is mainly unknown and also there is no investigation has been made on cluster numbers. But as clusters are very useful for fast investigation of digital documents, our paper consist of classical clustering algorithms also newest researches like consensus partitions. Here our paper is organized in following main points: Introduction Classical Clustering Algorithms and Pre-processing Steps Proposed System Limitations Conclusion References CLASSICAL CLUSTERING ALGORITHMS AND PRE-PROCESSING STEPS The important step before running the clustering algorithms is the preprocessing. Pre-processing consisting of stopword removal and stemming. Stopword removal is process in which particular words are removed which will not affect the meaning of document. This consists of removal of prepositions, pronouns, articles and other irrespective words. After preprocessing, the main statistical approach for text mining is adopted. In this Statistical approach, document represented in vector model format, each document is represented as vector model consisting of words according to their frequencies of occurrence. Reduction technique known as Term Variance (TV) is also used to increase efficiency of clustering algorithms. As clusters are formed which containing documents, TV are used to estimate top n words which have greatest occurrences over documents within clusters. So that this is very important factor for formation of cluster. Also it is important to find out distances between two documents when they are resides in different clusters. And for finding out distances between them, cosine-based distance and Levenshtein -based distance As to find out number of clusters, by looking for best set of partitions of data set from different clusters which gives best result by relative index or different combinations of attributes are selected and after various runs(k-mean), best is selected. So by considering that, there are various sets of data partitions with different clusters, from which we have to choose best one. And one of the ways to finding out best partition from sets is Levenshtein-based algorithms so it is a important component in our studies. Suppose there is object x in cluster A. And dissimilarity of x with other objects in same luster that is A is a (i). Now consider cluster C. And average dissimilarity of object x to cluster C is d(x, C). As we have to find o dissimilarities within neighbors clusters following technique is used. After computing the dissimilarity d(x, C) with all clusters except A, smallest one is selected that is b(x). Value of dissimilarity neighbors is finding out by formula S(x) = b(x)-a(x)/ max {a(x), b(x)} Index Copernicus Value: 3.0 - Articles can be sent to editor@impactjournals.us

Clustering for Forensic Analysis 131 Value S(x) is verified in between age of -1 to 1. If value of S(x) is higher, object x belongs to particular cluster but if S(x) is zero, then it is not clear that whether object belongs to current cluster or adjacent one. S(x) is carried out over i=1, 2,,n where n are no of objects and then average is computed. And best clustering is maximum S(x). Hence to finding out effective S(x), i.e. called simplified Silhouette, one can compute only the distances among the objects and the centroids of the clusters. So a(x) is dissimilarity correspond to or simply belongs to cluster (A) centroid. Now it is very easy to get only one distance rather than finding out whole dissimilarity between all objects within cluster. Also rather than finding out d (x C). C not equal to a, we will find only distance with centroid. Table 1: Information Found in Clusters Cluster Information C1 6 LIC policies C2 2 bank accounts C3 5 grocery list 1 loan agreement C4 2 check receipt C5 5 office applications C6 3 financial transactions C7 5 investment club status PROPOSED SYSTEM Consider one folder as an input. It contains six files. Our software will take one file at a time. And then it will go under following processes: Consider one folder as an input. It contains six files. Our software will take 1 st file at a time. And then it will go under following processes. Pre-Processing Module o o Fetch the Content of File: This will be our 1 st sub step in pre-processing module in which content of our input file will be fetched by our software for further processing. Stopword Removing: Fetched content of input file contains lot of stopwords i.e. the words which don t have important meaning. For example, suppose our input file contains sentence as Here we are learning java. In this sentence, we have words like here, we, are which are not important for further processing i.e. stemming. So we will remove those words from our original sentence and we just pass words like java, learning to further step i.e. stemming. o Stemming: This is the step where we bring down the word to its original base form. Consider the same example of sentence. Example- Here we are learning java. In this sentence, after removing of stopwords, we get words learning, java for stemming. Impact Factor(JCC): 1.3268 - This article can be downloaded from www.impactjournals.us

132 Sahil Khenat, Pratik Kolhatkar, Sarang Parit & Shardul Joshi In stemming, we will bring the word learning to its base form as to learn. For this, we use algorithms Port stemmer, but problem with this already existing algorithm is it don t return some words to its bas form with correct spelling or with correct meaning. For Example: The word Studied, It returns to studi, and not To study. And also words like String, which are already in base form containing -ing. That after removing -ing, words become meaningless. Preparing Cluster Vector From first module, we get pre-processed content of the input. We will calculate weight i.e. frequency (no of occurrences) of each word. And we will arrange these words in descending order according to their weights. Out of them, we take certain number of words as Top n words. And we will maintain its array, say Array A. In second module, usually the first sentence is the most important and which is probably it is most suitable s the title of the document. In such a way every document has its first sentence as its title which results in formation of Array B. Our input document may contain few numerical values which can provide us very important information, so we will gather those numerical values and we will maintain its different array, Array C. Now, we have three different arrays A, B, C of Top n words, title, numerical values respectively. So we will combine all three arrays and we will form one master vector Mv. Likewise we will create master vectors for all the remaining files. Forensic Analysis Now, we have master vectors for all the files in that particular input folder. So, here we going to compare all this master vectors with each others to find out the similarities between master vectors on the basis of accuracy. This process is named as mapping of master vectors. The accuracy is given by user according to their requirement. For example, consider file folder consisting of 5 text files, 2 image files, 3 video files. File formats other than text files will get filtered out. Only.txt,.doc,.docx files are considered for analysis of documents. Therefore 5 text files will have their respective 5 master vectors {MV1, MV2, MV3, MV4, MV5}. Now we apply the process of mapping on these master vectors. User will enter accuracy on the basis of which the further process will be executed. Constraint of accuracy is that 0<accuracy<1. Table 2: Statistical Mapping of Clusters MV1 MV2 MV3 MV4 MV5 MV1 0 0.56 0.98 0.45 0.3 MV2 0.56 0 0.63 0.52 0.41 MV3 0.98 0.63 0 0.75 0.18 MV4 0.45 0.52 0.75 0 0.23 MV5 0.3 0.41 0.18 0.23 0 Suppose, user enters accuracy as 0.5. Then all master vectors having accuracy quotient equal to or more than 0.5 will be combined together to form revised master vector (R.M.V.). So, in this case revised master vectors will be formed in following manner. Index Copernicus Value: 3.0 - Articles can be sent to editor@impactjournals.us

Clustering for Forensic Analysis 133 MV1 and MV2 RMV12 MV1 and MV3 RMV13 MV2 and MV3 RMV23 MV2 and MV4 RMV24 MV3 and MV4 RMV34 In such a way, we get five revised master vectors which contain data depending on accuracy entered by user. Note: Why accuracy should not be 1? If user enters accuracy as 1, then that means while mapping master vectors of files data from those master vectors should be exactly same, which is not practical and it is inefficient. Due to this, our system will skip very crucial information might help in investigation and carries important data. Algorithm for Cluster Vector Creation CL = [D, S, SVSM, Sim (d, s i ), C] Input D = {d 1, d 2 ; : : : ; d D } Where set of input documents to be clustered S = {s 1 ; s 2 ; : : : ; s n } set of n subjects, where each subject si is a set of weighted terms Sim (d, s i ) = similarity function C = {C 1, C 2 C n } Where c is set of n overlapping output clusters Cluster Vector T w =Term weight D = {d 0, d 1, d 2,.,..,..d n } Where, D=Document Set d 0, d 1,.,d n =words of documents N d =Numerical data Output V c = cluster vector Impact Factor(JCC): 1.3268 - This article can be downloaded from www.impactjournals.us

134 Sahil Khenat, Pratik Kolhatkar, Sarang Parit & Shardul Joshi Algorithm Step 0: Get T w and D Step 1: Sort T 100 in descending order Step 2: Add top 100 words to T 100 Step 3: Fetch 1st sentence as T t Step 4: For each d i Step 5: Check for numeric data Step 6: Then add into N d = {n 1, n 2, n 3,.,n n } Step 7: Merge T 100, T t, and N d Step 8: V c = {T 100, T t, N d } Algorithm for Forensic Analysis Input V c = Cluster Vector Output F r = {c 1, c 2, c 3,.,c n } Algorithm Step 0: Get V c = {T t, T 100, N d } Step 1: Get T t Step 2: for each sentence s i Step 3: Find title word of T t Step 4: Rank top 5 sentences Step 5: Get T 100 Step 6: For each sentence s i Step 7: Find frequency of T 100 words Step 8: Rank top 5 sentences R 100 Step 9: Get N d Step 10: for each sentence s i Step 11: Find N d in each sentences Step 12: Rank top 5 sentences, R nd Index Copernicus Value: 3.0 - Articles can be sent to editor@impactjournals.us

Clustering for Forensic Analysis 135 Step 13: Merge R t, R 100, R nd Step 14: Remove repeated sentences to get important (Imp) vectors Step 15: for each document Step 16: Compare Imp vector Step 17: If comparison is more than 50% then cluster the documents Step 18: Stop LIMITATIONS It is known that very well, that working of any clustering algorithm depends on data, but for estimated datasets some of our versions of clustering algorithms have shown good results. One of the prominent issues is scalability. In order to handle this issue, sampling and other techniques can be used. Also algorithms like bisecting k-means and associated approaches can be used. These algorithms can persuade dendograms. They have a similar inductive bias with respect to the various hierarchical methods in our work. More precisely, aiming at evading the computational difficulties, partitional clustering algorithms can be used for computing a hierarchical clustering solution by using repeated cluster bisectioning techniques. For illustration, bisecting k-means has comparatively low computational requirements. i.e. it is O(N.log N), versus the overall time complexity O(N^2.logN) for specially agglomerative methods. If the number of documents is extremely high for running an agglomerative algorithm, in that case, bisecting k-means and other related approaches can be used as a solution. When we consider the cost of computation for estimating the number of clusters, the silhouette introduced in [1] is based mainly on all the distances between objects. This leads to a computational cost of O (N^2.D), where N is number of objects in dataset and D is number of attributes. As a solution to this potential difficulty, especially when very large size datasets come into the picture, a simplified version of silhouette can be used. Simplified silhouette includes computation of distance between objects and cluster centroids. This reduces computational cost from O (N^2.D) to O (k.n.d) where k is the number of clusters. Here value of k is very less than value of N. In our work we have adopted silhouettes. Instead there are several relative criteria which can be used as a replacement or alternative for it. Such criteria have abilities that make each of them to outperform others in particular classes of problems. Practically, one can use or test different criteria to compute number of clusters by considering both the quality of data partitions and respective computational cost. We would also like to mention that practically it is not that important to use scalable methods. In our case, there are no severe restrictions to get data partitions. Instead of that, domain experts can utilize a lot of time to analyze their input data before reaching up to a particular conclusion. CONCLUSIONS We have introduced an approach which can become an ideal application for document clustering to forensic analysis of computers, laptops and hard disks which are seized from criminals during investigation of police. There are several practical results based on our work which are extremely useful for the experts working in forensic computing department. In our work, the algorithms known as Average Link and Complete Link yielded the best results. In spite of these algorithms having high computational costs, they are suitable for our work domain because dendrograms provides a Impact Factor(JCC): 1.3268 - This article can be downloaded from www.impactjournals.us

136 Sahil Khenat, Pratik Kolhatkar, Sarang Parit & Shardul Joshi neat summary of documents which are being inspected. All the textual documents are scanned thoroughly and corresponding output is given. When proper initialization is done, the partitional K-means and K-medoids algorithms also have satisfactory results. When estimation of number of clusters is to be done and approaches for doing the same are considered, at that time the simplified version of silhouette in less efficient as compared to its relative validity criterion which is far more accurate than simpler version. Additionally, in some results it was observed that making utilization of the file names along with the actual document information or content would prove to be very useful for document ensemble algorithms. Another important observation which we came across was that clustering algorithms tend to bring about clusters formed by either relevant or irrelevant document set. This leads to enhancement of expert examiner s job or task. Further ahead, our proposed approach has the capacity to boost the speed of computer inspection by an impressive factor. REFERENCES 1. Document Clustering for Forensic Analysis: An Approach for Improving Computer Inspection Luís Filipe da Cruz Nassif and Eduardo Raul Hruschka. 2. L. Kaufman and P. Rousseeuw, Finding Groups in Gata: An Introduction to Cluster Analysis. Hoboken, NJ: Wiley-Interscience, 1990. 3. B. S. Everitt, S. Landau, and M. Leese, Cluster Analysis. London, U.K.: Arnold, 2001. 4. B. Mirkin, Clustering for Data Mining: A Data Recovery Approach. London, U.K.: Chapman & Hall, 2005. Index Copernicus Value: 3.0 - Articles can be sent to editor@impactjournals.us