1. What are the uses of statistics in data mining? Statistics is used to Estimate the complexity of a data mining problem. Suggest which data mining



Similar documents
Data Mining: Concepts and Techniques. Jiawei Han. Micheline Kamber. Simon Fräser University К MORGAN KAUFMANN PUBLISHERS. AN IMPRINT OF Elsevier

DATA WAREHOUSING AND OLAP TECHNOLOGY

Introduction to Data Mining

OLAP and Data Mining. Data Warehousing and End-User Access Tools. Introducing OLAP. Introducing OLAP

Introduction to Data Mining

Chapter 5. Warehousing, Data Acquisition, Data. Visualization

Course DSS. Business Intelligence: Data Warehousing, Data Acquisition, Data Mining, Business Analytics, and Visualization

Introduction. A. Bellaachia Page: 1

An Overview of Knowledge Discovery Database and Data mining Techniques

Chapter 5 Business Intelligence: Data Warehousing, Data Acquisition, Data Mining, Business Analytics, and Visualization

Fluency With Information Technology CSE100/IMT100

International Journal of Computer Science Trends and Technology (IJCST) Volume 2 Issue 3, May-Jun 2014

CS1011: DATA WAREHOUSING AND MINING TWO MARKS QUESTIONS AND ANSWERS

DATA WAREHOUSE E KNOWLEDGE DISCOVERY

Information Management course

DATA MINING CONCEPTS AND TECHNIQUES. Marek Maurizio E-commerce, winter 2011

IMPROVING DATA INTEGRATION FOR DATA WAREHOUSE: A DATA MINING APPROACH

International Journal of Scientific & Engineering Research, Volume 5, Issue 4, April ISSN

CHAPTER-24 Mining Spatial Databases

Copyright 2007 Ramez Elmasri and Shamkant B. Navathe. Slide 29-1

Data Warehousing and Data Mining

DATA MINING TECHNIQUES AND APPLICATIONS

Data Mining System, Functionalities and Applications: A Radical Review

Search and Data Mining: Techniques. Applications Anya Yarygina Boris Novikov

Data Warehouse: Introduction

Foundations of Business Intelligence: Databases and Information Management

Building Data Cubes and Mining Them. Jelena Jovanovic

Data Warehousing and Data Mining in Business Applications

Data Mining Introduction

Data Mining. Vera Goebel. Department of Informatics, University of Oslo

Data Mining Solutions for the Business Environment

Importance or the Role of Data Warehousing and Data Mining in Business Applications

Chapter 20: Data Analysis

An Overview of Database management System, Data warehousing and Data Mining

Sanjeev Kumar. contribute

Principles of Data Mining by Hand&Mannila&Smyth

Mario Guarracino. Data warehousing

Overview. Background. Data Mining Analytics for Business Intelligence and Decision Support

Foundations of Business Intelligence: Databases and Information Management

Data Warehousing and Data Mining

Data Mining as Part of Knowledge Discovery in Databases (KDD)

Data Mining and Database Systems: Where is the Intersection?

Data Mining - Introduction

Social Media Mining. Data Mining Essentials

Chapter 6 FOUNDATIONS OF BUSINESS INTELLIGENCE: DATABASES AND INFORMATION MANAGEMENT Learning Objectives

(b) How data mining is different from knowledge discovery in databases (KDD)? Explain.

The basic data mining algorithms introduced may be enhanced in a number of ways.

SPATIAL DATA CLASSIFICATION AND DATA MINING

Data Mining Analytics for Business Intelligence and Decision Support

The Data Mining Process

A Review of Data Mining Techniques

Knowledge Discovery and Data Mining. Structured vs. Non-Structured Data

Adding New Level in KDD to Make the Web Usage Mining More Efficient. Abstract. 1. Introduction [1]. 1/10

COURSE RECOMMENDER SYSTEM IN E-LEARNING

Course MIS. Foundations of Business Intelligence

Data Warehousing and OLAP Technology for Knowledge Discovery

Oracle9i Data Warehouse Review. Robert F. Edwards Dulcian, Inc.

BUILDING BLOCKS OF DATAWAREHOUSE. G.Lakshmi Priya & Razia Sultana.A Assistant Professor/IT

Data Mining and Knowledge Discovery in Databases (KDD) State of the Art. Prof. Dr. T. Nouri Computer Science Department FHNW Switzerland

1. OLAP is an acronym for a. Online Analytical Processing b. Online Analysis Process c. Online Arithmetic Processing d. Object Linking and Processing

Classification algorithm in Data mining: An Overview

Data Mining Part 5. Prediction

Introduction to Data Mining and Machine Learning Techniques. Iza Moise, Evangelos Pournaras, Dirk Helbing

Customer Classification And Prediction Based On Data Mining Technique

Distance Learning and Examining Systems

Data Mining + Business Intelligence. Integration, Design and Implementation

not possible or was possible at a high cost for collecting the data.

Data Mining: Overview. What is Data Mining?

COMP3420: Advanced Databases and Data Mining. Classification and prediction: Introduction and Decision Tree Induction

CHAPTER 1 INTRODUCTION

Example application (1) Telecommunication. Lecture 1: Data Mining Overview and Process. Example application (2) Health

Week 3 lecture slides

Data Warehousing and Data Mining. A.A Datawarehousing & Datamining 1

Classification and Prediction

OLAP and OLTP. AMIT KUMAR BINDAL Associate Professor M M U MULLANA

Foundations of Business Intelligence: Databases and Information Management

Data Preprocessing. Week 2

B.Sc (Computer Science) Database Management Systems UNIT-V

Data Mining for Successful Healthcare Organizations

Chapter 2 Literature Review

PRACTICAL DATA MINING IN A LARGE UTILITY COMPANY

Data Mining for Knowledge Management. Classification

Data Mining. Knowledge Discovery, Data Warehousing and Machine Learning Final remarks. Lecturer: JERZY STEFANOWSKI

Statistical Models in Data Mining

Alexander Nikov. 5. Database Systems and Managing Data Resources. Learning Objectives. RR Donnelley Tries to Master Its Data

Databases in Organizations

When to consider OLAP?

Published by: PIONEER RESEARCH & DEVELOPMENT GROUP ( 28

John R. Vacca INSIDE

Data Warehouse Snowflake Design and Performance Considerations in Business Analytics

The University of Jordan

14. Data Warehousing & Data Mining

Foundations of Business Intelligence: Databases and Information Management

New Approach of Computing Data Cubes in Data Warehousing

Delivering Business Intelligence With Microsoft SQL Server 2005 or 2008 HDT922 Five Days

Data Mining for Retail Website Design and Enhanced Marketing

Data Mining Algorithms Part 1. Dejan Sarka

Chapter 6. Foundations of Business Intelligence: Databases and Information Management

Big Data: Rethinking Text Visualization

Unsupervised Data Mining (Clustering)

Transcription:

1. What are the uses of statistics in data mining? Statistics is used to Estimate the complexity of a data mining problem. Suggest which data mining techniques are most likely to be successful, and Identify data fields that contain the most surface information. 2. What is the main goal of statistics? The basic goal of statistics is to extend knowledge about a subset of a collection to the entire collection. 3. What are the factors to be considered while selecting the sample in statistics? The sample should be Large enough to be representative of the population. Small enough to be manageable. Accessible to the sampler. Free of bias. 4. Name some advanced database systems? Object-oriented databases. Object-relational databases. 5. Name some specific application oriented databases? Spatial databases. Time-series databases. Text databases Multimedia databases. 6. Define Relational databases? Relational databases are a collection of tables, each of which is assigned a unique name. Each table consists of a set of attributes (columns or fields) and usually stores a large set of tuples (rows or records). Each tuple in a relational table represents an object identified by a unique key and described by a set of attribute values. 7. Define Transactional databases? Transactional databases consist of a file where each record represents a transaction. A transaction typically includes a unique transaction identity number (trans_id), and a list of the items making up the transaction. 1

8. Define Spatial Databases? Spatial databases contain spatial-related information. Such databases include geographic (map) databases, VLSI chip design databases, and medical and satellite image databases. Spatial data may be represented in raster format, consisting of n-dimensional bit maps or pixel maps. 9. What is Temporal Database? Temporal database store time related data. It usually stores relational data that include time related attributes. These attributes may involve several time stamps, each having different semantics. 10. What is a Time-Series database? A Time-Series database stores sequences of values that change with time, such as data collected regarding the stock exchange. 11. What is Legacy database? A Legacy database is a group of heterogeneous databases that combines different kinds of data systems such as relational or objects oriented databases, hierarchical databases, network databases, and spread sheets, multimedia databases or file systems. 12. What are the steps in the data mining process? Data Cleaning Data Integration Data Selection Data Transformation Data Mining Pattern Evaluation Knowledge Representation 13. Define data cleaning? Data Cleaning means removing the inconsistent data or noise and collecting necessary information. 14. Define data mining? Data mining is a process of extracting or mining knowledge from huge amount of data. 15. Define pattern evaluation? Pattern evaluation is used to identify the truly interesting patterns representing knowledge based on some interesting measures. 2

16. Define Knowledge representation? Knowledge representation techniques are used to present the mined knowledge to the user. 17. Define class/ concept description? Data can be associated with classes or concepts. It can be useful to describe individual classes and concepts in summarized, concise and yet precise terms. Such description of a class or a concept is called class/ concept descriptions. 18. What is Data Characterization? Data Characterization is a summarization of the general characteristics or features of a target class of data. The data corresponding to the user specified class or typically collected by a database query. 19. What is data discrimination? Data discrimination is a comparison of the general features of target class data objects with the general features of objects from one or a set of contrasting classes. 20. What is Association analysis? Association analysis is the discovery of association rules showing attribute-value conditions that occur frequently together in a given set of data. 21. Define association rules? Association rules are of the form X Y, that is A1.. Am B1 Bn, where A i (for i {1,.,m}) and B j (for j {1,,n}) are attribute- value pairs. The association rule X Y is interpreted as database tuples that satisfy the condition in X are also likely to satisfy the conditions in Y. 22. List out the major components of a typical data mining system? The major components in the typical data mining system architecture are Database, Data warehouse, World Wide Web or other information repositories. Database or data warehouse server Knowledge base Data mining engine Pattern evaluation module User interface 3

23. How does a data warehouse differ from a database? How are they similar? Difference: A database system or DBMS, consists of a collection of interrelated data, know as a database and a set of software programs to manage and access the data. A data warehouse is a repository of information collected from multiple sources, stored under a unified schema. Similarity: Queries can be applied to both database and data warehouse. A data warehouse is modeled by multidimensional database structure. 24. What is concept description of hierarchies? Concept description generates description for the characterization and comparison of the data. It is some times called class description, when the concept to be described refers to a class of objects. 25. What is constraint based association mining? Specification of constraints or expectations to confine the search space of database mining process called constraint based association mining. The constraints can be, Knowledge based constraints Data constraints Dimension/level constraints Interestingness constraints Rule constraints. 26. What is linear regression? Linear regression involves finding the best line to fit two attributes, so that one attribute can be used to predict the other. Example: A random variable y (response variable) can be modeled as a linear function of another random variable x (predictor variable).with the equation y = wx+b. 27. What are the two data structures in cluster analysis? Two data structures in cluster analysis are, Data matrix (object by variable structure) Dissimilarity matrix (object by object structure) 4

28. How are concept hierarchies useful in OLAP? In the multidimensional model, data are organized onto multiple dimensions, and each dimension contains multiple levels of abstraction defined by concept hierarchies. This organization provides users with a flexibility to view data from different perspectives. LAP provides a user-friendly environment for interactive data analysis. 29. What do you mean by virtual warehouse? A virtual warehouse is a set of views over operational databases. For effective query processing only some of the possible summary views may be materialized. A virtual warehouse is easy to build but requires excess capacity on operational database servers. 30. List out five data mining tools IBM S Intelligent miner Data mined corporation Data mined Pilots Discovery server Tools from business objects and SAS Institute End user tools. 31. What is KDD? Knowledge discovery is a process and consists of an iterative sequence of the following steps. Data cleaning Data Integration Data Selection Data transformation Data Mining Pattern evaluation Knowledge presentation 32. List out the classification of data mining system? Classification according to the kinds of databases mined. Classification according to the kinds of knowledge mined. Classification according to the techniques utilized. Classification according to the application adapted. 5

33. What is concept description? Concept description is a form of data generalization. A concept typically refers to a collection of data such as frequent-buyers, graduate-students etc. Concept description generates descriptions for the characterization and comparison of the data. 34. What is association rule mining? It consists of first finding frequent item sets (set of items, such as A and B, satisfies a minimum support threshold, or percentage of the task-relevant tuples) from which strong association rules in the form of A=>B are generated. The rules also satisfy a minimum confidence threshold. Associative can be further analyzed to uncover correlation rules. 35. What is tree pruning? Tree pruning is used to remove the anomalies in the training data due to noise outliers. It addresses the problem of overfilling the data. Two approaches of tree pruning. Pre-pruning --Tree is pruned by halting its construction early. Post-pruning--Removes sub trees from a fully grown tree. 36. What is cluster analysis? The process of grouping a set of physical or abstract object into classes of similar objects is called clustering. A cluster is a collection of data objects that are similar to one another within the same cluster and are dissimilar to the objects in other clusters. 37. What is concept hierarchy? Concept hierarchy defines a sequence of mapping from a set of low-level concepts to higher-level, more general concepts. Concept hierarchies are implicit within the database schema. A concept hierarchy is a total or partial order among attribute in a database schema is called schema hierarchy. 38. What is Aggregation and metadata? Aggregation, where summary or aggregation operations are applied to the data. For example, the daily sales data may be aggregated so as to compute monthly and annual total amount. This is used in constructing a data cube for analysis of data at multiple granularities. Metadata are data about data which define warehouse objects. Metadata are created for the data names and definition of the given warehouse. 6

39. What is star schema and snow flake schema? Star schema is the most common modeling paradigm in which the data warehouse contains A large central table (fact table) containing the bulk of data with no redundancy. A set of smaller attendant tables, one for each dimension. Snowflake schema is a variant of the star schema model where some dimension tables are normalized; thereby further splitting the data into additional tables. 40. Write short notes on spatial clustering? Spatial data clustering identifies clusters or densely populated regions, according to some distance measurements in a large, multidimensional data set. 41. State the types of Linear Model and state its use? Generalized Linear model represent the theoretical foundation on which linear regression can be applied to the modeling of categorical response variables. The types of generalized linear model are Logistic regression Poisson regression 42. What are the goals of Time series analysis? Finding patterns in the data Predicting future values. 43. What is smoothing? Smoothing is an approach that is used to remove nonsystematic behaviors found in a time series. It can be used to detect trends in time series. 44. What is Lag? The time difference between related items is referred to as Lag. 45. Write the preprocessing steps that may be applied to the data for classification and prediction? Data cleaning Relevance analysis Data transformation 46. Define Data Classification? It is a two-step process. In the first step, a model is built describing a predetermined set of data classes or concepts. The model is constructed by analyzing 7

database tuples described by attributes. In the second step, the model is used for classification. 47. What are Bayesian Classifiers? Bayesian Classifiers are statistical classifiers. They can predict class membership probabilities, such as the probability that a given sample belongs to a particular class. 48. What is a decision tree? It is a flowchart like tree structure, where each internal node denotes a test on an attribute, each branch represents an outcome of the test, and leaf node represents classes or class distribution. Decision tree is a predictive model. Each branch of the tree is a classification question and leaves of the tree are partition of the data set with their classification. 49. Where are Decision Trees mainly used? Used for exploration of data set and business problems Data preprocessing for other predictive analysis Statisticians use decision trees for exploratory analysis. 50. How will you solve a classification problem using Decision Tree? Decision Tree Induction: Construct a decision tree using training data. For each ti D apply the decision tree to determine its class ti-tuple D-Database 51. How is association rules mined from large databases? Association rule mining is a two step process. Find all frequent itemsets. Generate strong association rules from the frequent itemsets. 52. What is the classification of association rules based on various criteria? 1. Based on the types of values handled in the rule a. Boolean association rule b. Quantitative association rule. 2. Based on the dimensions of data involved in the rule a. Single dimensional association rule b. Multidimensional association rule 3. Based on the levels of abstractions involved in the rule 8

a. Single level association rule b. Multilevel association rule 4. Based on various extensions to association mining a. Maxpatterns b. Frequent closed itemsets 53. What is Apriori algorithm? Apriori algorithm is an influential algorithm for mining frequent item sets for Boolean association rules using prior knowledge. Apriori algorithm uses prior knowledge of frequent itemset properties and it employees an iterative approach known as level-wise search where k-itemsets are used to explore (k+1)-itemsets. 54. Define a Data mart? Data mart is a pragmatic collection of related facts, but does not have to be exhaustive or exclusive. A data mart is both a kind of subject area and an application. Datamart is a collection of numeric facts. 55. What is data warehouse performance issue? The performance of data warehouse is largely a function of the quantity and the type of data stored within a database and the query/data loading work load placed upon the system. 56. What is Data Inconsistency Cleaning? This can be summarized as the process of cleaning up the small inconsistencies that introduce themselves into the data. Examples include duplicate keys and unreferenced foreign keys. 57. Merits of Data warehouse. * Ability to make effective decisions from database * Better Analysis of data and decision support * Discover trends and correlations that benefits business * Handle huge amount of data 58. What are the characteristics of data warehouse? * Separate * Available * Integrated * Subject oriented * Not dynamic * Consistency 9

* Iterative Development * Aggregation Performance 59. List some of the data warehouse tools. * OLAP (Online Analytic Processing) * ROLAP (Relational OLAP) * End User Data Access Tool * Ad Hoc Query Tool * Data Transformation Services * Replication 60. Explain OLAP. The general activity of querying and presenting text and number data from data warehouses, as well as a specifically dimensional style of querying and presenting that is exemplified by a number of "OLAP Vendors". The OLAP vendors technology is nonrelational and is almost always biased on an explicit multidimensional cube of data. OLAP databases are also known as multidimensional cube of databases. 61. Explain ROLAP. ROLAP is a set of user interfaces and applications that give a relational database, a dimensional flavor. ROLAP stands for Relational Online Analytic Processing. 62. Explain End User Data Access Tool? End User Data Access Tool is a client of the data warehouse. In a relational data warehouse, such as client maintains a session with the presentation server, sending a stream of separate SQL requests to the server. Eventually the End User Data Access Tool is done with the SQL session and turns around to present a screen of data or a report, a graph, or some other higher form of analysis to the user. An End User Data Access Tool can be as simple as an Ad Hoc Query Tool or can be complex as a sophisticated data mining or modeling application. 63. Explain Ad Hoc Query Tool? It is a specific kind of end user data access tool that invites the user to form their own queries by directly manipulating relational tables and their joins. Ad Hoc Query Tools, as powerful as they are, can only be effectively used and understood by about 10% of all the potential end users of a data warehouse. 64. Name some of the data mining applications. * Data mining for biomedical and DNA Data Analysis * Data Mining for Financial Data Analysis 10

* Data Mining for the Retail Industry * Data Mining for the Telecommunication Industry 65. What are the contributions of Data Mining to DNA Analysis? * Semantic Integration of heterogeneous, distributed genome databases * Similarity Search and Comparison among DNA Sequences * Association Analysis: identification of co-occurring gene sequences * Path Analysis: Linking genes to different stages of disease development * Visualization Tools and genetic data analysis 66. Name some examples of Data Mining in Retail Industry. * Design and Construction of Data Warehouses based on the benefits of Data Mining * Multidimensional Analysis of sales, customers, products, time and region * Analysis of the effectiveness of sales campaigns * Customer retention analysis of customer loyalty * Purchase recommendation and cross-reference of item 67. What is the difference between "supervised" and "unsupervised" learning scheme? In Data Mining during classification the class label of each training sample is provided, this type of training is called "supervised learning" i.e., the learning of the model is supervised in that it is told to which class each training sample belongs. E.g. Classification In unsupervised learning the class label of each training sample is not known and the member or set of classes to be learned may not be known in advance. E.g. Clustering 68. Discuss the importance of similarity metric clustering? Why is it difficult to handle categorical data for clustering? The process of grouping a set of physical or abstract objects into classes of similar objects is called "clustering". Similarity metric is important because it is used for outlier detection. The clustering algorithm which is main memory based can operate only on the following two data structures namely, a) Data Matrix b) Dissimilarity Matrix So it is difficult to handle categorical data. 11

69. Mention at least 3 advantages of Bayesian Networks for data analysis. Explain each one a) Bayesian Network is a graphical representation of unknown knowledge that is easy to construct and interpret. b) The representation has formal probabilistic semantics, making it suitable for statistical manipulation c) The representation is used for encoding uncertain expert knowledge in expert systems. 70. Why do we need to prune a decision tree? Why should we use a separate pruning data set instead of pruning the tree with the training database? When a decision tree is built, many of the branches will reflect animation in the training data due to noise or outliers. Tree pruning methods are needed to address this problem of over fitting the data. 71. Explain the various OLAP operations? a) Roll-up: The roll up operation performs aggregation on a data cube, either by climbing up a concept hierarchy for a dimension. b) Drill-down: It is the reverse of roll up. It navigates from less detailed data to more detailed data. c) Slice: Performs a selection on one dimension of the given cube, resulting in a sub cube. 72. Discuss the concepts of frequent itemset, support & confidence? A set of items is referred to as itemset. An itemset that contains k items is called k-itemset. An itemset that satisfies minimum support is referred to as frequent itemset. Support is the ratio of the number of transactions that include all items in the antecedent and consequent parts of the rule to the total number of transactions. Confidence is the ratio of the number of transactions that include all items in the consequent as well as antecedent to the number of transactions that include all items in antecedent. 73. Why is data quality so important in a data warehouse environment? Data quality is important in a data warehouse environment to facilitate decision- making. In order to support decision-making, the stored data should provide information from a historical perspective and in a summarized manner. 12

74. How can data visualization help in decision-making? Data visualization helps the analyst gain intuition about the data being observed. Visualization applications frequently assists the analyst in selecting display formats, viewer perspective and data representation schemas that faster deep intuitive understanding thus facilitating decision-making. 75. What do you mean by high performance data mining? Data mining refers to extracting or mining knowledge. It involves an integration of techniques from multiple disciplines like database technology, statistics, machine learning, neural networks, etc. when it involves techniques from high performance computing it is referred as high performance data mining. 76. What are the merits Of Data Warehouse? The merits of data warehouse are the following Ability to make effective decisions form the database. To discover trends and correlations as they provide benefit to the business. Better analysis of data and decision support. It leads to better understanding of the business and handle huge amount of data. There is a possibility of the customer being served better. Better understanding of the business risks. Improvement of the business process. Being able to make tailor made products and services. 77. What are the merits of spatial Data Warehouse? The merits of spatial data warehouse are the following Make dynamic geographic queries on data. To aggrgate your data to geographic areas. To analyse data and spatial reorganization of it. Visualization and presentation of data. 78. Describe the two common approaches of Tree Pruning? In the pre pruning approach a tree is pruned by halting its construction early. The second approach, post pruning, removes branches from a fully grown tree. A tree node is pruned by removing its branches. 13

79. What is clustering? Clustering is the process of grouping the data into classes or clusters so that objects within a cluster have high similarity in comparison to one another, but are very dissimilar to objects in other clusters. 80. What are the requirements of clustering? Scalability Ability to deal with different types of attributes Ability to deal with noisy data Minimal requirements for domain knowledge to determine input parameters Constraint based clustering Interpretability and usability 81. State the categories of clustering methods? Partitioning methods Hierarchical methods Density based methods Grid based methods Model based methods 82. Differentiate between lazy learner and eager learner? Nearest neighbor classifiers are lazy learners in that they store all of the training samples and do not build a classifier until a new (unlabeled) sample needs to be classified. In eager learning methods such as decision tree induction, back propagation constructs a generalized model before receiving a new sample to classify. 83. What is network pruning? The first step forwards extracting rules from neural networks pruning. This consists of removing weighted links that do not result in a decrease in the classification accuracy of the given network. 84. List the various criteria of classification in data mining system? Kinds of databases mined Kinds of knowledge mined Kinds of techniques utilized Application adapted 14

85. Name some data mining techniques? Statistics Machine learning Decision trees Hidden markov model Artificial neural networks Genetic algorithms Meta learning 86. Explain DBMiner tool in data mining? System Architecture Input and Output Data mining tasks supported by the system Support for task and method selection Support of the KDD process Main applications Current status 87. Define Iceberg query? It computes an aggregate function over an attribute or set of attributes in order to find aggregate values above some specified threshold. Given relation R with attributes a1,a2,..an and b, and an aggregate function, agg_f, an iceberg query is the form Select R.a1,R.a2,.,R.an, agg_f(r.b) from relation R group by R.a1,R.a2,.,R.an having agg_f(r.b)>= threshold 88. Define DBMiner? DBMiner is an Online Analytical Mining System, developed fro interactive mining of multiple level knowledge in large relational databases and data warehouses. 89. List out the DBMiner tasks? OLAP analyzer Association Classification Clustering Prediction Time series analysis. 15

90. Explain how data mining is used in Health care analysis? Healthcare data mining and its aims Healthcare data mining technique Segmenting patients into groups Identifying patients with recurring health problems Relation between disease and symptoms Curbing the treatment costs Predicting medical diagnosis Medical research Hospital administration Applications of data mining in Healthcare 91. Explain Data mining applications for financial data analysis? Loan payment prediction and customer credit policy analysis. Classification and clustering of customers for targeted marketing. Detection of money laundering and other financial crimes. 92. Explain Data mining applications for the Telecommunication industry? Multidimensional analysis of telecommunication data. Fraudulent pattern analysis and the identification of unusual patterns. Multidimensional association and sequential pattern analysis. Use of visualization tools in telecommunication data analysis. 93. Define Spatial Data Warehouse? A Spatial Data warehouse is a subject oriented, integrated, time variant and non-volatile collection of both spatial and non-spatial data in support of spatial data mining and spatial data related decision making process. 94. What are the different types of dimensions in a spatial data cube? A non spatial dimension. A spatial to non spatial dimensions. A spatial to spatial dimensions. 95. Define Spatial Association rule? A spatial association rule is of the form A B[s%, c%].where A,B are sets of spatial or non spatial predicates, s% is support of the rule and c% is the confidence of the rule. 16

96. Define Horizontal Parallelism? Horizontal Parallelism which means that the database is partitioned across multiple disks and parallel processing occurs within a specific task that is performed concurrently on different processors against different sets of data. 97. Define Vertical Parallelism? Vertical Parallelism which occurs among different tasks all component query operations is executed in parallel in a pipelined fashion. 98. What is the need for OLAP? To analyze data stored in database To analyze different dimensions in multidimensional database. 99. Explain the various types of variables used in clustering? Interval scaled variables Binary variables o Symmetric binary variables o Asymmetric binary variables Nominal variables Ordinal variables Ratio-scaled variables 100. Explain the hierarchical method of clustering? Agglomerative and Divisive hierarchical clustering BIRCH (Balanced Iterative Reducing and Clustering using Hierarchies) CURE(Clustering Using REpresentatives) Chameleon 17