A Protocol for Privacy Preserving Neural Network Learning on Horizontally Partitioned Data
|
|
|
- Eustace Rice
- 10 years ago
- Views:
Transcription
1 A Protocol for Privacy Preserving Neural Network Learning on Horizontally Partitioned Data Nico Schlitter Faculty of Computer Science Otto-von-Guericke-University Magdeburg, Germany Abstract. This article presents a new approach for privacy preserving neural network training. Several studies have been devoted to privacy preserving supervised model learning, but little work has been done to extend neural network learning with a privacy preserving protocol. Neural networks are popular for many applications, among else those calling for a robust learning algorithm. In this study, we elaborate on privacy preserving classification as well as regression with neural networks on horizontally partitioned data. We consider a scenario of more than two parties that are semi-honest but curious. We extend the neural network classification algorithm presented by Rumelhart et al. with protocols for secure sum and secure matrix addition. The extended algorithm does not fully guarantee privacy in the sense defined by Goldreich, but we show that the information revealed is not associated to a specific party and could also be derived by juxtaposing the local data and the final model. 1 Introduction In recent years a new trend in medical diagnosis came up: The so-called computer-aided diagnosis (CAD) procedure automatically analyzes physical examination results and provides, based on the results, diagnosis support information for medical personnel. In particular in the area of medical image recognition promising results were achieved [1]. CAD systems are widely used to recognize pathological regions in medical images. For this, neural networks are often used to extract the relevant patterns from images. The application of neural networks for this task is mainly motivated by their robustness to noisy data and their ability to determine general patterns by data generalization. Furthermore, neural networks support incremental learning, i.e., it is possible to update a trained network by presenting new training data to the network. The training data is a set of vectors generated by applying feature extraction methods on medical images. Each medical image is used to generate a feature vector that is then classified during the application phase by the network as healthy or pathological. For this, the neural network builds an internal model that maps a feature vector to one of the two classes. The trained network is applied in CAD systems in order to classify new medical images.
2 Data sources are image forming technologies, such as Mammography [2], Magnetic Resonance Imaging [3, 4] and Ultrasonography [5]. The mentioned techniques are used to support breast cancer diagnosis. Phonovibrography, a method for processing and visualizing the vocal fold vibrations was recently introduced by Lohscheller et al. [6]. Based on this approach, a method for automatically extracting numerical features describing the spatio-temporal and symmetrical behavior of the vocal folds was proposed in [7]. For CAD purposes, the extracted features could be used to train a classifier in order to distinguish between healthy and pathological vocal fold vibration patterns [8]. Chest Radiography [9], another image forming technique, provides information for lung cancer recognition via a neural network classifier. The classification quality of neural network depends, amongst others, on the number of training data. Therefore, the classification performance can be increased by using images from more than one health care institution. However, in this case, privacy issues come up because medical images are considered as sensitive data. As a consequence, conventional methods for knowledge discovery from data are not appropriate; in this case, their use is even restricted by privacy protection laws, such as the Health Insurance Portability and Accountability Act (HIPAA). Privacy preserving data mining methods offer the chance to build models and extract patterns without disclosing private data. Privacy preservation methods for knowledge discovery can be categorized into two groups, data perturbation methods and cryptographic methods. Methods of the first group use data distortion, such as adding uniform noise, with the purpose of hidding private data, or, more formally, of guaranteeing k-anonymity. Methods of the second group are used for collaborative model learning: P 2 parties contribute their data for the learning of a shared model according to protocols [10, 11] that prevent the disclosure of the contributed data. Our work belongs to the second group of methods. Cryptographic methods have been proposed for supervised learning, among else with decision trees, support vector machines and Bayesian networks, for unsupervised model learning, e.g., with K-means and for association rule discovery. We focus on supervised learning and discuss privacy preservation for neural network training. We assume the scenario of more than two semi-honest parties. The data contributed among these parties are horizontally partitioned, in the sense that the shared model is built upon the union of the contributed datasets. The rest of the paper is organized as follows. The next section gives a brief overview on privacy preserving data mining in general and on the rarely studied issue of privacy preserving neural network learning. In Section 3, we present a widely used neural network learning method - the backpropagation learning. The extension of this method to provide for privacy preservation is discussed in Section 4. Subsequently, we discuss the limits of the protocol and the implications for applications. Section 5 concludes with a summary and some directions for future work.
3 2 Related work A considerable amount of methods for privacy preservation in data mining use cryptography techniques from the Secure Multi-party Computation (SMC) area based on the seminal works by Yao [10] and Goldreich et al. [11]. In [12 14] several relevant operations for SMC were defined. These operations are applied, amongst others, in the following data mining algorithms to ensure privacy preservation: Decision tree induction was enhanced for vertically [15, 16] and horizontally [17] partitioned data to generate decision trees without data disclosure. Privacy preserving association rule mining [18] was proposed as well as clustering methods [19]. A summary of data mining applications and their privacy preserving solutions is given by Vaidya et al. [20]. However, in the area of neural networks, the aspect of privacy preservation is mostly disregarded. Wan et al. present a generic formulation for secure computation of gradient descent methods [21]. The authors discuss a multi-party-protocol for vertically partitioned data that can be used to train a neural network. To ensure privacy, the target function is defined as a composition of two functions. Thus, the weights can be adapted locally. A second protocol for a secure summation of two scalar products is also suggested as a part of the overall process. A privacy preserving version of self organizing maps (SOM) is presented by Han et al. [22]. SOMs belong to the class of unsupervised learning techniques and are applied, e.g., for dimension reduction. The authors present a two-party protocol to adapt the network weights iteratively for vertically partitioned data. Barni et al. address in [23] a two-party privacy preserving protocol for neural network based computation. In their setting, the first party owns the confidential data and the second party owns the confidential model that is applied on the first party s data. Both, the data and the network model, are kept private. The approach does not consider how the used network model is actually trained but assumes that it already exists. In the following, we present our approach which extends the presented works by proposing a multi-party protocol for privacy-preserving neural network learning on horizontally partitioned data. 3 Backpropagation algorithm for neural network learning Neural Networks are nature inspired computation models that are widely used for regression and classification tasks. A neural network consists of nodes and weighted edges. In general, we distinguish feed forward and recurrent neural networks. In a feed forward network, the information is transmitted only in one direction, from the input nodes, through the hidden nodes (if any) to the output nodes. A feed forward network has therefore no cycles or loops. In recurrent neural networks, connections between nodes form a directed cycle. Therefore, information might also be transmitted backwards. An example of a feed forward network is shown in Fig. 1. Each node, also called neuron, produces an output by applying the internal activation function A on the weighted neuron inputs. The
4 output is weighted by the edge weight and the updated value is forwarded to the next layer. By this, an input vector traverses from the input layer through the network and produces an output in the output layer. In other words, the function performed by the network is determined by the nodes activation function and the internal network weights. Fig. 1. Feed forward network The aim of neural network training is to approximate an applicationdependent objective function. A popular method for feed forward neural network learning is the backpropagation algorithm proposed by Rumelhart et al. [24] which we present in the following. We assume a data set D = {x n, y n n = 1, 2,..., N} where each vector x n = {x n (1), x n (2),..., x n (q)} represents the n-th training vector with dimension q. If the vector y n = {y n (1),..., y n (m)} corresponds to the m-dimensional target vector, we approximate the target function F : x n y n that maps x n to y n. The function approximation is a complex task and usually a deviation between the approximate f and the original F remains. In order to measure this difference the mean squared error (MSE) is defined in Equation 1. Here, the vector ŷ n = f(x n ) refers to the m-dimensional output of the neural network while x n is the vector that is used as the network input. MSE = 1 N N n=1 i=1 ( 2 m ŷ n (i) y n (i)) (1) The objective of neural network training is to adapt the internal network weights in order to minimize the MSE. Usually the backpropagation algorithm [24] is iteratively used to reach this goal. At first, the network weights are randomly initialized. In the next step, the weights are iteratively adapted in order to reach
5 a local and in the best case a global optimum. In each round t of the training process the algorithm adapts the weight Wij l between the i-th neuron in layer l and the j-th neuron in layer l + 1. W l ij(t + 1) = W l ij(t) + W l ij(t) (2) In the online-mode the weight adjustment Wij l is performed after one learning iteration. For this, after randomly choosing one training vector x n, the network is learned and the error is calculated. Afterwards, the error is propagated back to the input layer and the weights are recalculated accordingly. However, here we refer to the batch-mode where a weight adjustment is done after all N training vectors have been applied to learn the network. Wij l (t) is then the sum of N changes Wij l (t, n) caused by the n-th data record. W l ij(t) = N Wij(t, l n) (3) n=1 The calculation of the weight adjustment Wij l (t, n) in round t is shown in Equation 4 where Ni l (t, n) refers to the output of the i-th neuron in layer l while x n is the vector that is used as input for the whole network. The parameter γ controls the size of the adaptation steps on the error surface and is called learning rate. W l ij(t, n) = γ(t) N l i (t, n) δ l+1 j (t, n) (4) The error δ l+1 j (t, n) is the fraction of the network error that is caused by the j-th neuron in layer l + 1. The computation of the error depends on the location of the respective neuron in the network and is shown in Equation 5. δi(t, l n) = A l i A l i ( ) ( ) ŷ n (i) ŷ n (i) y n (i) if l = l max ) (N i l(t) j W (5) ij l (t) δl+1 j if l < l max In case the neuron is located in the output layer l max the error signal is calculated as the product of the first derivative of the neuron s activation function A and the current network error. The error signal of the inner neurons is computed recursively as the product of two factors. The first factor is the output of the first derivative of the activation function applied to the neuron s output. The second factor is the weighted sum of the error signals that are propagated through the network. The iterative weight adaption described in Equation 2 is stopped when the maximum number of training rounds is reached or the MSE (cf. Equation 1) falls below a previously defined threshold. A more detailed description of the backpropagation algorithm and the proof of convergence can be found in [24].
6 4 A protocol for privacy preserving backpropagation In this section we propose a protocol for privacy preserving neural network training. We present a formalization of the given problem and introduce our training approach based on SMC methods - in particular the Secure Sum Protocol. 4.1 Problem setting We assume a scenario with P parties that want to collaborate for a common data analysis task using a neural network. The data is horizontally partitioned, meaning that all parties share the same data schema and each party holds a subset of the data containing several records of the total data set D. Analogously to the definitions in Section 3, the data set of the p-th party is specified as D p = {x p n, yn} p and contains N p = D p records. As a consequence the data set D = P p=1 D p only exists in virtual form: it is distributed over P parties and is never merged because of privacy concerns. Additionally we suppose a semi-honest behavior of the participating parties. Thus, we assume, that all parties adhere to the given protocol. However, each party is curious and attempts to derive additional information about the other parties. We are also aware of colluding parties that work together against one or more participants in order to obtain private information about other parties. Consequently the proposed protocol needs to deal with the described situation and has to protect the participants privacy. The common objective of all parties is to train the neural network in order to minimize the overall MSE: MSE = 1 P p=1 N p N P p ( 2 m ŷn(i) p yn(i)) p (6) p=1 n=1 i=1 In the next Section, we present our proposed protocol which ensures, that the underlying private datasets D p and related knowledge are kept private. 4.2 The protocol execution In the following, we describe our proposed protocol for privacy preserving neural network training. The procedure is shown in Algorithm 1. Basically we extend Rumelhart s original algorithm [24] to the multi-party case. Thus, we adapt the formula for the weight adjustment (cf. Equation 3) in order to handle the weight adaptations that are determined by different parties. In Equations 7 to 9 we show, that Equation 3 is equivalent to 9. Wij(t) l = N W ij(t, l n) for N = N p (7) n=1 = P Np W ij(t, l n) (8) p=1 n=1 = P lp Wij (t) (9) p=1
7 Each party holds a subset of the complete data set D and computes the weight adjustments based on its own local data records. Equations 7 to 9 show that the sum of these weight changes is equal to the sum of the subtotals of weight changes as a consequence of the associative law. Hence, the neural network trained by our protocol is equal to a network generated by Rumelhart s algorithm. Algorithm 1: Privacy preserving backpropagation network training input : distributed data set D = P p=1 Dp output: weight matrix W l ij 1 All parties agree on network topology and type of activation function; Initiate the global network 2 foreach party P do Randomly choosing of local weight matrix 3 W lp ij = rand(); Create global weight matrix by using secure matrix addition Wij l = 4 P p=1 W lp ij ; Start network training 5 repeat 6 foreach party p do Copy global to local weight matrix 7 W lp ij = W ij; l Compute local weight changes 8 9 = N p n=1 γ N i l (n) δ l+1 j (n); Compute local squared error MSE p = ( N p 2 m n=1 i=1 ŷn(i) p yn(i)) p ; W lp ij Compute global weight changes by using secure matrix addition Wij l = 10 P lp p=1 Wij ; Adapt global weight matrix Wij l = Wij l + Wij; l Compute MSE by using secure sum protocol 1 MSE = P Pp=1 N p p=1 MSEp ; until MSE is small enough ; return Wij; l The execution of the proposed protocol works as follows. According to the original algorithm the parties randomly initiate the network weights (cf. line 2 in Algorithm 1). For this, the parties have to agree upon a network topology by defining the number of network layers, the number of neurons per layer and the applied activation functions (cf. line 1 in Algorithm 1). In the next step, each party p creates a local initial weight matrix W lp ij (0) by choosing a random number for each edge weight (cf. line 3 in Algorithm 1). Thus, each party builds
8 a local network and the global weight matrix is computed as the sum of all local weight matrices (cf. line 4 in Algorithm 1): W l ij(0) = P p=1 W lp ij (0) (10) In the learning step, the weights are iteratively adapted in order to minimize the total MSE (cf. line 6-9 in Algorithm 1). For this, each party presents its training vectors x p n to the network s input layer. The vector is processed through the different network layers and at the network s output layer, the error between the actual and the desired output is measured. This error is afterwards minimized by back propagating it through the network and by adjusting the weight matrices according to Equation 3. By adding up the local weight changes, the global changes are computed as shown in Equation 9. The resulting weight change matrix Wij l (t) is published to all participating parties (cf. line 10 in Algorithm 1). Thus, the new weight matrix of the network is adapted by each party locally (cf. line 11 in Algorithm 1): W l ij(t + 1) = W lp ij (t + 1) = W l ij(t) + W l ij(t) (11) The weight adaption is repeated until the total MSE reaches a minimum or a convergence criterion is met. In order to calculate this total MSE Equation 6 is used (cf. line 12 in Algorithm 1). To ensure privacy, the sums in Equations 10 and 9 are computed by using the secure sum (cf. line 12 in Algorithm 1) and the secure matrix addition protocol (cf. line 4 and 10 in Algorithm 1) which we discuss in the following. Secure sum protocol The secure sum protocol as presented in [13] allows the secure summation of distributed summands. During the protocol s execution only the final sum but not the single summands are published. Assuming a multiparty scenario with P > 2 parties and respectively P private values v 1,..., v P the protocol s objective is to compute the sum S = P i=1 v i that is known to be in the interval [0, S max ]. In the following, we briefly sketch the procedure. Party 1 randomly chooses a uniformly distributed value R and sends the sum R + v 1 (mod S max ) to party 2. The combination of the uniformly distributed value R and the modulo operation ensures, that party 2 learns nothing about v 1. Party 2 adds its own private value v 2 and sends the result to the next party. In general each party p receives R + p 1 i=1 v i (mod S max ) from its predecessor and sends R+ p i=1 v i (mod S max ) to its successor. The last party sends the masked sum back to party 1 which is able to compute the correct sum by subtracting its randomly chosen value R: S = R + P v i R (mod S max ) = i=1 P v i (12) The secure sum protocol is vulnerable in case two or more parties work together (colluders). For example, if parties p 1 and p + 1 collaborate, they are able to i=1
9 calculate p s private value v p : p 1 v p = R + v i i=1 p v i (mod S max ) (13) i=1 This problem can be solved by splitting the private value of each party into different shares. The sum of the shares of all parties is then computed separately. However, the order of the participating parties is permuted for each share in order to guarantee different predecessor and successor for each iteration. To compute the private value v p all predecessors and successors of party p have to collude. The protocol s security depends on the number of iterations and the number of colluding parties. The more iterations are performed, the more parties might collude without risking privacy breaches. Thus, the fraction of colluders to semihonest parties and the number of iterations determine the security level of the protocol. Secure matrix addition Based on the secure sum protocol Yu and Vaidya propose a protocol for secure matrix addition [25]. For all components of the P private matrices W 1,..., W P the secure sum protocol is simultaneously executed. By this, the matrix sum P p=1 W p is calculated in one step instead of an iterative computation for each matrix component. Since the protocol also uses the secure sum protocol, the same security restrictions apply, i.e., the privacy also depends on the fraction of colluders. 4.3 Security analysis The security of a protocol for Secure Multi-party Computation with respect to information disclosure is usually evaluated with a procedure called privacy by simulation: A protocol is considered secure if the view of each party during the execution can be simulated by using only its local input and the final output of the protocol. This is equivalent to saying that a party learns nothing beyond what it knew originally and what is inherent in the final result. Details on the notion of security in SMC can be found in [26]. Our algorithm does not fulfill the security requirements of SMC as presented above, because each party obtains more data during the execution of protocol than is inherent in the final model. However, the parties still do not obtain information about any specific other party. In particular: In line 10 of Algorithm 1 the global weight change matrix W (t) is computed by using the secure matrix addition protocol. The new weight matrix W (t + 1) is calculated (cf. line 11) and published to all parties (cf. line 7). The new weight matrix is used for next local network training iteration. By subtracting the new weight matrix W (t+1) and the local weight changes W p (t) from the old weight matrix W (t) each party is able to deduce the sum of weight changes W p (t) proposed by all other participating parties: W p (t) = W (t + 1) W (t) W p (t) (14) = W (t) W p (t) (15)
10 The same argumentation holds for the calculation of the number of records and the MSE. Both values can be inferred by using the secure sums computed in line 12. The number of records hold by the other parties can be determined as ( P ) N p = N i N p (16) The MSE caused by the other parties data can be calculated as [( MSE p = 1 P ) ] MSE i MSE p N p i=1 i=1 (17) We would like to emphasize, that W p (t) contains the weight changes of all other parties excluding party p, N p refers to the number of data records that are not held by p and MSE p represents the MSE of the other parties. Therefore, a single attacker is only able to obtain group related information and no information about individual parties. Using the inferred group related information, an attacker is able to detect differences between data set characteristics in its own data set D p and the remaining data set i p D i. For example in each iteration, the attacker can compare its own MSE p with the MSE p of the other parties. If both MSEs are similar, the attacker can assume that the data set characteristics are similar as well. Otherwise, in case the MSEs differ considerably, it can be speculated, that the characteristics vary as well. Analogously, an attacker might use the weight change matrix in order to compare its own data set characteristics with the other party s. However, an attacker is not able to compare its data set to another specific party or even to reconstruct other party s data records. If parties collude, the security depends on the security of the secure sum protocol. Even if the highest security level for the secure sum protocol is applied, the private information of a single party can be revealed if all P 1 parties collude. Therefore, the security of our proposed protocol is ensured if all parties are semi-honest and do not collude. Furthermore, the protocol is secure if the number of colluders is sufficiently small with respect to the security level of the secure sum protocol. Even though, our protocol does not fulfill the security requirements stated by the SMC community, it fulfills a more general privacy definition which is nevertheless relevant for practical applications. Vaidya et al. define two requirements for privacy preserving data mining techniques: a) Any information disclosed cannot be traced to an individual or b) does not constitute an intrusion. As discussed above, during the execution of our protocol, only group-related information is disclosed and the data revealed in each iteration does not contain any information about individual parties. Actually, the information that can be inferred from the data which is disclosed after each iteration has no more value to any party than the information that is included in the final model. Following this perspective, one could argue, that the protocol does also fulfill SMC privacy requirements since only data but no further information than that provided by the final model is revealed during the execution of the protocol.
11 5 Conclusion In this paper we proposed a privacy preserving protocol for neural network training on horizontally partitioned data. We used the secure sum and secure matrix addition protocol in order to extend Rumelhart s algorithm to a multi-party scenario. By this, the weight adaptations of the participating parties are iteratively included in the global neural network weight matrix. The protocol is not secure in the strict sense of privacy-by-simulation but the disclosed information is not associated to a single party (for P > 2 parties). Rather, a party can only compare the characteristic of the own local data set to the characteristic of the other parties data sets - an operation that can be performed also by juxtaposing the local data to the final model. On this account, we consider the proposed protocol as unsecure in the strict SMC sense but sufficiently secure for practical applications. Designing a SMC-secure privacy preserving protocol for neural network learning is still an open task for future work. We consider it also important to identify scenarios, where a SMC-secure protocol is indispensable, and study the interplay of data skew and SMC-security. References 1. Giger, M.L.: Computer-aided diagnosis of breast lesions in medical images. Computing in Science and Engineering 02(5) (2000) Lasztovicza, L., Pataki, B., Szekely, N., Toth, N.: Neural network based microcalcification detection in a mammographic cad system. Intelligent Data Acquisition and Advanced Computing Systems: Technology and Applications, Proceedings of the Second IEEE International Workshop on (Sept. 2003) Arbach, L., Stolpen, A., Reinhardt, J.M.: Classification of breast mri lesions using a backpropagation neural network (bnn). In: ISBI, IEEE (2004) Meinel, L.A., Stolpen, A.H., Berbaum, K.S., Fajardo, L.L., Reinhardt, J.M.: Breast mri lesion classification: Improved performance of human readers with a backpropagation neural network computer-aided diagnosis (cad) system. Journal of Magnetic Resonance Imaging 25 (2006) Chen, D.R., Chang, R.F., Huang, Y.L.: Computer-aided Diagnosis Applied to US of Solid Breast Nodules by Using Neural Networks. Radiology 213(2) (1999) Lohscheller, J., Eysholdt, U., Toy, H., Dllinger, M.: Phonovibrography: Mapping high-speed movies of vocal fold vibrations into 2-d diagrams for visualizing and analyzing the underlying laryngeal dynamics. Medical Imaging, IEEE Transactions on 27(3) (March 2008) Doellinger, M., Lohscheller, J., McWorther, A., Voigt, D., Kunduk, M.: Support vector machine classification of different phonatory tasks revealed by phonovibrogram analysis and acoustic perturbation measurements. The Journal of the Acoustical Society of America (2008) 8. Lohscheller, J., Voigt, D., Doellinger, M.: Classification of normal and pathological vocal fold vibrations using phonvibrograms. The Journal of the Acoustical Society of America (2008)
12 9. Lo, S.C.B., Lin, J.S.J., Freedman, M.T., Mun, S.K.: Application of artificial neural networks to medical image pattern recognition: Detection of clustered microcalcifications on mammograms and lung cancer on chest radiographs. J. VLSI Signal Process. Syst. 18(3) (1998) Yao, A.C.C.: How to generate and exchange secrets (extended abstract). In: FOCS 86: Proceedings of the 27th Annual Symposium on Foundations of Computer Science. (1986) Goldreich, O., Micali, S., Wigderson, A.: How to play any mental game. In: STOC 87: Proceedings of the nineteenth annual ACM conference on Theory of computing, New York, NY, USA, ACM Press (1987) Du, W., Atallah, M.J.: Secure multi-party computation problems and their applications: A review and open problems. In: Proceedings of the 2001 Workshop on New Security Paradigms, Cloudcroft, NM, ACM Press (September 2001) Clifton, C., Kantarcioglu, M., Vaidya, J., Lin, X., Zhu, M.Y.: Tools for privacy preserving distributed data mining. SIGKDD Explor. Newsl. 4(2) (2002) Pinkas, B.: Cryptographic techniques for privacy-preserving data mining. SIGKDD Explor. Newsl. 4(2) (2002) Vaidya, J., Clifton, C.: Privacy-preserving decision trees over vertically partitioned data. In: The 19th Annual IFIP WG 11.3 Working Conference on Data and Applications Security, Storrs, Connecticut, Springer ( ) 16. Du, W., Zhan, Z.: Building decision tree classifier on private data. In: CRPIT 14: Proceedings of the IEEE international conference on Privacy, security and data mining, Darlinghurst, Australia, Australia, Australian Computer Society, Inc. (2002) Lindell, Y., Pinkas, B.: Privacy preserving data mining. In: Advances in Cryptology (CRYPTO 00). Volume 1880 of Lecture Notes in Computer Science., Springer- Verlag (2000) Kantarcioglu, M.: Privacy-preserving distributed mining of association rules on horizontally partitioned data. IEEE Transactions on Knowledge and Data Engineering 16(9) (2004) Vaidya, J., Clifton, C.: Privacy-preserving k-means clustering over vertically partitioned data. In: KDD 03: Proceedings of the ninth ACM SIGKDD international conference on Knowledge discovery and data mining, New York, NY, USA, ACM Press (2003) Vaidya, J., Clifton, C., Zhu, M.: Privacy Preserving Data Mining. Volume 19 of Advances in Information Security. Springer (2006) 21. Wan, L., Ng, W.K., Han, S., Lee, V.C.S.: Privacy-preservation for gradient descent methods. In Berkhin, P., Caruana, R., Wu, X., eds.: KDD, New York, NY, USA, ACM Press (2007) Han, S., Ng, W.K.: Privacy-preserving self-organizing map. In Song, I.Y., Eder, J., Nguyen, T.M., eds.: DaWaK. Volume 4654 of Lecture Notes in Computer Science., Springer (2007) Barni, M., Orlandi, C., Piva, A.: A privacy-preserving protocol for neural-networkbased computation. In: MM&Sec 06: Proceeding of the 8th workshop on Multimedia and security, New York, NY, USA, ACM Press (2006) Rumelhart, D.E., Hinton, G.E., Williams, R.J.: Learning Interior Representation by Error Propagation. Parallel Distributed Processing 1 (1986) Yu, H., Vaidya, J.: Secure matrix addition. Technical Report UIOWA CS TR04-04, University of Iowa (November 2004) 26. Goldreich, O.: Foundations of Cryptography: Volume 2, Basic Applications. Cambridge University Press, New York, NY, USA (2004)
Experimental Analysis of Privacy-Preserving Statistics Computation
Experimental Analysis of Privacy-Preserving Statistics Computation Hiranmayee Subramaniam 1, Rebecca N. Wright 2, and Zhiqiang Yang 2 1 Stevens Institute of Technology graduate, [email protected]. 2
EFFICIENT DATA PRE-PROCESSING FOR DATA MINING
EFFICIENT DATA PRE-PROCESSING FOR DATA MINING USING NEURAL NETWORKS JothiKumar.R 1, Sivabalan.R.V 2 1 Research scholar, Noorul Islam University, Nagercoil, India Assistant Professor, Adhiparasakthi College
Social Media Mining. Data Mining Essentials
Introduction Data production rate has been increased dramatically (Big Data) and we are able store much more data than before E.g., purchase data, social media data, mobile phone data Businesses and customers
MapReduce Approach to Collective Classification for Networks
MapReduce Approach to Collective Classification for Networks Wojciech Indyk 1, Tomasz Kajdanowicz 1, Przemyslaw Kazienko 1, and Slawomir Plamowski 1 Wroclaw University of Technology, Wroclaw, Poland Faculty
Enhancement of Security in Distributed Data Mining
Enhancement of Security in Distributed Data Mining Sharda Darekar 1, Prof.D.K.Chitre, 2 1,2 Department Of Computer Engineering, Terna Engineering College,Nerul,Navi Mumbai. 1 [email protected],
Data Outsourcing based on Secure Association Rule Mining Processes
, pp. 41-48 http://dx.doi.org/10.14257/ijsia.2015.9.3.05 Data Outsourcing based on Secure Association Rule Mining Processes V. Sujatha 1, Debnath Bhattacharyya 2, P. Silpa Chaitanya 3 and Tai-hoon Kim
6.2.8 Neural networks for data mining
6.2.8 Neural networks for data mining Walter Kosters 1 In many application areas neural networks are known to be valuable tools. This also holds for data mining. In this chapter we discuss the use of neural
A Secure Model for Medical Data Sharing
International Journal of Database Theory and Application 45 A Secure Model for Medical Data Sharing Wong Kok Seng 1,1,Myung Ho Kim 1, Rosli Besar 2, Fazly Salleh 2 1 Department of Computer, Soongsil University,
NEURAL NETWORKS IN DATA MINING
NEURAL NETWORKS IN DATA MINING 1 DR. YASHPAL SINGH, 2 ALOK SINGH CHAUHAN 1 Reader, Bundelkhand Institute of Engineering & Technology, Jhansi, India 2 Lecturer, United Institute of Management, Allahabad,
Comparison of K-means and Backpropagation Data Mining Algorithms
Comparison of K-means and Backpropagation Data Mining Algorithms Nitu Mathuriya, Dr. Ashish Bansal Abstract Data mining has got more and more mature as a field of basic research in computer science and
Predicting the Risk of Heart Attacks using Neural Network and Decision Tree
Predicting the Risk of Heart Attacks using Neural Network and Decision Tree S.Florence 1, N.G.Bhuvaneswari Amma 2, G.Annapoorani 3, K.Malathi 4 PG Scholar, Indian Institute of Information Technology, Srirangam,
PRIVACY PRESERVING BACK-PROPAGATION NEURAL NETWORK IN CLOUD COMPUTING
PRIVACY PRESERVING BACK-PROPAGATION NEURAL NETWORK IN CLOUD COMPUTING Dr.P.Kumar M.sc., M.Tech., Ph.d, R.Muthu Vijay Deepak M.Tech., Abstract we find the problem of Privacy Preserving Back Propagation
Comparison of Supervised and Unsupervised Learning Classifiers for Travel Recommendations
Volume 3, No. 8, August 2012 Journal of Global Research in Computer Science REVIEW ARTICLE Available Online at www.jgrcs.info Comparison of Supervised and Unsupervised Learning Classifiers for Travel Recommendations
Keywords: Image complexity, PSNR, Levenberg-Marquardt, Multi-layer neural network.
Global Journal of Computer Science and Technology Volume 11 Issue 3 Version 1.0 Type: Double Blind Peer Reviewed International Research Journal Publisher: Global Journals Inc. (USA) Online ISSN: 0975-4172
SELECTING NEURAL NETWORK ARCHITECTURE FOR INVESTMENT PROFITABILITY PREDICTIONS
UDC: 004.8 Original scientific paper SELECTING NEURAL NETWORK ARCHITECTURE FOR INVESTMENT PROFITABILITY PREDICTIONS Tonimir Kišasondi, Alen Lovren i University of Zagreb, Faculty of Organization and Informatics,
Neural Networks and Back Propagation Algorithm
Neural Networks and Back Propagation Algorithm Mirza Cilimkovic Institute of Technology Blanchardstown Blanchardstown Road North Dublin 15 Ireland [email protected] Abstract Neural Networks (NN) are important
Visualization of Breast Cancer Data by SOM Component Planes
International Journal of Science and Technology Volume 3 No. 2, February, 2014 Visualization of Breast Cancer Data by SOM Component Planes P.Venkatesan. 1, M.Mullai 2 1 Department of Statistics,NIRT(Indian
Advanced Ensemble Strategies for Polynomial Models
Advanced Ensemble Strategies for Polynomial Models Pavel Kordík 1, Jan Černý 2 1 Dept. of Computer Science, Faculty of Information Technology, Czech Technical University in Prague, 2 Dept. of Computer
Performance Evaluation of Artificial Neural. Networks for Spatial Data Analysis
Contemporary Engineering Sciences, Vol. 4, 2011, no. 4, 149-163 Performance Evaluation of Artificial Neural Networks for Spatial Data Analysis Akram A. Moustafa Department of Computer Science Al al-bayt
Analecta Vol. 8, No. 2 ISSN 2064-7964
EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,
Role of Neural network in data mining
Role of Neural network in data mining Chitranjanjit kaur Associate Prof Guru Nanak College, Sukhchainana Phagwara,(GNDU) Punjab, India Pooja kapoor Associate Prof Swami Sarvanand Group Of Institutes Dinanagar(PTU)
Chapter 12 Discovering New Knowledge Data Mining
Chapter 12 Discovering New Knowledge Data Mining Becerra-Fernandez, et al. -- Knowledge Management 1/e -- 2004 Prentice Hall Additional material 2007 Dekai Wu Chapter Objectives Introduce the student to
Open Access Research on Application of Neural Network in Computer Network Security Evaluation. Shujuan Jin *
Send Orders for Reprints to [email protected] 766 The Open Electrical & Electronic Engineering Journal, 2014, 8, 766-771 Open Access Research on Application of Neural Network in Computer Network
A Simple Feature Extraction Technique of a Pattern By Hopfield Network
A Simple Feature Extraction Technique of a Pattern By Hopfield Network A.Nag!, S. Biswas *, D. Sarkar *, P.P. Sarkar *, B. Gupta **! Academy of Technology, Hoogly - 722 *USIC, University of Kalyani, Kalyani
To improve the problems mentioned above, Chen et al. [2-5] proposed and employed a novel type of approach, i.e., PA, to prevent fraud.
Proceedings of the 5th WSEAS Int. Conference on Information Security and Privacy, Venice, Italy, November 20-22, 2006 46 Back Propagation Networks for Credit Card Fraud Prediction Using Stratified Personalized
International Journal of Scientific & Engineering Research, Volume 4, Issue 10, October-2013 ISSN 2229-5518 1582
1582 AN EFFICIENT CRYPTOGRAPHIC APPROACH FOR PRESERVING PRIVACY IN DATA MINING T.Sujitha 1, V.Saravanakumar 2, C.Saravanabhavan 3 1. M.E. Student, [email protected] 2. Assistant Professor, [email protected]
Neural Networks and Support Vector Machines
INF5390 - Kunstig intelligens Neural Networks and Support Vector Machines Roar Fjellheim INF5390-13 Neural Networks and SVM 1 Outline Neural networks Perceptrons Neural networks Support vector machines
PLAANN as a Classification Tool for Customer Intelligence in Banking
PLAANN as a Classification Tool for Customer Intelligence in Banking EUNITE World Competition in domain of Intelligent Technologies The Research Report Ireneusz Czarnowski and Piotr Jedrzejowicz Department
How To Use Neural Networks In Data Mining
International Journal of Electronics and Computer Science Engineering 1449 Available Online at www.ijecse.org ISSN- 2277-1956 Neural Networks in Data Mining Priyanka Gaur Department of Information and
Lecture 6. Artificial Neural Networks
Lecture 6 Artificial Neural Networks 1 1 Artificial Neural Networks In this note we provide an overview of the key concepts that have led to the emergence of Artificial Neural Networks as a major paradigm
DATA MINING - 1DL360
DATA MINING - 1DL360 Fall 2013" An introductory class in data mining http://www.it.uu.se/edu/course/homepage/infoutv/per1ht13 Kjell Orsborn Uppsala Database Laboratory Department of Information Technology,
Data Mining - Evaluation of Classifiers
Data Mining - Evaluation of Classifiers Lecturer: JERZY STEFANOWSKI Institute of Computing Sciences Poznan University of Technology Poznan, Poland Lecture 4 SE Master Course 2008/2009 revised for 2010
Homomorphic Encryption Schema for Privacy Preserving Mining of Association Rules
Homomorphic Encryption Schema for Privacy Preserving Mining of Association Rules M.Sangeetha 1, P. Anishprabu 2, S. Shanmathi 3 Department of Computer Science and Engineering SriGuru Institute of Technology
An Analysis on Density Based Clustering of Multi Dimensional Spatial Data
An Analysis on Density Based Clustering of Multi Dimensional Spatial Data K. Mumtaz 1 Assistant Professor, Department of MCA Vivekanandha Institute of Information and Management Studies, Tiruchengode,
Classification algorithm in Data mining: An Overview
Classification algorithm in Data mining: An Overview S.Neelamegam #1, Dr.E.Ramaraj *2 #1 M.phil Scholar, Department of Computer Science and Engineering, Alagappa University, Karaikudi. *2 Professor, Department
International Journal of Computer Science Trends and Technology (IJCST) Volume 2 Issue 3, May-Jun 2014
RESEARCH ARTICLE OPEN ACCESS A Survey of Data Mining: Concepts with Applications and its Future Scope Dr. Zubair Khan 1, Ashish Kumar 2, Sunny Kumar 3 M.Tech Research Scholar 2. Department of Computer
MANAGING QUEUE STABILITY USING ART2 IN ACTIVE QUEUE MANAGEMENT FOR CONGESTION CONTROL
MANAGING QUEUE STABILITY USING ART2 IN ACTIVE QUEUE MANAGEMENT FOR CONGESTION CONTROL G. Maria Priscilla 1 and C. P. Sumathi 2 1 S.N.R. Sons College (Autonomous), Coimbatore, India 2 SDNB Vaishnav College
Prediction of Cancer Count through Artificial Neural Networks Using Incidence and Mortality Cancer Statistics Dataset for Cancer Control Organizations
Using Incidence and Mortality Cancer Statistics Dataset for Cancer Control Organizations Shivam Sidhu 1,, Upendra Kumar Meena 2, Narina Thakur 3 1,2 Department of CSE, Student, Bharati Vidyapeeth s College
Impact of Feature Selection on the Performance of Wireless Intrusion Detection Systems
2009 International Conference on Computer Engineering and Applications IPCSIT vol.2 (2011) (2011) IACSIT Press, Singapore Impact of Feature Selection on the Performance of ireless Intrusion Detection Systems
Data quality in Accounting Information Systems
Data quality in Accounting Information Systems Comparing Several Data Mining Techniques Erjon Zoto Department of Statistics and Applied Informatics Faculty of Economy, University of Tirana Tirana, Albania
CHAPTER 5 PREDICTIVE MODELING STUDIES TO DETERMINE THE CONVEYING VELOCITY OF PARTS ON VIBRATORY FEEDER
93 CHAPTER 5 PREDICTIVE MODELING STUDIES TO DETERMINE THE CONVEYING VELOCITY OF PARTS ON VIBRATORY FEEDER 5.1 INTRODUCTION The development of an active trap based feeder for handling brakeliners was discussed
SURVIVABILITY ANALYSIS OF PEDIATRIC LEUKAEMIC PATIENTS USING NEURAL NETWORK APPROACH
330 SURVIVABILITY ANALYSIS OF PEDIATRIC LEUKAEMIC PATIENTS USING NEURAL NETWORK APPROACH T. M. D.Saumya 1, T. Rupasinghe 2 and P. Abeysinghe 3 1 Department of Industrial Management, University of Kelaniya,
Secure Collaborative Privacy In Cloud Data With Advanced Symmetric Key Block Algorithm
Secure Collaborative Privacy In Cloud Data With Advanced Symmetric Key Block Algorithm Twinkle Graf.F 1, Mrs.Prema.P 2 1 (M.E- CSE, Dhanalakshmi College of Engineering, Chennai, India) 2 (Asst. Professor
A Study of Web Log Analysis Using Clustering Techniques
A Study of Web Log Analysis Using Clustering Techniques Hemanshu Rana 1, Mayank Patel 2 Assistant Professor, Dept of CSE, M.G Institute of Technical Education, Gujarat India 1 Assistant Professor, Dept
Using multiple models: Bagging, Boosting, Ensembles, Forests
Using multiple models: Bagging, Boosting, Ensembles, Forests Bagging Combining predictions from multiple models Different models obtained from bootstrap samples of training data Average predictions or
Performing Data Mining in (SRMS) through Vertical Approach with Association Rules
Performing Data Mining in (SRMS) through Vertical Approach with Association Rules Mr. Ambarish S. Durani 1 and Miss. Rashmi B. Sune 2 MTech (III rd Sem), Vidharbha Institute of Technology, Nagpur, Nagpur
ARTIFICIAL INTELLIGENCE (CSCU9YE) LECTURE 6: MACHINE LEARNING 2: UNSUPERVISED LEARNING (CLUSTERING)
ARTIFICIAL INTELLIGENCE (CSCU9YE) LECTURE 6: MACHINE LEARNING 2: UNSUPERVISED LEARNING (CLUSTERING) Gabriela Ochoa http://www.cs.stir.ac.uk/~goc/ OUTLINE Preliminaries Classification and Clustering Applications
Privacy-Preserving Outsourcing Support Vector Machines with Random Transformation
Privacy-Preserving Outsourcing Support Vector Machines with Random Transformation Keng-Pei Lin Ming-Syan Chen Department of Electrical Engineering, National Taiwan University, Taipei, Taiwan Research Center
Chapter 4: Artificial Neural Networks
Chapter 4: Artificial Neural Networks CS 536: Machine Learning Littman (Wu, TA) Administration icml-03: instructional Conference on Machine Learning http://www.cs.rutgers.edu/~mlittman/courses/ml03/icml03/
Clustering through Decision Tree Construction in Geology
Nonlinear Analysis: Modelling and Control, 2001, v. 6, No. 2, 29-41 Clustering through Decision Tree Construction in Geology Received: 22.10.2001 Accepted: 31.10.2001 A. Juozapavičius, V. Rapševičius Faculty
Load balancing in a heterogeneous computer system by self-organizing Kohonen network
Bull. Nov. Comp. Center, Comp. Science, 25 (2006), 69 74 c 2006 NCC Publisher Load balancing in a heterogeneous computer system by self-organizing Kohonen network Mikhail S. Tarkov, Yakov S. Bezrukov Abstract.
Standardization and Its Effects on K-Means Clustering Algorithm
Research Journal of Applied Sciences, Engineering and Technology 6(7): 399-3303, 03 ISSN: 040-7459; e-issn: 040-7467 Maxwell Scientific Organization, 03 Submitted: January 3, 03 Accepted: February 5, 03
Specific Usage of Visual Data Analysis Techniques
Specific Usage of Visual Data Analysis Techniques Snezana Savoska 1 and Suzana Loskovska 2 1 Faculty of Administration and Management of Information systems, Partizanska bb, 7000, Bitola, Republic of Macedonia
Self Organizing Maps: Fundamentals
Self Organizing Maps: Fundamentals Introduction to Neural Networks : Lecture 16 John A. Bullinaria, 2004 1. What is a Self Organizing Map? 2. Topographic Maps 3. Setting up a Self Organizing Map 4. Kohonen
Robust Outlier Detection Technique in Data Mining: A Univariate Approach
Robust Outlier Detection Technique in Data Mining: A Univariate Approach Singh Vijendra and Pathak Shivani Faculty of Engineering and Technology Mody Institute of Technology and Science Lakshmangarh, Sikar,
An Overview of Knowledge Discovery Database and Data mining Techniques
An Overview of Knowledge Discovery Database and Data mining Techniques Priyadharsini.C 1, Dr. Antony Selvadoss Thanamani 2 M.Phil, Department of Computer Science, NGM College, Pollachi, Coimbatore, Tamilnadu,
Quality Assessment in Spatial Clustering of Data Mining
Quality Assessment in Spatial Clustering of Data Mining Azimi, A. and M.R. Delavar Centre of Excellence in Geomatics Engineering and Disaster Management, Dept. of Surveying and Geomatics Engineering, Engineering
Prediction Model for Crude Oil Price Using Artificial Neural Networks
Applied Mathematical Sciences, Vol. 8, 2014, no. 80, 3953-3965 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.43193 Prediction Model for Crude Oil Price Using Artificial Neural Networks
Using Data Mining for Mobile Communication Clustering and Characterization
Using Data Mining for Mobile Communication Clustering and Characterization A. Bascacov *, C. Cernazanu ** and M. Marcu ** * Lasting Software, Timisoara, Romania ** Politehnica University of Timisoara/Computer
MALLET-Privacy Preserving Influencer Mining in Social Media Networks via Hypergraph
MALLET-Privacy Preserving Influencer Mining in Social Media Networks via Hypergraph Janani K 1, Narmatha S 2 Assistant Professor, Department of Computer Science and Engineering, Sri Shakthi Institute of
Data Mining Algorithms Part 1. Dejan Sarka
Data Mining Algorithms Part 1 Dejan Sarka Join the conversation on Twitter: @DevWeek #DW2015 Instructor Bio Dejan Sarka ([email protected]) 30 years of experience SQL Server MVP, MCT, 13 books 7+ courses
SPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING
AAS 07-228 SPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING INTRODUCTION James G. Miller * Two historical uncorrelated track (UCT) processing approaches have been employed using general perturbations
Prediction of DDoS Attack Scheme
Chapter 5 Prediction of DDoS Attack Scheme Distributed denial of service attack can be launched by malicious nodes participating in the attack, exploit the lack of entry point in a wireless network, and
Comparative Analysis of EM Clustering Algorithm and Density Based Clustering Algorithm Using WEKA tool.
International Journal of Engineering Research and Development e-issn: 2278-067X, p-issn: 2278-800X, www.ijerd.com Volume 9, Issue 8 (January 2014), PP. 19-24 Comparative Analysis of EM Clustering Algorithm
American International Journal of Research in Science, Technology, Engineering & Mathematics
American International Journal of Research in Science, Technology, Engineering & Mathematics Available online at http://www.iasir.net ISSN (Print): 2328-349, ISSN (Online): 2328-3580, ISSN (CD-ROM): 2328-3629
Statistical Data Mining. Practical Assignment 3 Discriminant Analysis and Decision Trees
Statistical Data Mining Practical Assignment 3 Discriminant Analysis and Decision Trees In this practical we discuss linear and quadratic discriminant analysis and tree-based classification techniques.
Categorical Data Visualization and Clustering Using Subjective Factors
Categorical Data Visualization and Clustering Using Subjective Factors Chia-Hui Chang and Zhi-Kai Ding Department of Computer Science and Information Engineering, National Central University, Chung-Li,
Support Vector Machines with Clustering for Training with Very Large Datasets
Support Vector Machines with Clustering for Training with Very Large Datasets Theodoros Evgeniou Technology Management INSEAD Bd de Constance, Fontainebleau 77300, France [email protected] Massimiliano
degrees of freedom and are able to adapt to the task they are supposed to do [Gupta].
1.3 Neural Networks 19 Neural Networks are large structured systems of equations. These systems have many degrees of freedom and are able to adapt to the task they are supposed to do [Gupta]. Two very
An Algorithm for Automatic Base Station Placement in Cellular Network Deployment
An Algorithm for Automatic Base Station Placement in Cellular Network Deployment István Törős and Péter Fazekas High Speed Networks Laboratory Dept. of Telecommunications, Budapest University of Technology
Sensitivity Analysis for Data Mining
Sensitivity Analysis for Data Mining J. T. Yao Department of Computer Science University of Regina Regina, Saskatchewan Canada S4S 0A2 E-mail: [email protected] Abstract An important issue of data mining
A Data Generator for Multi-Stream Data
A Data Generator for Multi-Stream Data Zaigham Faraz Siddiqui, Myra Spiliopoulou, Panagiotis Symeonidis, and Eleftherios Tiakas University of Magdeburg ; University of Thessaloniki. [siddiqui,myra]@iti.cs.uni-magdeburg.de;
A survey on Data Mining based Intrusion Detection Systems
International Journal of Computer Networks and Communications Security VOL. 2, NO. 12, DECEMBER 2014, 485 490 Available online at: www.ijcncs.org ISSN 2308-9830 A survey on Data Mining based Intrusion
Design call center management system of e-commerce based on BP neural network and multifractal
Available online www.jocpr.com Journal of Chemical and Pharmaceutical Research, 2014, 6(6):951-956 Research Article ISSN : 0975-7384 CODEN(USA) : JCPRC5 Design call center management system of e-commerce
Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets
Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets Macario O. Cordel II and Arnulfo P. Azcarraga College of Computer Studies *Corresponding Author: [email protected]
EVALUATION OF NEURAL NETWORK BASED CLASSIFICATION SYSTEMS FOR CLINICAL CANCER DATA CLASSIFICATION
EVALUATION OF NEURAL NETWORK BASED CLASSIFICATION SYSTEMS FOR CLINICAL CANCER DATA CLASSIFICATION K. Mumtaz Vivekanandha Institute of Information and Management Studies, Tiruchengode, India S.A.Sheriff
A Review of Data Mining Techniques
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 3, Issue. 4, April 2014,
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
Distributed forests for MapReduce-based machine learning
Distributed forests for MapReduce-based machine learning Ryoji Wakayama, Ryuei Murata, Akisato Kimura, Takayoshi Yamashita, Yuji Yamauchi, Hironobu Fujiyoshi Chubu University, Japan. NTT Communication
Neural Networks in Data Mining
IOSR Journal of Engineering (IOSRJEN) ISSN (e): 2250-3021, ISSN (p): 2278-8719 Vol. 04, Issue 03 (March. 2014), V6 PP 01-06 www.iosrjen.org Neural Networks in Data Mining Ripundeep Singh Gill, Ashima Department
Ensembles and PMML in KNIME
Ensembles and PMML in KNIME Alexander Fillbrunn 1, Iris Adä 1, Thomas R. Gabriel 2 and Michael R. Berthold 1,2 1 Department of Computer and Information Science Universität Konstanz Konstanz, Germany [email protected]
Data Mining using Artificial Neural Network Rules
Data Mining using Artificial Neural Network Rules Pushkar Shinde MCOERC, Nasik Abstract - Diabetes patients are increasing in number so it is necessary to predict, treat and diagnose the disease. Data
Programming Exercise 3: Multi-class Classification and Neural Networks
Programming Exercise 3: Multi-class Classification and Neural Networks Machine Learning November 4, 2011 Introduction In this exercise, you will implement one-vs-all logistic regression and neural networks
Lecture 10: Regression Trees
Lecture 10: Regression Trees 36-350: Data Mining October 11, 2006 Reading: Textbook, sections 5.2 and 10.5. The next three lectures are going to be about a particular kind of nonlinear predictive model,
TOWARDS SIMPLE, EASY TO UNDERSTAND, AN INTERACTIVE DECISION TREE ALGORITHM
TOWARDS SIMPLE, EASY TO UNDERSTAND, AN INTERACTIVE DECISION TREE ALGORITHM Thanh-Nghi Do College of Information Technology, Cantho University 1 Ly Tu Trong Street, Ninh Kieu District Cantho City, Vietnam
