Neural networks in data analysis
|
|
|
- Victoria George
- 10 years ago
- Views:
Transcription
1 ISAPP Summer Institute 2009 Neural networks in data analysis Michal Kreps ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 1/38
2 Outline What are the neural networks 1 Basic principles of NN 2 For what are they useful 3 4 Training of NN 5 1 Available packages 6 Comparison to other methods 7 Some examples of NN usage 8 input layer hidden layer output layer I m not expert on neural networks, only advanced user who understands principles. Goal: Help you to understand principles so that: neural networks don t look as black box to you you have starting point to use neural networks by yourself ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 2/38
3 Different tasks in nature Tasks to solve Some easy to solve algorithmically, some difficult or impossible Reading hand written text Lot of personal specifics Recognition often needs to work only with partial information Practically impossible in algorithmic way But our brains can solve this tasks extremely fast Use brain as inspiration ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 3/38
4 Tasks to solve Why (astro-)particle physicist would care about recognition of hand written text? Because it is classification task on not very regular data Do we see now letter A or H? Classification tasks are very often in our analysis Is this event signal or background? Neural networks have some advantages in this task: Can learn from known examples (data or simulation) Can evolve to complex non-linear models Easily can deal with complicated correlations amongs inputs Going to more inputs does not cost additional examples ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 4/38
5 Down to principles Detailed look shows huge number of neurons Neurons are interconnected, each having many inputs from other neurons Zoom in To build artificial neural network: Build model of neuron Connect many neurons together ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 5/38
6 Model of neuron Get inputs Give output Combine input to single quantity ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 6/38
7 Model of neuron Get inputs Give output Combine input to single quantity ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 7/38
8 Model of neuron Get inputs Give output Combine input to single quantity S = i w ix i ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 7/38
9 Model of neuron Get inputs Give output Pass S through activation function Most common: 2 A(x) = 1+e 1 x Combine input to single quantity S = i w ix i ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 7/38
10 Building neural network Complicated structure ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 8/38
11 Building neural network Even more complicated structure ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 8/38
12 Building neural network Or simple structures input layer hidden layer output layer ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 8/38
13 Output of Feed-forward NN Output of node k in second layer: o k = A ( i w 1 2 ik x i ) Output of node j in output layer: ) o j = A( k w kj 2 3 x k o j = A[ k w kj 2 3 A ( i w ) ] ik 1 2 x i 6 More layers usually not needed 7 If more layers involved, output follows same logic 8 input layer hidden layer output layer In following, stick to only three layers as here Knowledge is stored in weights ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 9/38
14 Tasks we can solve Classification Binary decision whether event belongs to one of the two classes Is this particle kaon or not? Is this event candidate for Higgs? Will Germany become soccer world champion in 2010? Output layer contains single node Conditional probability density Get probability density for each event What is decay time of a decay What is energy of particle given the observed shower What is probability that customer will cause car accident Output layer contains many nodes ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 10/38
15 NN work flow ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 11/38
16 Neural network training After assembling neural networks, weights wik 1 2 unknown Initialised to default value Random value and w 2 3 kj Before using neural network for planned task, need to find best set of weights Need some measure, which tells which set of weights performs best loss function Training is process, where based on known examples of events we search for set of weights which minimise loss function Most important part in using neural networks Bad training can result in bad or even wrong solutions Many traps in the process, so need to be careful ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 12/38
17 Loss function Practically any function which allows you to decide, which neural network is better In particle physics it can be one which results in best measurement In practical applications, two commonly used functions (T ji is true value with 1 for one class and 1 for other, o ji is actual neural network output) Quadratic loss function χ 2 = j w jχ 2 j = 1 j 2 i (T ji o ji ) 2 χ 2 = 0 for correct decisions and χ 2 = 2 for wrong Entropy loss function E D = j w je j D = j i log( 1 2 (1 + T jio ji )) E D = 0 for correct decisions and E D = for wrong ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 13/38
18 Conditional probability density Classification is easy case, we need single output node and if output is above chosen value, it is class 1 (signal) otherwise class 2 (background) Now question is how to obtain probability density? Use many nodes in output layer, each answering question, whether probability is larger than some value As example, lets have n = 10 nodes and true value of 0.63 Vector of true values for output nodes: (1, 1, 1, 1, 1, 1, 1, 1, 1, 1) Node j answers question whether probability is larger than j/10 The output vector that represents integrated probability density After differentiation we obtain desired probability density ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 14/38
19 Interpretation of NN output In classification tasks in general higher NN output means more signal like event NN output rescaled to interval [0; 1] can be interpreted as probability Quadratic/entropy loss function is used NN is well trained loss function minimised Purity P(o) = N selected signal (o)/n selected (o) should be linear function of NN output Looking to expression for output o j = A[ k w kj 2 3 A ( i w ) ] ik 1 2 x i NN is function, which maps n- dimensional space to 1-dimensional space purity Network output ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 15/38
20 How to check quality of training Define quantities: Signal efficiency: N(selected signal) ǫ s = N(all signal) Efficiency: ǫ = N(selected) N(all) Purity: N(selected signal) P = N(selected) ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 16/38
21 How to check quality of training Define quantities: Signal efficiency: N(selected signal) ǫ s = N(all signal) Efficiency: ǫ = N(selected) N(all) Purity: N(selected signal) P = N(selected) Distribution of NN output for two classes Network output How good is separation? ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 16/38
22 How to check quality of training Define quantities: Signal efficiency: N(selected signal) ǫ s = N(all signal) Efficiency: ǫ = N(selected) N(all) Purity: N(selected signal) P = N(selected) purity Network output Are we at minimum of the loss function? ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 16/38
23 How to check quality of training Define quantities: Signal efficiency: N(selected signal) ǫ s = N(all signal) Efficiency: ǫ = N(selected) N(all) Purity: N(selected signal) P = N(selected) Signal purity Each point has requirement on out > cut Best point is (1,1) Training S/B Signal efficiency Each point has requirement on out < cut ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 16/38
24 How to check quality of training Define quantities: Signal efficiency: N(selected signal) ǫ s = N(all signal) Efficiency: ǫ = N(selected) N(all) Purity: N(selected signal) P = N(selected) Signal efficiency Each point has requirement on out > cut Best achievable Random decision Efficiency ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 16/38
25 Issues in training Main issue: Finding minimum in high dimensional space 5 input nodes with 5 nodes in hidden layer and one output node gives 5x5 + 5 = 30 weights Imagine task to find deepest valley in the Alps (only 2 dimensions) Being put to some random place, easy to find some minimum ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 17/38
26 Issues in training Main issue: Finding minimum in high dimensional space 5 input nodes with 5 nodes in hidden layer and one output node gives 5x5 + 5 = 30 weights Imagine task to find deepest valley in the Alps (only 2 dimensions) Being put to some random place, easy to find some minimum But what is in next valley? ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 17/38
27 Issues in training In training, we start with: Large number of weights which are far from optimal value Input values in unknown ranges Large chance for numerical issues in calculation of loss function Statistical issues coming from low statistics in training In many places in physics not that much issue as training statistics can be large In industry can be quite big issue to obtain large enough training dataset But also in physics it can be sometimes very costly to obtain large enough training sample Learning by heart, when NN has enough weights to do that ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 18/38
28 How to treat issues All issues mentioned can be treated by neural network package in principle Regularisation for numerical issues at the beginning of training Weight decay - forgetting Helps in overtraining on statistical fluctuations Idea is that effects which are not significant are not shown often enough to keep weight to remember them Pruning of connections - remove connections, which don t carry significant information Preprocessing numerical issues search for good minimum by good starting point and decorrelation ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 19/38
29 Neurobayes preprocessing events input node 5 flat Bin to bins of equal statistics 1.2 spline fit purity bin n Calculate purity for each bin and do spline fit final events Use purity instead of original value, plus transform to µ = 0 and RMS = 1 ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 20/38
30 Electron ID at CDF Comparing two different NN packages with exactly same inputs Preprocessing is important and can improve result significantly ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 21/38
31 Not exhaustive list Available packages NeuroBayes Best I know Commercial, but fee is small for scientific use Developed originally in high energy physics experiment ROOT/TMVA Part of the ROOT package Can be used for playing, but I would advise to use something else for serious applications SNNS: Stuttgart Neural Network Simulator Written in C, but ROOT interface exist Best open source package I know about ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 22/38
32 NN vs. other multivariate techniques Several times I saw statements of type, my favourite method is better than other multivariate techniques Usually lot of time is spend to understand and properly train favourite method Other methods are tried just for quick comparison without understanding or tuning Often you will even not learn any details of how other methods were setup With TMVA it is now easy to compare different methods - often I saw that Fischer discriminant won in those From some tests we found that NN implemented in TMVA is relativily bad No wonder that it is usually worst than other methods Usually it is useful to understand method you are using, none of them is black box ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 23/38
33 Practical tips No need for several hidden layers, one should be sufficient for you Number of nodes in hidden layer should not be very large Risk of learning by heart Number of nodes in hidden layer should not be very small Not enough capacity to learn your problem Usually O(number of inputs) in hidden layer Understand your problem, NN gets numbers and can dig information out of them, but you know meaning of those numbers ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 24/38
34 Practical tips Don t throw away knowledge you acquired about problem before starting using NN NN can do lot, but you don t need that it discovers what you already know If you already know that given event is background, throw it away If you know about good inputs, use them rather than raw data from which they are calculated Use neural network to learn what you don t know yet ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 25/38
35 Physics examples Only mentioning work done by Karlsruhe HEP group "Stable" particle identification at Delphi, CDF and now Belle Properties (E, φ, θ, Q-value) of inclusively reconstructed B mesons Decay time in semileptonic B s decays Heavy quark resonances studies: X(3872), B, B s, Λ c, Σ c B flavour tagging and B s mixing (Delphi, CDF, Belle) CPV in B s J/ψφ decays (CDF) B tagging for high p T physics (CDF, CMS) Single top and Higgs search (CDF) B hadron full reconstruction (Belle) Optimisation of KEKB accelerator ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 26/38
36 B s at CDF CDF Run fb CDF Run fb Candidates ** B s B K [J/ψ K ] K Signal Background 2 Candidates per 1.25 MeV/c B K B s1 B s2 + + B K Signal Background Neural Network Output M(B K )-M(B )-M(K ) [GeV/c ] Background from data Signal from MC First observation of B s1 Most precise masses ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 27/38
37 X(3872) mass at CDF Candidates CDF II Preliminary Signal Background Network Output fb Powerful selection at CDF of largest sample leading to best mass measurement X(3872) first of charmonium like states Lot of effort to understand its nature 2 Candidates per 2 MeV/c CDF II Preliminary Total Rejected Selected J/ψππ Mass (GeV/c ) fb ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 28/38
38 Single top search Event Fraction TLC 2Jets 1Tag -1 CDF II Preliminary 3.2 fb single top tt Wbb+Wcc Wc Wqq Diboson Z+jets QCD KIT Flavor Sep. Output Final data plot normalized to unit area Part of big program leading to discovery of single top production Complicated analysis with many different multivariate techniques Tagging whether jet is b-quark jet or not using NN Candidate Events All Channels CDF II Preliminary 3.2 fb single top 80 tt 70 Wbb+Wcc 60 Wc 50 Wqq 40 Diboson 30 Z+jets 20 QCD 10 data NN Output MC normalized to SM prediction ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 29/38
39 B flavour tagging - CDF Flavour tagging at hadron colliders is difficult task Lot of potential for improvements In plot, concentrate on continuous lines Example of not that good separation ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 30/38
40 Training with weights Candidates per 0.4 MeV CDF Run II preliminary, L = 2.4 fb #Signal Events #Background Events Mass(Λ c π)-mass(λ c ) [GeV/c ] One knows statistically whether event is signal or background Or has sample of mixture of signal and background and pure sample for one class Can use data only as example of application ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 31/38
41 Training with weights Candidates per 0.4 MeV CDF Run II preliminary, L = 2.4 fb #Signal Events #Background Events Mass(Λ c π)-mass(λ c ) [GeV/c ] Candidates per 1.30 MeV CDF Run II preliminary, L = 2.4 fb Data Fit Function ++ Σ c (2455) Λ c ++ + Σ c (2520) Λ c + π + π Use only data in selection Can compare data to simulation Can reweight simulation to match data (data - fit) data Mass(Λ c π + + )-Mass(Λ c ) [GeV/c 2 ] ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 32/38
42 Examples outside physics Medicine and Pharma research e.g. effects and undesirable effects of drugs early tumor recognition Banks e.g. credit-scoring (Basel II), finance time series prediction, valuation of derivates, risk minimised trading strategies, client valuation Insurances e.g. risk and cost prediction for individual clients, probability of contract cancellation, fraud recognition, justice in tariffs Trading chain stores: turnover prognosis for individual articles/stores ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 33/38
43 World s largest students competition: Data-Mining-Cup Very successful in competition with other data-mining methods 2005: Fraud detection in internet trading 2006: price prediction in ebay auctions 2007: coupon redemption prediction 2008: lottery customer behaviour prediction Data Mining Cup ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 34/38
44 NeuroBayes Turnover prognosis ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 35/38
45 Individual health costs Pilot project for a large private health insurance Prognosis of costs in following year for each person insured with confidence intervals 4 years of training, test on following year Results: Probability density for each customer/tariff combination Very good test results! Potential for an objective cost reduction in health management ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 36/38
46 Prognosis of sport events Results: Probabilities for home - tie - guest ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 37/38
47 Conclusions Neural network is powerful tool for your analysis They are not genius, but can do marvellous things for genius Lecture of this kind cannot really make you expert Hope that I was successful with: Give you enough insight so that neural network is not black box anymore Teaching you enough not to be afraid to try Feel free to talk to me, I m at Campus Süd, building 30.23, room 9-11 ISAPP Summer Institute 2009 M. Kreps, KIT Neural networks in data analysis p. 38/38
Maximum Likelihood vs. Least Squares
Precondition Basis Maximum Likelihood vs. Least Squares pdf exactly known Height of pdf Mean and variance known Deviation from mean Efficiency Complexity Robustness Correlated measurements Special case
Top rediscovery at ATLAS and CMS
Top rediscovery at ATLAS and CMS on behalf of ATLAS and CMS collaborations CNRS/IN2P3 & UJF/ENSPG, LPSC, Grenoble, France E-mail: [email protected] We describe the plans and strategies of the
Spin-Off from Physics Research to Business
From Delphi to Phi-T Spin-Off from Physics Research to Business Prof. Dr. Michael Feindt KCETA - Centrum für Elementarteilchen- und Astroteilchenphysik IEKP, Universität Karlsruhe, Karlsruhe Institute
D A T A M I N I N G C L A S S I F I C A T I O N
D A T A M I N I N G C L A S S I F I C A T I O N FABRICIO VOZNIKA LEO NARDO VIA NA INTRODUCTION Nowadays there is huge amount of data being collected and stored in databases everywhere across the globe.
Measurement of the Mass of the Top Quark in the l+ Jets Channel Using the Matrix Element Method
Measurement of the Mass of the Top Quark in the l+ Jets Channel Using the Matrix Element Method Carlos Garcia University of Rochester For the DØ Collaboration APS Meeting 2007 Outline Introduction Top
Gerard Mc Nulty Systems Optimisation Ltd [email protected]/0876697867 BA.,B.A.I.,C.Eng.,F.I.E.I
Gerard Mc Nulty Systems Optimisation Ltd [email protected]/0876697867 BA.,B.A.I.,C.Eng.,F.I.E.I Data is Important because it: Helps in Corporate Aims Basis of Business Decisions Engineering Decisions Energy
Neural Networks and Back Propagation Algorithm
Neural Networks and Back Propagation Algorithm Mirza Cilimkovic Institute of Technology Blanchardstown Blanchardstown Road North Dublin 15 Ireland [email protected] Abstract Neural Networks (NN) are important
Theory versus Experiment. Prof. Jorgen D Hondt Vrije Universiteit Brussel [email protected]
Theory versus Experiment Prof. Jorgen D Hondt Vrije Universiteit Brussel [email protected] Theory versus Experiment Pag. 2 Dangerous cocktail!!! Pag. 3 The basics in these lectures Part1 : Theory meets
How To Use Neural Networks In Data Mining
International Journal of Electronics and Computer Science Engineering 1449 Available Online at www.ijecse.org ISSN- 2277-1956 Neural Networks in Data Mining Priyanka Gaur Department of Information and
Data Mining for Business Analytics
Data Mining for Business Analytics Lecture 2: Introduction to Predictive Modeling Stern School of Business New York University Spring 2014 MegaTelCo: Predicting Customer Churn You just landed a great analytical
Data Mining and Neural Networks in Stata
Data Mining and Neural Networks in Stata 2 nd Italian Stata Users Group Meeting Milano, 10 October 2005 Mario Lucchini e Maurizo Pisati Università di Milano-Bicocca [email protected] [email protected]
Chapter 12 Discovering New Knowledge Data Mining
Chapter 12 Discovering New Knowledge Data Mining Becerra-Fernandez, et al. -- Knowledge Management 1/e -- 2004 Prentice Hall Additional material 2007 Dekai Wu Chapter Objectives Introduce the student to
EFFICIENT DATA PRE-PROCESSING FOR DATA MINING
EFFICIENT DATA PRE-PROCESSING FOR DATA MINING USING NEURAL NETWORKS JothiKumar.R 1, Sivabalan.R.V 2 1 Research scholar, Noorul Islam University, Nagercoil, India Assistant Professor, Adhiparasakthi College
TOWARDS SIMPLE, EASY TO UNDERSTAND, AN INTERACTIVE DECISION TREE ALGORITHM
TOWARDS SIMPLE, EASY TO UNDERSTAND, AN INTERACTIVE DECISION TREE ALGORITHM Thanh-Nghi Do College of Information Technology, Cantho University 1 Ly Tu Trong Street, Ninh Kieu District Cantho City, Vietnam
Statistics for BIG data
Statistics for BIG data Statistics for Big Data: Are Statisticians Ready? Dennis Lin Department of Statistics The Pennsylvania State University John Jordan and Dennis K.J. Lin (ICSA-Bulletine 2014) Before
Calorimetry in particle physics experiments
Calorimetry in particle physics experiments Unit n. 8 Calibration techniques Roberta Arcidiacono Lecture overview Introduction Hardware Calibration Test Beam Calibration In-situ Calibration (EM calorimeters)
Data Mining mit der JMSL Numerical Library for Java Applications
Data Mining mit der JMSL Numerical Library for Java Applications Stefan Sineux 8. Java Forum Stuttgart 07.07.2005 Agenda Visual Numerics JMSL TM Numerical Library Neuronale Netze (Hintergrund) Demos Neuronale
Neural Networks and Support Vector Machines
INF5390 - Kunstig intelligens Neural Networks and Support Vector Machines Roar Fjellheim INF5390-13 Neural Networks and SVM 1 Outline Neural networks Perceptrons Neural networks Support vector machines
Data Mining Techniques Chapter 7: Artificial Neural Networks
Data Mining Techniques Chapter 7: Artificial Neural Networks Artificial Neural Networks.................................................. 2 Neural network example...................................................
Artificial Neural Networks and Support Vector Machines. CS 486/686: Introduction to Artificial Intelligence
Artificial Neural Networks and Support Vector Machines CS 486/686: Introduction to Artificial Intelligence 1 Outline What is a Neural Network? - Perceptron learners - Multi-layer networks What is a Support
A Property & Casualty Insurance Predictive Modeling Process in SAS
Paper AA-02-2015 A Property & Casualty Insurance Predictive Modeling Process in SAS 1.0 ABSTRACT Mei Najim, Sedgwick Claim Management Services, Chicago, Illinois Predictive analytics has been developing
Study of the B D* ℓ ν with the Partial Reconstruction Technique
Study of the B D* ℓ ν with the Partial Reconstruction Technique + University of Ferrara / INFN Ferrara Dottorato di Ricerca in Fisica Ciclo XVII Mirco Andreotti 4 March 25 Measurement of B(B D*ℓν) from
Modelling and Big Data. Leslie Smith ITNPBD4, October 10 2015. Updated 9 October 2015
Modelling and Big Data Leslie Smith ITNPBD4, October 10 2015. Updated 9 October 2015 Big data and Models: content What is a model in this context (and why the context matters) Explicit models Mathematical
Data analysis in Par,cle Physics
Data analysis in Par,cle Physics From data taking to discovery Tuesday, 13 August 2013 Lukasz Kreczko - Bristol IT MegaMeet 1 $ whoami Lukasz (Luke) Kreczko Par,cle Physicist Graduated in Physics from
Implications of CMS searches for the Constrained MSSM A Bayesian approach
Implications of CMS searches for the Constrained MSSM A Bayesian approach Małgorzata Kazana, Yue-Lin Sming Tsai On behalf of the BayesFITS group National Centre for Nuclear Research Warsaw, Poland BayesFITS,
degrees of freedom and are able to adapt to the task they are supposed to do [Gupta].
1.3 Neural Networks 19 Neural Networks are large structured systems of equations. These systems have many degrees of freedom and are able to adapt to the task they are supposed to do [Gupta]. Two very
Classification Problems
Classification Read Chapter 4 in the text by Bishop, except omit Sections 4.1.6, 4.1.7, 4.2.4, 4.3.3, 4.3.5, 4.3.6, 4.4, and 4.5. Also, review sections 1.5.1, 1.5.2, 1.5.3, and 1.5.4. Classification Problems
Stabilization by Conceptual Duplication in Adaptive Resonance Theory
Stabilization by Conceptual Duplication in Adaptive Resonance Theory Louis Massey Royal Military College of Canada Department of Mathematics and Computer Science PO Box 17000 Station Forces Kingston, Ontario,
Testing the In-Memory Column Store for in-database physics analysis. Dr. Maaike Limper
Testing the In-Memory Column Store for in-database physics analysis Dr. Maaike Limper About CERN CERN - European Laboratory for Particle Physics Support the research activities of 10 000 scientists from
NEURAL NETWORKS IN DATA MINING
NEURAL NETWORKS IN DATA MINING 1 DR. YASHPAL SINGH, 2 ALOK SINGH CHAUHAN 1 Reader, Bundelkhand Institute of Engineering & Technology, Jhansi, India 2 Lecturer, United Institute of Management, Allahabad,
Social Media Mining. Data Mining Essentials
Introduction Data production rate has been increased dramatically (Big Data) and we are able store much more data than before E.g., purchase data, social media data, mobile phone data Businesses and customers
Lecture 6. Artificial Neural Networks
Lecture 6 Artificial Neural Networks 1 1 Artificial Neural Networks In this note we provide an overview of the key concepts that have led to the emergence of Artificial Neural Networks as a major paradigm
Real Time Tracking with ATLAS Silicon Detectors and its Applications to Beauty Hadron Physics
Real Time Tracking with ATLAS Silicon Detectors and its Applications to Beauty Hadron Physics Carlo Schiavi Dottorato in Fisica - XVII Ciclo Outline The ATLAS Experiment The SiTrack Algorithm Application
Data Mining Algorithms Part 1. Dejan Sarka
Data Mining Algorithms Part 1 Dejan Sarka Join the conversation on Twitter: @DevWeek #DW2015 Instructor Bio Dejan Sarka ([email protected]) 30 years of experience SQL Server MVP, MCT, 13 books 7+ courses
INTELLIGENT ENERGY MANAGEMENT OF ELECTRICAL POWER SYSTEMS WITH DISTRIBUTED FEEDING ON THE BASIS OF FORECASTS OF DEMAND AND GENERATION Chr.
INTELLIGENT ENERGY MANAGEMENT OF ELECTRICAL POWER SYSTEMS WITH DISTRIBUTED FEEDING ON THE BASIS OF FORECASTS OF DEMAND AND GENERATION Chr. Meisenbach M. Hable G. Winkler P. Meier Technology, Laboratory
Measurement of Neutralino Mass Differences with CMS in Dilepton Final States at the Benchmark Point LM9
Measurement of Neutralino Mass Differences with CMS in Dilepton Final States at the Benchmark Point LM9, Katja Klein, Lutz Feld, Niklas Mohr 1. Physikalisches Institut B RWTH Aachen Introduction Fast discovery
Using Data Mining for Mobile Communication Clustering and Characterization
Using Data Mining for Mobile Communication Clustering and Characterization A. Bascacov *, C. Cernazanu ** and M. Marcu ** * Lasting Software, Timisoara, Romania ** Politehnica University of Timisoara/Computer
Joseph Twagilimana, University of Louisville, Louisville, KY
ST14 Comparing Time series, Generalized Linear Models and Artificial Neural Network Models for Transactional Data analysis Joseph Twagilimana, University of Louisville, Louisville, KY ABSTRACT The aim
Blue Yonder Research Papers
Blue Yonder Research Papers Why cutting edge technology matters for Blue Yonder solutions Prof. Dr. Michael Feindt, Chief Scientific Advisor Abstract This article gives an overview of the stack of predictive
Data Mining. 1 Introduction 2 Data Mining methods. Alfred Holl Data Mining 1
Data Mining 1 Introduction 2 Data Mining methods Alfred Holl Data Mining 1 1 Introduction 1.1 Motivation 1.2 Goals and problems 1.3 Definitions 1.4 Roots 1.5 Data Mining process 1.6 Epistemological constraints
MACHINE LEARNING IN HIGH ENERGY PHYSICS
MACHINE LEARNING IN HIGH ENERGY PHYSICS LECTURE #1 Alex Rogozhnikov, 2015 INTRO NOTES 4 days two lectures, two practice seminars every day this is introductory track to machine learning kaggle competition!
How To Make A Credit Risk Model For A Bank Account
TRANSACTIONAL DATA MINING AT LLOYDS BANKING GROUP Csaba Főző [email protected] 15 October 2015 CONTENTS Introduction 04 Random Forest Methodology 06 Transactional Data Mining Project 17 Conclusions
Government of Russian Federation. Faculty of Computer Science School of Data Analysis and Artificial Intelligence
Government of Russian Federation Federal State Autonomous Educational Institution of High Professional Education National Research University «Higher School of Economics» Faculty of Computer Science School
DATA MINING TECHNIQUES AND APPLICATIONS
DATA MINING TECHNIQUES AND APPLICATIONS Mrs. Bharati M. Ramageri, Lecturer Modern Institute of Information Technology and Research, Department of Computer Application, Yamunanagar, Nigdi Pune, Maharashtra,
From Jet Scaling to Jet Vetos
From Jet Scaling to Jet Vetos Heidelberg DESY, 2/202 LHC Higgs analyses Two problems for LHC Higgs analyses [talks Rauch, Englert] observe H b b decays [fat Higgs jets, Marcel s talk] 2 understand jet
Azure Machine Learning, SQL Data Mining and R
Azure Machine Learning, SQL Data Mining and R Day-by-day Agenda Prerequisites No formal prerequisites. Basic knowledge of SQL Server Data Tools, Excel and any analytical experience helps. Best of all:
Neural Network and Genetic Algorithm Based Trading Systems. Donn S. Fishbein, MD, PhD Neuroquant.com
Neural Network and Genetic Algorithm Based Trading Systems Donn S. Fishbein, MD, PhD Neuroquant.com Consider the challenge of constructing a financial market trading system using commonly available technical
Weather forecast prediction: a Data Mining application
Weather forecast prediction: a Data Mining application Ms. Ashwini Mandale, Mrs. Jadhawar B.A. Assistant professor, Dr.Daulatrao Aher College of engg,karad,[email protected],8407974457 Abstract
Combining GLM and datamining techniques for modelling accident compensation data. Peter Mulquiney
Combining GLM and datamining techniques for modelling accident compensation data Peter Mulquiney Introduction Accident compensation data exhibit features which complicate loss reserving and premium rate
Neural Networks for Machine Learning. Lecture 13a The ups and downs of backpropagation
Neural Networks for Machine Learning Lecture 13a The ups and downs of backpropagation Geoffrey Hinton Nitish Srivastava, Kevin Swersky Tijmen Tieleman Abdel-rahman Mohamed A brief history of backpropagation
An Introduction to Data Mining. Big Data World. Related Fields and Disciplines. What is Data Mining? 2/12/2015
An Introduction to Data Mining for Wind Power Management Spring 2015 Big Data World Every minute: Google receives over 4 million search queries Facebook users share almost 2.5 million pieces of content
Back Propagation Neural Networks User Manual
Back Propagation Neural Networks User Manual Author: Lukáš Civín Library: BP_network.dll Runnable class: NeuralNetStart Document: Back Propagation Neural Networks Page 1/28 Content: 1 INTRODUCTION TO BACK-PROPAGATION
STA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! [email protected]! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct
Cross section, Flux, Luminosity, Scattering Rates
Cross section, Flux, Luminosity, Scattering Rates Table of Contents Paul Avery (Andrey Korytov) Sep. 9, 013 1 Introduction... 1 Cross section, flux and scattering... 1 3 Scattering length λ and λ ρ...
How To Find The Higgs Boson
Dezső Horváth: Search for Higgs bosons Balaton Summer School, Balatongyörök, 07.07.2009 p. 1/25 Search for Higgs bosons Balaton Summer School, Balatongyörök, 07.07.2009 Dezső Horváth MTA KFKI Research
How to use Big Data in Industry 4.0 implementations. LAURI ILISON, PhD Head of Big Data and Machine Learning
How to use Big Data in Industry 4.0 implementations LAURI ILISON, PhD Head of Big Data and Machine Learning Big Data definition? Big Data is about structured vs unstructured data Big Data is about Volume
Network Machine Learning Research Group. Intended status: Informational October 19, 2015 Expires: April 21, 2016
Network Machine Learning Research Group S. Jiang Internet-Draft Huawei Technologies Co., Ltd Intended status: Informational October 19, 2015 Expires: April 21, 2016 Abstract Network Machine Learning draft-jiang-nmlrg-network-machine-learning-00
!"!!"#$$%&'()*+$(,%!"#$%$&'()*""%(+,'-*&./#-$&'(-&(0*".$#-$1"(2&."3$'45"
!"!!"#$$%&'()*+$(,%!"#$%$&'()*""%(+,'-*&./#-$&'(-&(0*".$#-$1"(2&."3$'45"!"#"$%&#'()*+',$$-.&#',/"-0%.12'32./4'5,5'6/%&)$).2&'7./&)8'5,5'9/2%.%3%&8':")08';:
Applied Data Mining Analysis: A Step-by-Step Introduction Using Real-World Data Sets
Applied Data Mining Analysis: A Step-by-Step Introduction Using Real-World Data Sets http://info.salford-systems.com/jsm-2015-ctw August 2015 Salford Systems Course Outline Demonstration of two classification
DATA MINING TECHNIQUES SUPPORT TO KNOWLEGDE OF BUSINESS INTELLIGENT SYSTEM
INTERNATIONAL JOURNAL OF RESEARCH IN COMPUTER APPLICATIONS AND ROBOTICS ISSN 2320-7345 DATA MINING TECHNIQUES SUPPORT TO KNOWLEGDE OF BUSINESS INTELLIGENT SYSTEM M. Mayilvaganan 1, S. Aparna 2 1 Associate
Using reporting and data mining techniques to improve knowledge of subscribers; applications to customer profiling and fraud management
Using reporting and data mining techniques to improve knowledge of subscribers; applications to customer profiling and fraud management Paper Jean-Louis Amat Abstract One of the main issues of operators
Method of Combining the Degrees of Similarity in Handwritten Signature Authentication Using Neural Networks
Method of Combining the Degrees of Similarity in Handwritten Signature Authentication Using Neural Networks Ph. D. Student, Eng. Eusebiu Marcu Abstract This paper introduces a new method of combining the
Machine Learning for Data Science (CS4786) Lecture 1
Machine Learning for Data Science (CS4786) Lecture 1 Tu-Th 10:10 to 11:25 AM Hollister B14 Instructors : Lillian Lee and Karthik Sridharan ROUGH DETAILS ABOUT THE COURSE Diagnostic assignment 0 is out:
Potential Value of Data Mining for Customer Relationship Marketing in the Banking Industry
Advances in Natural and Applied Sciences, 3(1): 73-78, 2009 ISSN 1995-0772 2009, American Eurasian Network for Scientific Information This is a refereed journal and all articles are professionally screened
Intrusion Detection via Machine Learning for SCADA System Protection
Intrusion Detection via Machine Learning for SCADA System Protection S.L.P. Yasakethu Department of Computing, University of Surrey, Guildford, GU2 7XH, UK. [email protected] J. Jiang Department
Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.
Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C
Data, Measurements, Features
Data, Measurements, Features Middle East Technical University Dep. of Computer Engineering 2009 compiled by V. Atalay What do you think of when someone says Data? We might abstract the idea that data are
Lecture 8 February 4
ICS273A: Machine Learning Winter 2008 Lecture 8 February 4 Scribe: Carlos Agell (Student) Lecturer: Deva Ramanan 8.1 Neural Nets 8.1.1 Logistic Regression Recall the logistic function: g(x) = 1 1 + e θt
Statistical Data Mining. Practical Assignment 3 Discriminant Analysis and Decision Trees
Statistical Data Mining Practical Assignment 3 Discriminant Analysis and Decision Trees In this practical we discuss linear and quadratic discriminant analysis and tree-based classification techniques.
International Journal of Computer Science Trends and Technology (IJCST) Volume 2 Issue 3, May-Jun 2014
RESEARCH ARTICLE OPEN ACCESS A Survey of Data Mining: Concepts with Applications and its Future Scope Dr. Zubair Khan 1, Ashish Kumar 2, Sunny Kumar 3 M.Tech Research Scholar 2. Department of Computer
PHYSICS WITH LHC EARLY DATA
PHYSICS WITH LHC EARLY DATA ONE OF THE LAST PROPHETIC TALKS ON THIS SUBJECT HOPEFULLY We may have some two month of the Machine operation in 2008 LONG HISTORY... I will extensively use: Fabiola GIANOTTI
Machine learning for algo trading
Machine learning for algo trading An introduction for nonmathematicians Dr. Aly Kassam Overview High level introduction to machine learning A machine learning bestiary What has all this got to do with
Comparison of K-means and Backpropagation Data Mining Algorithms
Comparison of K-means and Backpropagation Data Mining Algorithms Nitu Mathuriya, Dr. Ashish Bansal Abstract Data mining has got more and more mature as a field of basic research in computer science and
Web Mining using Artificial Ant Colonies : A Survey
Web Mining using Artificial Ant Colonies : A Survey Richa Gupta Department of Computer Science University of Delhi ABSTRACT : Web mining has been very crucial to any organization as it provides useful
Data quality in Accounting Information Systems
Data quality in Accounting Information Systems Comparing Several Data Mining Techniques Erjon Zoto Department of Statistics and Applied Informatics Faculty of Economy, University of Tirana Tirana, Albania
Data Mining Techniques
15.564 Information Technology I Business Intelligence Outline Operational vs. Decision Support Systems What is Data Mining? Overview of Data Mining Techniques Overview of Data Mining Process Data Warehouses
ARTIFICIAL INTELLIGENCE (CSCU9YE) LECTURE 6: MACHINE LEARNING 2: UNSUPERVISED LEARNING (CLUSTERING)
ARTIFICIAL INTELLIGENCE (CSCU9YE) LECTURE 6: MACHINE LEARNING 2: UNSUPERVISED LEARNING (CLUSTERING) Gabriela Ochoa http://www.cs.stir.ac.uk/~goc/ OUTLINE Preliminaries Classification and Clustering Applications
Neural Networks for Sentiment Detection in Financial Text
Neural Networks for Sentiment Detection in Financial Text Caslav Bozic* and Detlef Seese* With a rise of algorithmic trading volume in recent years, the need for automatic analysis of financial news emerged.
Self Organizing Maps: Fundamentals
Self Organizing Maps: Fundamentals Introduction to Neural Networks : Lecture 16 John A. Bullinaria, 2004 1. What is a Self Organizing Map? 2. Topographic Maps 3. Setting up a Self Organizing Map 4. Kohonen
TMVA, the Toolkit for Multivariate Data Analysis with ROOT
with ROOT Max-Planck-Institut für Kernphysik, Heidelberg E-mail: [email protected] Andreas Höcker CERN, Switzerland E-mail: [email protected] Jörg Stelzer CERN, Switzerland E-mail: [email protected]
Visualization methods for patent data
Visualization methods for patent data Treparel 2013 Dr. Anton Heijs (CTO & Founder) Delft, The Netherlands Introduction Treparel can provide advanced visualizations for patent data. This document describes
White Paper. Data Mining for Business
White Paper Data Mining for Business January 2010 Contents 1. INTRODUCTION... 3 2. WHY IS DATA MINING IMPORTANT?... 3 FUNDAMENTALS... 3 Example 1...3 Example 2...3 3. OPERATIONAL CONSIDERATIONS... 4 ORGANISATIONAL
Theoretical Particle Physics FYTN04: Oral Exam Questions, version ht15
Theoretical Particle Physics FYTN04: Oral Exam Questions, version ht15 Examples of The questions are roughly ordered by chapter but are often connected across the different chapters. Ordering is as in
MS1b Statistical Data Mining
MS1b Statistical Data Mining Yee Whye Teh Department of Statistics Oxford http://www.stats.ox.ac.uk/~teh/datamining.html Outline Administrivia and Introduction Course Structure Syllabus Introduction to
Practical Data Science with Azure Machine Learning, SQL Data Mining, and R
Practical Data Science with Azure Machine Learning, SQL Data Mining, and R Overview This 4-day class is the first of the two data science courses taught by Rafal Lukawiecki. Some of the topics will be
Physik des Higgs Bosons. Higgs decays V( ) Re( ) Im( ) Figures and calculations from A. Djouadi, Phys.Rept. 457 (2008) 1-216
: Higgs decays V( ) Re( ) Im( ) Figures and calculations from A. Djouadi, Phys.Rept. 457 (2008) 1-216 1 Reminder 10.6.2014 Higgs couplings: 2 Reminder 10.6.2014 Higgs BF as a function of mh Higgs total
Artificial Neural Network, Decision Tree and Statistical Techniques Applied for Designing and Developing E-mail Classifier
International Journal of Recent Technology and Engineering (IJRTE) ISSN: 2277-3878, Volume-1, Issue-6, January 2013 Artificial Neural Network, Decision Tree and Statistical Techniques Applied for Designing
Protein Protein Interaction Networks
Functional Pattern Mining from Genome Scale Protein Protein Interaction Networks Young-Rae Cho, Ph.D. Assistant Professor Department of Computer Science Baylor University it My Definition of Bioinformatics
Top-Quark Studies at CMS
Top-Quark Studies at CMS Tim Christiansen (CERN) on behalf of the CMS Collaboration ICHEP 2010, Paris 35th International Conference on High-Energy Physics tt 2 km 22 28 July 2010 Single-top 4 km New Physics
Keywords data mining, prediction techniques, decision making.
Volume 5, Issue 4, April 2015 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Analysis of Datamining
Application of Neural Network in User Authentication for Smart Home System
Application of Neural Network in User Authentication for Smart Home System A. Joseph, D.B.L. Bong, D.A.A. Mat Abstract Security has been an important issue and concern in the smart home systems. Smart
HT2015: SC4 Statistical Data Mining and Machine Learning
HT2015: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Bayesian Nonparametrics Parametric vs Nonparametric
Lecture 10: Regression Trees
Lecture 10: Regression Trees 36-350: Data Mining October 11, 2006 Reading: Textbook, sections 5.2 and 10.5. The next three lectures are going to be about a particular kind of nonlinear predictive model,
AUTOMATION OF ENERGY DEMAND FORECASTING. Sanzad Siddique, B.S.
AUTOMATION OF ENERGY DEMAND FORECASTING by Sanzad Siddique, B.S. A Thesis submitted to the Faculty of the Graduate School, Marquette University, in Partial Fulfillment of the Requirements for the Degree
Neural Networks in Data Mining
IOSR Journal of Engineering (IOSRJEN) ISSN (e): 2250-3021, ISSN (p): 2278-8719 Vol. 04, Issue 03 (March. 2014), V6 PP 01-06 www.iosrjen.org Neural Networks in Data Mining Ripundeep Singh Gill, Ashima Department
OBJECTIVE ASSESSMENT OF FORECASTING ASSIGNMENTS USING SOME FUNCTION OF PREDICTION ERRORS
OBJECTIVE ASSESSMENT OF FORECASTING ASSIGNMENTS USING SOME FUNCTION OF PREDICTION ERRORS CLARKE, Stephen R. Swinburne University of Technology Australia One way of examining forecasting methods via assignments
