Neural Nets. General Model Building

Save this PDF as:

Size: px
Start display at page:

Transcription

1 Neural Nets To give you an idea of how new this material is, let s do a little history lesson. The origins are typically dated back to the early 1940 s and work by two physiologists, McCulloch and Pitts. Hebb s work came later, when he formulated his rule for learning. In the 1950 s came the perceptron. The perceptron is what we called a linear neural network- it became clear that the perceptron could only classify certain types of data (what we now call linearly separable data). This shortcoming led to a stop in neural net research for many years- It was not really until the 1980 s that neural net research really took off from a coming together of many disparate strands of research- There was a group in Finland led by T. Kohonen that was working with self-organizing maps (SOM); there was the work from the 70s with Stephen Grossberg, and finally several researchers were discovering methods for training new networks that could be put in parallel (back-propagation). It was in the 80 s and 90 s that researchers were able to prove that a neural network was a universal function approximator- That is, for any function with certain nice properties, we are able to find a neural network that will approximate that function with arbitrarily small error. It is interesting to note that the use of a neural network could be used to solve Hilbert s 13th Problem. Every continuous function of two or more real variables can be written as the superposition of continuous functions of one real variable along with addition. As we discussed earlier, the term neural network has come to be an umbrella term covering a broad range of different function approximation or pattern classification methods. The classic type of neural network can be defined simply as a connected graph with weighted edges and computational nodes- We ve seen two types: a linear neural network (using Widrow-Hoff training rule), and the RBF (using Orthogonal Least Squares). We now turn to the workhorse of the neural network community: The feed forward neural network. General Model Building As with our other types of neural networks, we assume that we have p data pairs, (x 1, t 1 ), (x 2, t 2 ),, (x p, t p ) (the letter t is for target), and we are looking to build a function F so that ideally, F (x i ) = t i However, we will typically allow for error, and we typically model the error ɛ i using a normal distribution: t i = F (x i ) + ɛ i 1

2 If we assume that the function F depends on parameters (weights and biases, like the RBF), then we might state it as an optimization problem- Find the function F that minimizes the error function: E = 1 2 p t i F (x i ) 2 i=1 so that E is a function of the parameters of F, and we would go about determining the values of the parameters that minimize the error- And this is what we will do. Earlier in this class, F (x i = W x i + b. This was the case of the linear neural network. More recently, we worked with the RBFs, so that F (x i ) = W φ(edm(x i, centers)) + b We will now work through the construction of the multi-layered feed forward network. The Feed-Forward Neural Network As in the linear network, we will assume that along the dendrites, our signals can be scaled or re-polarized. If we use k neurons, then this is a linear mapping from IR n IR k, and W 1 is a k n matrix. A vector b 1 represents the standing voltage of the neuron, and can be added as a bias term: x W 1 x + b 1 Rather than stopping here, we will now call this the prestate of the layer of neurons. Next, a nonlinear transfer function is applied, σ(r). This is typically called a sigmoidal function because of its shape. The result is a vector in IR k known as the state of the layer of neurons. Adding that to our diagram, we have: x W 1 x + b 1 = P S = σ( P ) Finally, the signal can be recombined in a linear way to produce an output vector y IR m using W 2 that is m k and another bias vector, b 2 IR m : x W 1 x + b 1 = P i S i = σ( P i ) W 2 Si + b 2 Putting it all together, we could write the function F explicitly: F (x i ) = W 2 (σ (W 1 x i + b 1 )) + b 2 so that, with σ defined, F becomes a function of the weights W 1, W 2, and the biases b 1, b 2. 2

3 Defining the Network Architecture We have constructed what many people call a two layer network (although I typically say it is three layers- Some people don t count the input layer as a real layer): The first layer is called the input layer. If x i IR n, then the input layer has n nodes. The next layer is called the hidden layer, and it consists of k nodes (where k is the number of neurons we re using). The mapping from the input layer to the hidden layer is performed by our first affine map, then σ is applied to that vector. The last layer is called the output layer, and if y IR m, then the output layer has m nodes. We did not need to stop with only a single hidden layer- Some researchers like to use two hidden layers as a default neural network. In that case, the mapping (in stages) would look like: x W 1 x + b 1 Prestate of layer 1 σ(w 1 x + b 1 ) State of layer 1 W 2 (σ(w 1 x + b 1 )) + b 2 Prestate of layer 2 σ(w 2 (σ(w 1 x + b 1 )) + b 2 ) State of layer 2 W 3 σ(w 2 (σ(w 1 x + b 1 )) + b 2 ) + b 3 State of layer 3 (output) One can imagine that many layers are possible. However, it has been shown that, at least theoretically, one needs to have only one hidden layer to perform the function approximation to arbitrarily small error. Practically, sometimes it is more efficient to use multiple layers of a small number of neurons than a single layer with a large number of neurons. Definition: The architecture of the neural network is typically defined by stating the number of neurons in each layer. For example, a network has one hidden network of three neurons, and maps IR 2 to IR 4. The parameters of a neural network We have seen that, to define a neural network, we need to define: The weights W 1, W 2 (and more, if we use more layers), the biases b 1, b 2, and the transfer function φ. Although we could define a different φ for every neuron, we typically will use the same transfer function for all the neurons in a single layer. Therefore, when we train a neural network, we will usually pre-define φ, and then we will find the weights and biases that will (hopefully!) minimize the error function. 3

4 E(W 1, W 2, b 1, b 2 ) = 1 2 p t i y i 2 where y i is the output of the neural net. In the case of a single hidden layer, we have: y i = W 2 (σ (W 1 x i + b 1 )) + b 2 The transfer function In a neuron, the incoming signals to the cell body must usually surpass some lowest trigger value before the signal is sent out. A graph of this would be a step function, where the step is at trigger. This is not a good function using notions from Calculus because the voltage function is not continuous and not differentiable at the trigger. We replace the step function by any function that is: Increasing. Differentiable i=1 Has finite horizontal asymptotes at ±. Such a function generally looks like an extended S - We call it a sigmoidal function. There are many ways we could define a sigmoidal, but here are some standard choices (going from most to least used): 1 σ(r) = 1 + e r Matlab calls this the logsig function. Matlab does not use this one. σ(r) = arctan(r) σ(r) = tanh(r) = e2r 1 e 2r + 1 Matlab calls this one tansig. 4

5 Exercises 1. Compute the limits as x ± for the two types of sigmoidal functions that Matlab uses. Show that they are also monotonically increasing functions. 2. Let σ(x) = 1 1+e βx. Show that σ (x) = βσ(x)(1 σ(x)) 3. If x IR n and our targets t IR m, and we use k nodes in the hidden layer, how many unknown parameters do we have to find? As you are constructing your network, keep this number in mind. In particular, you should have at least several data points for each unknown parameter that you are looking for. 4. We have to be somewhat careful when our data is badly scaled. For example, complete this table of values: x tanh(x) logsig(x) What do you see as x becomes very large? This phenomenon goes by the name of saturation. 5. Some people like to scale the sigmoidal function by an extra parameter, β, that is σ(βx). Show by sketching what happens to the graph of the sigmoidal (either the tansig or logsig) as you change β. It is not necessary to scale the sigmoidal, as this is equivalent to scaling the data instead (via the weights). 6. Show, using the definition of the hyperbolic sine and cosine, that the hyperbolic tangent can be written as: tanh(x) = sinh(x) cosh(x) 7. Show that the hyperbolic tangent can be computed as: tanh(x) = e 2x 1 (Matlab claims that this version is faster, but warns about possible numerical error) 5

6 8. Extensions of the transfer function Some other interesting transfer functions can be used at the nodes. Here are a couple of unique ones- They are used to encode circular or spherical information: (a) The Circular Node (two inputs, two outputs per node): ( ) x σ(x, y) = x2 + y, y 2 x2 + y 2 (b) The Spherical Node (three inputs, three outputs): ( ) x σ(x, y, z) = x2 + y 2 + z, y 2 x2 + y 2 + z, z 2 x2 + y 2 + z 2 See [?,?] for examples of how to implement the last two transfer function types. Training a Neural Network As we said previously, training a neural network means to find weights and biases that minimize the error function. There are several techniques available to us for doing this- among them: Method of Steepest Descent (or Gradient Descent) Newton s Method (an indirect method, solving for where the derivative of the error is 0). Conjuage Gradient (Search along the eigenvectors of the Hessian of the error) Levenburg-Marquardt (A combination of the techniques above). For us, we can practice using Gradient Descent, and for the other techniques we ll rely on Matlab. Going into the next section, recall that in the exercises, we showed that, if β = 1, then: σ (x) = σ(x) (1 σ(x)) 6

7 Backpropagation of Error We start with a simple example: A network: x w 1 x + b 1 = P σ(w 1 x + b 1 ) = S w 2 σ(w 1 x + b 1 ) + b 2 where P represents the prestate, and S represents the state of the node. Given target t, the error is E(w 1, w 2, b 1, b 2 ) = 1 2 (t y)2 = 1 2 (t (w 2σ(w 1 x + b 1 ) + b 2 )) 2 We want to minimize the error, so we move in the opposite direction of the gradient. Suppose we let the symbol u denote a generic parameter (either a weight or a bias). Then given a particular value of the parameter, it is updated to (hopefully) get a better error. Using gradient descent, u is updated by: u new = u old α E u = u old + α u where α is called the learning rate, and the change in u is computed via the chain rule on the error. Notice that we incorporated the negative sign into u- It will become clear why we did that (its because of the (t y) term- the derivative will always be negative t y). In particular, u = E y y u = (t y) y u Now let us compute these partial derivatives: = (t y) y u y w 2 y w 1 w 1 = S w 2 = (t y)s = w 2 σ (P ) x = (t y)w 2 σ (P ) x y b 2 = 1 b 2 = (t y) y b 1 = w 2 σ (P ) b 1 = (t y)w 2 σ (P ) Matlab and the Feedforward Network Here is an example training session: P=-1:0.1:1; T=sin(pi*P)+0.1*randn(size(P)); net = newff(p,t,10,{ tansig, logsig }); Y = sim(net,p); plot(p,t,p,y, o ); title( Before Training ); net = train(net,p,t); 7

8 X = linspace(-1,1,200); Y = sim(net,x); figure(2) plot(p,t, ko,x,y) title( After Training ); Bad things that might happen A particularly bad random set of initial weights and biases might be assigned. You should always try training several times to make sure that your error is as small as it should be- It is easy to get locked into a local minimum! 2. Saturation. This is when the data is badly scaled. The problem has to do with our sigmoidal function. An example might be in order: Consider the table of values x σ(x) We see that the transfer function begins to output 1 for ANY large number, so we say that it has lost its ability to distinguish between input patters (the function has become saturated). The same behavior happens for very negative input values as well. If your network begins to output the same numbers for wildly different inputs, then this is probably the reason (the weights could be large- See below). 3. Your data may be badly scaled. For example, suppose you have 4 dimensional input, and it represents temperatures from 200 degrees to 300 degrees in the first dimension, error values from to in the second dimension, altitudes like 2000 to 5000 feet in the third dimension, and integers from 1 to 10 in the fourth. Here are some samples: (250, , 2450, 6) The second column will disappear in terms of the network- the error minimization will end up focusing almost entirely on the third column. If possible, try to keep all of the scalings similar. For example, scale so that each dimension has mean zero and unit standard deviation (mean subtract and scale by the inverse of the standard deviation). 4. Too many nodes in the hidden layer: Use the test set/validation set to be sure you re not memorizing the data (the default settings work pretty well). 8

9 5. You should take a quick look at the magnitude of the numbers in the weights and biases- They should all be reasonable - Numbers that are out of scale with respect to the others should be suspect (and can lead to saturation). Recommended: hintonw(net.iw{1,1}) and hintonw(net.lw{2,1}). The colors are negative (versus positive), and the sizes of the squares are related to the magnitude of the numbers. 9

Neural Networks. Neural network is a network or circuit of neurons. Neurons can be. Biological neurons Artificial neurons

Neural Networks Neural network is a network or circuit of neurons Neurons can be Biological neurons Artificial neurons Biological neurons Building block of the brain Human brain contains over 10 billion

Feed-Forward mapping networks KAIST 바이오및뇌공학과 정재승

Feed-Forward mapping networks KAIST 바이오및뇌공학과 정재승 How much energy do we need for brain functions? Information processing: Trade-off between energy consumption and wiring cost Trade-off between energy consumption

Learning. Artificial Intelligence. Learning. Types of Learning. Inductive Learning Method. Inductive Learning. Learning.

Learning Learning is essential for unknown environments, i.e., when designer lacks omniscience Artificial Intelligence Learning Chapter 8 Learning is useful as a system construction method, i.e., expose

Artificial Neural Networks and Support Vector Machines. CS 486/686: Introduction to Artificial Intelligence

Artificial Neural Networks and Support Vector Machines CS 486/686: Introduction to Artificial Intelligence 1 Outline What is a Neural Network? - Perceptron learners - Multi-layer networks What is a Support

Introduction to Machine Learning and Data Mining. Prof. Dr. Igor Trajkovski trajkovski@nyus.edu.mk

Introduction to Machine Learning and Data Mining Prof. Dr. Igor Trakovski trakovski@nyus.edu.mk Neural Networks 2 Neural Networks Analogy to biological neural systems, the most robust learning systems

Building MLP networks by construction

University of Wollongong Research Online Faculty of Informatics - Papers (Archive) Faculty of Engineering and Information Sciences 2000 Building MLP networks by construction Ah Chung Tsoi University of

Application of Neural Network in User Authentication for Smart Home System

Application of Neural Network in User Authentication for Smart Home System A. Joseph, D.B.L. Bong, D.A.A. Mat Abstract Security has been an important issue and concern in the smart home systems. Smart

Neural networks. Chapter 20, Section 5 1

Neural networks Chapter 20, Section 5 Chapter 20, Section 5 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural networks Chapter 20, Section 5 2 Brains 0 neurons of

Neural Networks and Support Vector Machines

INF5390 - Kunstig intelligens Neural Networks and Support Vector Machines Roar Fjellheim INF5390-13 Neural Networks and SVM 1 Outline Neural networks Perceptrons Neural networks Support vector machines

CSC 321 H1S Study Guide (Last update: April 3, 2016) Winter 2016

1. Suppose our training set and test set are the same. Why would this be a problem? 2. Why is it necessary to have both a test set and a validation set? 3. Images are generally represented as n m 3 arrays,

PMR5406 Redes Neurais e Lógica Fuzzy Aula 3 Multilayer Percetrons

PMR5406 Redes Neurais e Aula 3 Multilayer Percetrons Baseado em: Neural Networks, Simon Haykin, Prentice-Hall, 2 nd edition Slides do curso por Elena Marchiori, Vrie Unviersity Multilayer Perceptrons Architecture

Neural network software tool development: exploring programming language options

INEB- PSI Technical Report 2006-1 Neural network software tool development: exploring programming language options Alexandra Oliveira aao@fe.up.pt Supervisor: Professor Joaquim Marques de Sá June 2006

Chapter 7. Diagnosis and Prognosis of Breast Cancer using Histopathological Data

Chapter 7 Diagnosis and Prognosis of Breast Cancer using Histopathological Data In the previous chapter, a method for classification of mammograms using wavelet analysis and adaptive neuro-fuzzy inference

Chapter 4: Artificial Neural Networks

Chapter 4: Artificial Neural Networks CS 536: Machine Learning Littman (Wu, TA) Administration icml-03: instructional Conference on Machine Learning http://www.cs.rutgers.edu/~mlittman/courses/ml03/icml03/

3 An Illustrative Example

Objectives An Illustrative Example Objectives - Theory and Examples -2 Problem Statement -2 Perceptron - Two-Input Case -4 Pattern Recognition Example -5 Hamming Network -8 Feedforward Layer -8 Recurrent

INTRODUCTION TO NEURAL NETWORKS

INTRODUCTION TO NEURAL NETWORKS Pictures are taken from http://www.cs.cmu.edu/~tom/mlbook-chapter-slides.html http://research.microsoft.com/~cmbishop/prml/index.htm By Nobel Khandaker Neural Networks An

Artificial neural networks

Artificial neural networks Now Neurons Neuron models Perceptron learning Multi-layer perceptrons Backpropagation 2 It all starts with a neuron 3 Some facts about human brain ~ 86 billion neurons ~ 10 15

Lecture 8 February 4

ICS273A: Machine Learning Winter 2008 Lecture 8 February 4 Scribe: Carlos Agell (Student) Lecturer: Deva Ramanan 8.1 Neural Nets 8.1.1 Logistic Regression Recall the logistic function: g(x) = 1 1 + e θt

An Introduction to Neural Networks

An Introduction to Vincent Cheung Kevin Cannons Signal & Data Compression Laboratory Electrical & Computer Engineering University of Manitoba Winnipeg, Manitoba, Canada Advisor: Dr. W. Kinsner May 27,

6.2.8 Neural networks for data mining

6.2.8 Neural networks for data mining Walter Kosters 1 In many application areas neural networks are known to be valuable tools. This also holds for data mining. In this chapter we discuss the use of neural

degrees of freedom and are able to adapt to the task they are supposed to do [Gupta].

1.3 Neural Networks 19 Neural Networks are large structured systems of equations. These systems have many degrees of freedom and are able to adapt to the task they are supposed to do [Gupta]. Two very

Machine Learning and Data Mining -

Machine Learning and Data Mining - Perceptron Neural Networks Nuno Cavalheiro Marques (nmm@di.fct.unl.pt) Spring Semester 2010/2011 MSc in Computer Science Multi Layer Perceptron Neurons and the Perceptron

Neural Network Toolbox

Neural Network Toolbox A Tutorial for the Course Computational Intelligence http://www.igi.tugraz.at/lehre/ci Stefan Häusler Institute for Theoretical Computer Science Inffeldgasse 16b/I Abstract This

Introduction to Neural Networks : Revision Lectures

Introduction to Neural Networks : Revision Lectures John A. Bullinaria, 2004 1. Module Aims and Learning Outcomes 2. Biological and Artificial Neural Networks 3. Training Methods for Multi Layer Perceptrons

1 SELF-ORGANIZATION MECHANISM IN THE NETWORKS

Mathematical literature reveals that the number of neural network structures, concepts, methods, and their applications have been well known in neural modeling literature for sometime. It started with

Analecta Vol. 8, No. 2 ISSN 2064-7964

EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,

6. Feed-forward mapping networks

6. Feed-forward mapping networks Fundamentals of Computational Neuroscience, T. P. Trappenberg, 2002. Lecture Notes on Brain and Computation Byoung-Tak Zhang Biointelligence Laboratory School of Computer

Lecture 6. Artificial Neural Networks

Lecture 6 Artificial Neural Networks 1 1 Artificial Neural Networks In this note we provide an overview of the key concepts that have led to the emergence of Artificial Neural Networks as a major paradigm

Machine Learning: Multi Layer Perceptrons

Machine Learning: Multi Layer Perceptrons Prof. Dr. Martin Riedmiller Albert-Ludwigs-University Freiburg AG Maschinelles Lernen Machine Learning: Multi Layer Perceptrons p.1/61 Outline multi layer perceptrons

A TUTORIAL. BY: Negin Yousefpour PhD Student Civil Engineering Department TEXAS A&M UNIVERSITY

ARTIFICIAL NEURAL NETWORKS: A TUTORIAL BY: Negin Yousefpour PhD Student Civil Engineering Department TEXAS A&M UNIVERSITY Contents Introduction Origin Of Neural Network Biological Neural Networks ANN Overview

Lecture 8 Artificial neural networks: Unsupervised learning

Lecture 8 Artificial neural networks: Unsupervised learning Introduction Hebbian learning Generalised Hebbian learning algorithm Competitive learning Self-organising computational map: Kohonen network

Solving for Voltage and Current

Chapter 3 Solving for Voltage and Current Nodal Analysis If you know Ohm s Law, you can solve for all the voltages and currents in simple resistor circuits, like the one shown below. In this chapter, we

Lecture 1: Introduction to Neural Networks Kevin Swingler / Bruce Graham

Lecture 1: Introduction to Neural Networks Kevin Swingler / Bruce Graham kms@cs.stir.ac.uk 1 What are Neural Networks? Neural Networks are networks of neurons, for example, as found in real (i.e. biological)

Neural Networks. Introduction to Artificial Intelligence CSE 150 May 29, 2007

Neural Networks Introduction to Artificial Intelligence CSE 150 May 29, 2007 Administration Last programming assignment has been posted! Final Exam: Tuesday, June 12, 11:30-2:30 Last Lecture Naïve Bayes

TIME SERIES FORECASTING WITH NEURAL NETWORK: A CASE STUDY OF STOCK PRICES OF INTERCONTINENTAL BANK NIGERIA

www.arpapress.com/volumes/vol9issue3/ijrras_9_3_16.pdf TIME SERIES FORECASTING WITH NEURAL NETWORK: A CASE STUDY OF STOCK PRICES OF INTERCONTINENTAL BANK NIGERIA 1 Akintola K.G., 2 Alese B.K. & 2 Thompson

Recurrent Neural Networks

Recurrent Neural Networks Neural Computation : Lecture 12 John A. Bullinaria, 2015 1. Recurrent Neural Network Architectures 2. State Space Models and Dynamical Systems 3. Backpropagation Through Time

Basics Inversion and related concepts Random vectors Matrix calculus. Matrix algebra. Patrick Breheny. January 20

Matrix algebra January 20 Introduction Basics The mathematics of multiple regression revolves around ordering and keeping track of large arrays of numbers and solving systems of equations The mathematical

Using Neural Networks for Pattern Classification Problems

Using Neural Networks for Pattern Classification Problems Converting an Image Camera captures an image Image needs to be converted to a form that can be processed by the Neural Network Converting an Image

Introduction to Neural Networks for Senior Design

Introduction to Neural Networks for Senior Design Intro-1 Neural Networks: The Big Picture Artificial Intelligence Neural Networks Expert Systems Machine Learning not ruleoriented ruleoriented Intro-2

COPYRIGHT NOTICE: David A. Kendrick, P. Ruben Mercado, and Hans M. Amman: Computational Economics is published by Princeton University Press and copyrighted, 2006, by Princeton University Press. All rights

The Phase Plane. Phase portraits; type and stability classifications of equilibrium solutions of systems of differential equations

The Phase Plane Phase portraits; type and stability classifications of equilibrium solutions of systems of differential equations Phase Portraits of Linear Systems Consider a systems of linear differential

One Solution to XOR problem using Multilayer Perceptron having Minimum Configuration

International Journal of Science and Engineering Volume 3, Number 2-2015 PP: 32-41 IJSE Available at www.ijse.org ISSN: 2347-2200 One Solution to XOR problem using Multilayer Perceptron having Minimum

PSS 27.2 The Electric Field of a Continuous Distribution of Charge

Chapter 27 Solutions PSS 27.2 The Electric Field of a Continuous Distribution of Charge Description: Knight Problem-Solving Strategy 27.2 The Electric Field of a Continuous Distribution of Charge is illustrated.

Gradient Methods Rafael E. Banchs INTRODUCTION This report discuss one class of the local search algorithms to be used in the inverse modeling of the time harmonic field electric logging problem, the Gradient

Role of Neural network in data mining

Role of Neural network in data mining Chitranjanjit kaur Associate Prof Guru Nanak College, Sukhchainana Phagwara,(GNDU) Punjab, India Pooja kapoor Associate Prof Swami Sarvanand Group Of Institutes Dinanagar(PTU)

Data Mining Techniques Chapter 7: Artificial Neural Networks

Data Mining Techniques Chapter 7: Artificial Neural Networks Artificial Neural Networks.................................................. 2 Neural network example...................................................

Regression Analysis: Basic Concepts

The simple linear model Regression Analysis: Basic Concepts Allin Cottrell Represents the dependent variable, y i, as a linear function of one independent variable, x i, subject to a random disturbance

Chapter 12 Discovering New Knowledge Data Mining

Chapter 12 Discovering New Knowledge Data Mining Becerra-Fernandez, et al. -- Knowledge Management 1/e -- 2004 Prentice Hall Additional material 2007 Dekai Wu Chapter Objectives Introduce the student to

Predict Influencers in the Social Network

Predict Influencers in the Social Network Ruishan Liu, Yang Zhao and Liuyu Zhou Email: rliu2, yzhao2, lyzhou@stanford.edu Department of Electrical Engineering, Stanford University Abstract Given two persons

AN APPLICATION OF TIME SERIES ANALYSIS FOR WEATHER FORECASTING

AN APPLICATION OF TIME SERIES ANALYSIS FOR WEATHER FORECASTING Abhishek Agrawal*, Vikas Kumar** 1,Ashish Pandey** 2,Imran Khan** 3 *(M. Tech Scholar, Department of Computer Science, Bhagwant University,

: In this Module we will consider instruments with an analog readout. You will be using a ruler to measure the diameter of a coin. Imagine we are measuring some position with a ruler. A portion of the

SUCCESSFUL PREDICTION OF HORSE RACING RESULTS USING A NEURAL NETWORK

SUCCESSFUL PREDICTION OF HORSE RACING RESULTS USING A NEURAL NETWORK N M Allinson and D Merritt 1 Introduction This contribution has two main sections. The first discusses some aspects of multilayer perceptrons,

TRAINING A LIMITED-INTERCONNECT, SYNTHETIC NEURAL IC

777 TRAINING A LIMITED-INTERCONNECT, SYNTHETIC NEURAL IC M.R. Walker. S. Haghighi. A. Afghan. and L.A. Akers Center for Solid State Electronics Research Arizona State University Tempe. AZ 85287-6206 mwalker@enuxha.eas.asu.edu

Math 4310 Handout - Quotient Vector Spaces

Math 4310 Handout - Quotient Vector Spaces Dan Collins The textbook defines a subspace of a vector space in Chapter 4, but it avoids ever discussing the notion of a quotient space. This is understandable

DERIVATIVES AS MATRICES; CHAIN RULE

DERIVATIVES AS MATRICES; CHAIN RULE 1. Derivatives of Real-valued Functions Let s first consider functions f : R 2 R. Recall that if the partial derivatives of f exist at the point (x 0, y 0 ), then we

Introduction to Artificial Neural Networks MAE-491/591

Introduction to Artificial Neural Networks MAE-491/591 Artificial Neural Networks: Biological Inspiration The brain has been extensively studied by scientists. Vast complexity prevents all but rudimentary

NEURAL NETWORKS A Comprehensive Foundation

NEURAL NETWORKS A Comprehensive Foundation Second Edition Simon Haykin McMaster University Hamilton, Ontario, Canada Prentice Hall Prentice Hall Upper Saddle River; New Jersey 07458 Preface xii Acknowledgments

Operational Amplifiers

Operational Amplifiers Introduction The operational amplifier (op-amp) is a voltage controlled voltage source with very high gain. It is a five terminal four port active element. The symbol of the op-amp

Comparison Between Multilayer Feedforward Neural Networks and a Radial Basis Function Network to Detect and Locate Leaks in Pipelines Transporting Gas

A publication of 1375 CHEMICAL ENGINEERINGTRANSACTIONS VOL. 32, 2013 Chief Editors:SauroPierucci, JiříJ. Klemeš Copyright 2013, AIDIC ServiziS.r.l., ISBN 978-88-95608-23-5; ISSN 1974-9791 The Italian Association

Neural Networks. CAP5610 Machine Learning Instructor: Guo-Jun Qi

Neural Networks CAP5610 Machine Learning Instructor: Guo-Jun Qi Recap: linear classifier Logistic regression Maximizing the posterior distribution of class Y conditional on the input vector X Support vector

THREE DIMENSIONAL REPRESENTATION OF AMINO ACID CHARAC- TERISTICS

THREE DIMENSIONAL REPRESENTATION OF AMINO ACID CHARAC- TERISTICS O.U. Sezerman 1, R. Islamaj 2, E. Alpaydin 2 1 Laborotory of Computational Biology, Sabancı University, Istanbul, Turkey. 2 Computer Engineering

BASIC ELECTRONICS PROF. T.S. NATARAJAN DEPT OF PHYSICS IIT MADRAS LECTURE-4 SOME USEFUL LAWS IN BASIC ELECTRONICS

BASIC ELECTRONICS PROF. T.S. NATARAJAN DEPT OF PHYSICS IIT MADRAS LECTURE-4 SOME USEFUL LAWS IN BASIC ELECTRONICS Hello everybody! In a series of lecture on basic electronics, learning by doing, we now

Horse Racing Prediction Using Artificial Neural Networks

Horse Racing Prediction Using Artificial Neural Networks ELNAZ DAVOODI, ALI REZA KHANTEYMOORI Mathematics and Computer science Department Institute for Advanced Studies in Basic Sciences (IASBS) Gavazang,

[1] Diagonal factorization

8.03 LA.6: Diagonalization and Orthogonal Matrices [ Diagonal factorization [2 Solving systems of first order differential equations [3 Symmetric and Orthonormal Matrices [ Diagonal factorization Recall:

ALGEBRA. sequence, term, nth term, consecutive, rule, relationship, generate, predict, continue increase, decrease finite, infinite

ALGEBRA Pupils should be taught to: Generate and describe sequences As outcomes, Year 7 pupils should, for example: Use, read and write, spelling correctly: sequence, term, nth term, consecutive, rule,

SELECTING NEURAL NETWORK ARCHITECTURE FOR INVESTMENT PROFITABILITY PREDICTIONS

UDC: 004.8 Original scientific paper SELECTING NEURAL NETWORK ARCHITECTURE FOR INVESTMENT PROFITABILITY PREDICTIONS Tonimir Kišasondi, Alen Lovren i University of Zagreb, Faculty of Organization and Informatics,

Advanced Control System Design Prof. Radhakant Padhi Department of Aerospace Engineering Indian Institute of Science, Bangalore Lecture No. # 09 Representation of Dynamical Systems - I Previously, we have

Self Organizing Maps: Fundamentals

Self Organizing Maps: Fundamentals Introduction to Neural Networks : Lecture 16 John A. Bullinaria, 2004 1. What is a Self Organizing Map? 2. Topographic Maps 3. Setting up a Self Organizing Map 4. Kohonen

A Time Series ANN Approach for Weather Forecasting

A Time Series ANN Approach for Weather Forecasting Neeraj Kumar 1, Govind Kumar Jha 2 1 Associate Professor and Head Deptt. Of Computer Science,Nalanda College Of Engineering Chandi(Bihar) 2 Assistant

Review of some concepts in predictive modeling

Review of some concepts in predictive modeling Brigham and Women s Hospital Harvard-MIT Division of Health Sciences and Technology HST.951J: Medical Decision Support A disjoint list of topics? Naïve Bayes

Notes on Support Vector Machines

Notes on Support Vector Machines Fernando Mira da Silva Fernando.Silva@inesc.pt Neural Network Group I N E S C November 1998 Abstract This report describes an empirical study of Support Vector Machines

Simple Linear Regression Chapter 11

Simple Linear Regression Chapter 11 Rationale Frequently decision-making situations require modeling of relationships among business variables. For instance, the amount of sale of a product may be related

Introduction to Artificial Neural Networks. Introduction to Artificial Neural Networks

Introduction to Artificial Neural Networks v.3 August Michel Verleysen Introduction - Introduction to Artificial Neural Networks p Why ANNs? p Biological inspiration p Some examples of problems p Historical

Roots and Coefficients of a Quadratic Equation Summary

Roots and Coefficients of a Quadratic Equation Summary For a quadratic equation with roots α and β: Sum of roots = α + β = and Product of roots = αβ = Symmetrical functions of α and β include: x = and

5 Homogeneous systems

5 Homogeneous systems Definition: A homogeneous (ho-mo-jeen -i-us) system of linear algebraic equations is one in which all the numbers on the right hand side are equal to : a x +... + a n x n =.. a m

Dynamic neural network with adaptive Gauss neuron activation function

Dynamic neural network with adaptive Gauss neuron activation function Dubravko Majetic, Danko Brezak, Branko Novakovic & Josip Kasac Abstract: An attempt has been made to establish a nonlinear dynamic

Quick Reference Guide to Linear Algebra in Quantum Mechanics

Quick Reference Guide to Linear Algebra in Quantum Mechanics Scott N. Walck September 2, 2014 Contents 1 Complex Numbers 2 1.1 Introduction............................ 2 1.2 Real Numbers...........................

IBM SPSS Neural Networks 22

IBM SPSS Neural Networks 22 Note Before using this information and the product it supports, read the information in Notices on page 21. Product Information This edition applies to version 22, release 0,

129: Artificial Neural Networks. Ajith Abraham Oklahoma State University, Stillwater, OK, USA 1 INTRODUCTION TO ARTIFICIAL NEURAL NETWORKS

129: Artificial Neural Networks Ajith Abraham Oklahoma State University, Stillwater, OK, USA 1 Introduction to Artificial Neural Networks 901 2 Neural Network Architectures 902 3 Neural Network Learning

4.1 Learning algorithms for neural networks

4 Perceptron Learning 4.1 Learning algorithms for neural networks In the two preceding chapters we discussed two closely related models, McCulloch Pitts units and perceptrons, but the question of how to

Power Prediction Analysis using Artificial Neural Network in MS Excel

Power Prediction Analysis using Artificial Neural Network in MS Excel NURHASHINMAH MAHAMAD, MUHAMAD KAMAL B. MOHAMMED AMIN Electronic System Engineering Department Malaysia Japan International Institute

Position, Velocity, and Acceleration

Chapter 2 Position, Velocity, and Acceleration In this section, we will define the first set of quantities (position, velocity, and acceleration) that will form a foundation for most of what we do in this

varies, if the classication is performed with a certain accuracy, and if a complex input-output association. There exist theoretical guarantees

Training a Sigmoidal Network is Dicult Barbara Hammer, University of Osnabruck, Dept. of Mathematics/Comp. Science, Albrechtstrae 28, 49069 Osnabruck, Germany Abstract. In this paper we show that the loading

CSC321 Introduction to Neural Networks and Machine Learning. Lecture 21 Using Boltzmann machines to initialize backpropagation.

CSC321 Introduction to Neural Networks and Machine Learning Lecture 21 Using Boltzmann machines to initialize backpropagation Geoffrey Hinton Some problems with backpropagation The amount of information

15.062 Data Mining: Algorithms and Applications Matrix Math Review

.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop

Solving Nonlinear Equations Using Recurrent Neural Networks

Solving Nonlinear Equations Using Recurrent Neural Networks Karl Mathia and Richard Saeks, Ph.D. Accurate Automation Corporation 71 Shallowford Road Chattanooga, Tennessee 37421 Abstract A class of recurrent

Modular Neural Networks

16 Modular Neural Networks In the previous chapters we have discussed different models of neural networks linear, recurrent, supervised, unsupervised, self-organizing, etc. Each kind of network relies

v w is orthogonal to both v and w. the three vectors v, w and v w form a right-handed set of vectors.

3. Cross product Definition 3.1. Let v and w be two vectors in R 3. The cross product of v and w, denoted v w, is the vector defined as follows: the length of v w is the area of the parallelogram with

Math 56 Homework 4 Michael Downs

Fourier series theory. As in lecture, let be the L 2 -norm on (, 2π). (a) Find a unit-norm function orthogonal to f(x) = x on x < 2π. [Hint: do what you d do in vector-land, eg pick some simple new function

Learning Objectives for Math 165

Learning Objectives for Math 165 Chapter 2 Limits Section 2.1: Average Rate of Change. State the definition of average rate of change Describe what the rate of change does and does not tell us in a given

Artificial Neural Network for Speech Recognition

Artificial Neural Network for Speech Recognition Austin Marshall March 3, 2005 2nd Annual Student Research Showcase Overview Presenting an Artificial Neural Network to recognize and classify speech Spoken

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical

Threshold Logic. 2.1 Networks of functions

2 Threshold Logic 2. Networks of functions We deal in this chapter with the simplest kind of computing units used to build artificial neural networks. These computing elements are a generalization of the

SMORN-VII REPORT NEURAL NETWORK BENCHMARK ANALYSIS RESULTS & FOLLOW-UP 96. Özer CIFTCIOGLU Istanbul Technical University, ITU. and

NEA/NSC-DOC (96)29 AUGUST 1996 SMORN-VII REPORT NEURAL NETWORK BENCHMARK ANALYSIS RESULTS & FOLLOW-UP 96 Özer CIFTCIOGLU Istanbul Technical University, ITU and Erdinç TÜRKCAN Netherlands Energy Research

Introduction to Finite Systems: Z 6 and Z 7

Introduction to : Z 6 and Z 7 The main objective of this discussion is to learn more about solving linear and quadratic equations. The reader is no doubt familiar with techniques for solving these equations

Performance Evaluation of Artificial Neural. Networks for Spatial Data Analysis

Contemporary Engineering Sciences, Vol. 4, 2011, no. 4, 149-163 Performance Evaluation of Artificial Neural Networks for Spatial Data Analysis Akram A. Moustafa Department of Computer Science Al al-bayt

Able Enrichment Centre - Prep Level Curriculum

Able Enrichment Centre - Prep Level Curriculum Unit 1: Number Systems Number Line Converting expanded form into standard form or vice versa. Define: Prime Number, Natural Number, Integer, Rational Number,

Predicting Results of Brazilian Soccer League Matches

University of Wisconsin-Madison ECE/CS/ME 539 Introduction to Artificial Neural Networks and Fuzzy Systems Predicting Results of Brazilian Soccer League Matches Student: Alberto Trindade Tavares Email: