Accepted Notation. x i w ij. Net input to unit Y j : s j = b j + i
|
|
- Linette Richard
- 7 years ago
- Views:
Transcription
1 Accepted Notation x i input signal of activation unit X i. y j y j = F (s j ) for an activation unit Y j. w i,j Weight on connection from unit X i to unit Y j. b j Bias on unit Y j. s j Net input to unit Y j : s j = b j + i x i w ij W weight matrix {w ij }. w j Vector of weights w j = (w 1j, w 2j,..., w nj ) T. u j Threshold for activation of neuron Y j. x Input vector for classification or response x = (x 1,..., x i,..., x n ). w ij Change in w ij. w ij = [w ij (new) w ij (old)]. Dr. E.C. Kulasekere () Neural Networks 1 / 36
2 Mathematical Models of Activation Functions Typically the same activation function is used for all neurons in any particular layer (this is not a requirement). In a multi-layer network if the neurons have linear activation functions the capabilities are no better than a single layer network with a linear activation function. Hence in most cases nonlinear activation functions are used. Linear Activation Function f (x) = x, x. Dr. E.C. Kulasekere () Neural Networks 2 / 36
3 Mathematical Models of Activation Functions... Binary Step Function with Threshold θ { 1 if x θ f (x) = 0 if x < θ f(x) 1 θ x Dr. E.C. Kulasekere () Neural Networks 3 / 36
4 Mathematical Models of Activation Functions... Binary Sigmoid Function f (x) = exp ( σx). f (x) = σf (x)[1 f (x)]. f(x) 1 This is especially useful when training the backpropagation neural network where the derivatives are used to obtain the output value. Dr. E.C. Kulasekere () Neural Networks 4 / 36 x
5 Mathematical Models of Activation Functions... Bipolar Sigmoid Function g(x) = 2f (x) 1 = 1 exp ( σx) 1 + exp ( σx). g (x) = σ [1 + g(x)][1 g(x)]. 2 The above is actually the binary sigmoid scaled to have a set of values ranging from -1 to +1. The bipolar sigmoid is also closely related to the hyperbolic tangent function. exp (x) exp ( x) h(x) = tanh (x) = exp (x) + exp ( x) 1 exp ( 2x) = 1 + exp ( 2x). Dr. E.C. Kulasekere () Neural Networks 5 / 36
6 Matrix Manipulations 1 X 1 w 1j b j X i w ij Y j X n w nj s j = w j x t = n i=1 x iw ij, if the bias is neglected If the bias is not neglected, the x includes the bias value. Here row vectors for the input and weights are assumed. Ex is a direct application of the material learned thus far. Dr. E.C. Kulasekere () Neural Networks 6 / 36
7 McCulloch-Pitts Neuron Characteristics The McCulloch-Pitt neuron is considered to be the first neural network. It is a fixed weight network that can be used to implement boolean functions. Its characteristics are: Binary activation (1 ON, 0 OFF). i.e. it either fires with an activation 1 or does not fire with an activation of 0. Neurons are connected by directed weighted paths. If w > 0, excitatory, else inhibitory. Excitatory weights are identical; inhibitory weights too. Each neuron has a fixed threshold for firing. That is if the net input to the neuron is greater than the threshold it fires. The threshold is set such that the inhibition is absolute. Any non-zero inhibitory input will prevent it from firing. It takes one time step for the signal to pass over one link. Dr. E.C. Kulasekere () Neural Networks 7 / 36
8 McCulloch-Pitts Neuron Architecture X 1 w The inhibition should be absolute. Hence the threshold satisfies θ > nw p X n X n+1 X n+m w p p Y The output neuron will fire if it receives k or more excitatory inputs with no inhibitory inputs, where kw θ > (k 1)w The activation function { is 1 if s θ f (s) = 0 if s < θ Dr. E.C. Kulasekere () Neural Networks 8 / 36
9 Neural net to perform the AND function X 1 X Y Write the truth table for the above network. Threshold = 2 Given another configuration of the AND function with Threshold 1. Dr. E.C. Kulasekere () Neural Networks 9 / 36
10 Neural net to perform the OR function X 1 X Y Write the truth table for the above network. Threshold=2 Give another configuration of the OR function with Threshold 1. Dr. E.C. Kulasekere () Neural Networks 10 / 36
11 Neural net to perform the AND NOT function X 1 X Y Write the truth table for the AND NOT function. Threshold=2 Can you find another configuration for the weights to implement the AND NOT function? Dr. E.C. Kulasekere () Neural Networks 11 / 36
12 Neural net to perform the XOR function X Z 1 2 Y X Z 2 2 Write the truth table for the XOR function. Threshold=2 Write the layer equations for the XOR function and show that it is AND NOT combined with an OR function. Dr. E.C. Kulasekere () Neural Networks 12 / 36
13 Geometric View of Neural Architectures Single output neuron Note that the threshold function can be reduced to the discrete activation function when the threshold is considered as a bias. Then the synaptic input can be written as s = wx t. For two input neurons with inputs x 1 and x 2, s = 0 is a straight line. On the x 1 x 2 -plane, s > 0 gives one half-plane and s < 0 gives another half-plane. If the weights and threshold are multiplied by a constant c, then the only difference is that the half-planes in which the earlier classification occurred will be switched. (pp.27) If you have more that two inputs you will be plotting on a x 1 x 2... x n -hyper plane. For example if n = 3 you will end up with a plane. Dr. E.C. Kulasekere () Neural Networks 13 / 36
14 Geometric View of Neural Architectures Multiple output neurons Each output neuron will have a summation block of s j = wx t. For j = 2 two input neurons the above will give two straight lines in the x 1, x 2 -plane. For j = 2 with more than two input neurons, the above will give two n-dimensional hyper-planes. For j > 2 and several inputs will give higher order planes. Dr. E.C. Kulasekere () Neural Networks 14 / 36
15 Example Note that the threshold function can be converted to the discrete activation function by converting s > 0 to s u > 0. By doing this we are converting a threshold to a bias. We have to mark the area in which the output turns out to be 1. That is by looking at the vertices, we determine it is x 1 > 0, x 1 < 1, x 2 > 0, and, x 2 < 1. In order to obtain the required output the threshold function has to satisfy s u > 0. That is w 1 x 1 + w 2 x 2 u > 0. Now compare the above two items to obtain the required w 1, w 2 and u Even for Exercise we have a region bonded by a triangle in which the same computation can be carried out. Even when you are building a two layer network the arguments used in MCP networks can be used for example the question in pp.33. All of these techniques produce fixed weight ANNs. Dr. E.C. Kulasekere () Neural Networks 15 / 36
16 Pattern Classification Using ANNs This is the simplest task a neural network can be trained to perform. Each input vector pattern either belongs or does not belong to a class. The correct input data is assumed to be known. The activation function is either binary or bipolar. The output of the trained network will be 1 when presented with a pattern from the class. If the pattern is not in the class a -1 (or 0 for a binary output) is output. Since it is either one set of outputs the activation is a hard step. Dr. E.C. Kulasekere () Neural Networks 16 / 36
17 Biases and Thresholds The bias acts exactly as a weight on a connection from a unit whose activation is always 1. Increasing the bias increases the net input to to the unit. The Activation with a bias is f (s) = { 1 if s 0; 1 if s < 0; The Activation with a threshold is where s = b + i x i w i,. f (s) = { 1 if s θ; 1 if s < θ; where s = i x i w i,. Essentially the above are equivalent for most cases. However the bias becomes essential for certain problems associated with linear separability. Dr. E.C. Kulasekere () Neural Networks 17 / 36
18 Requirement of Linear Separability In pattern classification, the network is trained to output an indicator to show the presented input is either a member of one class or another class. This will depend on the decision boundary that will either fire the neuron or not fire it. This can be determined by setting y_in = 0 for the case with a bias. The decision boundary is b + i x i w i = 0. If the system is linearly separable, the two classes should lie on either side of the decision boundary (y_in > 0 or y_in < 0). The boundary is non unique x 2 = w 1 w 2 x 1 b w 2. Dr. E.C. Kulasekere () Neural Networks 18 / 36
19 Requirement of Linear Separability... For bipolar signals the outputs for the two classes are -1 and +1. For unipolar signals it is 0 and 1. Depending on the number of inputs the decision boundary can be a line, plane or a hyperplane. Eg. For two inputs its a line and for three inputs its a plane. If all of the training input vectors for which the correct response is +1 lie on one side of the decision boundary we say that this system is linearly separable. It has been shown that a single layer network can only learn linearly separable problems. The trained weights are not unique. Dr. E.C. Kulasekere () Neural Networks 19 / 36
20 Response Regions for the AND Function The AND function for Bipolar inputs and targets is Input (x 1, x 2 ) Output (t) (1,1) +1 (1,-1) -1 (-1,1) -1 (-1,-1) -1 x 2 = x is the decision boundary where w 1 = w 2 = 1 and b = 1. This boundary is not unique. - - x x 1 Dr. E.C. Kulasekere () Neural Networks 20 / 36
21 Response Regions for the OR function The OR function for Bipolar inputs and targets is Input (x 1, x 2 ) Output (t) (1,1) +1 (1,-1) +1 One (-1,1) +1 (-1,-1) -1 possible decision boundary is given by x 2 = x 1 1 where b = 1, w 1 = 1 = w 2. Draw the above graphically. Importance of Bias If the bias weight was not included in the previous example the decision boundary would have been forced to go through the origin. This changes the problem from a solvable one to an unsolvable problem. Dr. E.C. Kulasekere () Neural Networks 21 / 36
22 Response Regions for the XOR Function The XOR function for bipolar inputs and targets is given by Input (x 1, x 2 ) Output (t) (1,1) -1 (1,-1) +1 This (-1,1) +1 (-1,-1) -1 is not linearly separable. + x 2 - x Dr. E.C. Kulasekere () Neural Networks 22 / 36
23 Binary/Bipolar Data Representation In most cases the binary input data can be modified to bipolar data. However the form of the data can change the problem from one that is solvable to a problem that cannot be solved. Binary representation is also not as good as the bipolar if we want the net to generalize. i.e. to respond to input data that is similar but not the same as the training data. Bipolar inputs and targets are also better during training where the adaptive weight change is computed. For example w = xy is the weight change in the Hebb learning algorithm. If either the training input vector or the target vector is binary (unipolar) the update becomes zero and the learning stops. Using bipolar missing data can be distinguished from the mistaken data by assigning 0 to missing data and a mistake by changing -1 to +1 or vise versa. Dr. E.C. Kulasekere () Neural Networks 23 / 36
24 The HEBB Net This is the earliest and the most simplest learning rule for the neural networks. Hebb proposed that if two interconnected neurons are both on at the same time, then the weight between them should be increased. However the original statement did not discuss neurons that are connected but do not fire together. Later this was also included in the Hebb rule so that the weight increased of not firing together and the original rule was made a strongest learning algorithm. The weight update can be represented as w i (new) = w i (old) + x i y. Note that if the signals are binary, the update rule cannot distinguish between a pair of inputs outputs for the following conditions. Input is on and the target is off. Both the input and the target units are off. Dr. E.C. Kulasekere () Neural Networks 24 / 36
25 Hebb Algorithm Step 0 Initialize all weights: w i = 0, i = 1,..., n. Step 1 For each input training vector and target output pair, s : t, do steps 2-4. Step 2 Set activations for input units: x i = s i, i = 1,..., n. Step 3 Set activation for output unit: y = t. Step 4 Adjust the weight and bias: w i (new) = w i (old) + x i y for i = 1,..., n b(new) = b(old) + y. If the bias is considered to be an input signal that is always 1, the weight change can be written as where w = xy. w(new) = w(old) + w Dr. E.C. Kulasekere () Neural Networks 25 / 36
26 Hebb Net for AND Function Binary inputs and targets Input Target (x 1, x 2, 1) The truth table for the AND function is (1,1,1) 1 (1,0,1) 0 (0,1,1) 0 (0,0,1) 0 The initial values are w(old) = (0, 0) and b(old) = 0. The first step of the algorithm is Input Target Weight Changes Weights (x 1, x 2, 1) ( w 1, w 2, b) (w 1, w 2, b) (0,0,0) (1,1,1) 1 (1,1,1) (1,1,1) The separating line after the first step is x 2 = x 1 1 Dr. E.C. Kulasekere () Neural Networks 26 / 36
27 Hebb Net for AND Function... Binary inputs and targets Now if we present the second, third and fourth training vectors the weight change is given by Input Target Weight Changes Weights (x 1, x 2, 1) ( w 1, w 2, b) (w 1, w 2, b) (1,1,1) (1,0,1) 0 (0,0,0) (1,1,1) (0,1,1) 0 (0,0,0) (1,1,1) (0,0,1) 0 (0,0,0) (1,1,1) We note that in the above problem when the target value is zero, no learning occurs and hence no weight change. With this we can determine that the bipolar inputs have resulted in a short coming in the learning method. Dr. E.C. Kulasekere () Neural Networks 27 / 36
28 Hebb Net for the AND function Binary inputs, bipolar targets The truth table in this case is Input Target (x 1, x 2, 1) (1,1,1) +1 (1,0,1) -1 (0,1,1) -1 (0,0,1) -1 Presenting the first input Input Target Weight Changes Weights (x 1, x 2, 1) ( w 1, w 2, b) (w 1, w 2, b) (0,0,0) (1,1,1) 1 (1,1,1) (1,1,1) The separating line becomes x 2 = x 1 1. This is the correct classification for the first input. Dr. E.C. Kulasekere () Neural Networks 28 / 36
29 Hebb Net for the AND function... Binary inputs, bipolar targets x x 1 Dr. E.C. Kulasekere () Neural Networks 29 / 36
30 Hebb Net for the AND function... Binary inputs, bipolar targets Presenting the other inputs we have the following Input Target Weight Changes Weights (x 1, x 2, 1) ( w 1, w 2, b) (w 1, w 2, b) (1,0,1) -1 (-1,0,-1) (0,1,0) (0,1,1) -1 (0,-1,-1) (0,0,-1) (0,0,1) -1 (0,0,-1) (0,0,-2) Again we see that the problem has not correctly classified the output. Hence alternate training data should be used. Dr. E.C. Kulasekere () Neural Networks 30 / 36
31 Hebb Net for the AND function Bipolar inputs and targets The truth table is given by Input Target (x 1, x 2, 1) (1,1,1) +1 (1,-1,1) -1 (-1,1,1) -1 (-1,-1,1) -1 Presenting the first input:target pair yields the same result as before with the separation line being x 2 = x 1 1. Now the classification is correct for the point (1,1) and also for (-1,-1). - - x x 1 Dr. E.C. Kulasekere () Neural Networks 31 / 36
32 Hebb Net for the AND function... Bipolar inputs and targets For the second input:target pair we obtain the separating line as x 2 = 0. Input Target Weight Changes Weights (x 1, x 2, 1) ( w 1, w 2, b) (w 1, w 2, b) (1,1,1) (1,-1,1) -1 (-1,1,-1) (0,2,0) With the third pair we obtain the separating line as x 2 = x Input Target Weight Changes Weights (x 1, x 2, 1) ( w 1, w 2, b) (w 1, w 2, b) (0,2,0) (-1,1,1) -1 (1,-1,-1) (1,1,-1) Presenting the fourth pair, the separating line does not change from what we obtained previously even though the weights have changed. Input Target Weight Changes Weights (x 1, x 2, 1) ( w 1, w 2, b) (w 1, w 2, b) (1,1,-1) (-1,-1,1) -1 (1,1,-1) (2,2,-2) Dr. E.C. Kulasekere () Neural Networks 32 / 36
33 Hebb Net for the AND function... Bipolar inputs and targets x x Figure: Decision boundary for bipolar AND function using Hebb rule after third/fourth training. Draw the diagram for the intermediate stages as well. Dr. E.C. Kulasekere () Neural Networks 33 / 36
34 Character Recognition Example Using Hebb Net The Problem: Distinguish between the two patterns. # # # # # # # # # # # # # # # # # # # # # The steps involved are as follows: We use the bipolar representation to convert the patterns into the input values that can be used for training. We assign # to 1 and to -1. The correct target for the first pattern is +1 and for the second pattern it is -1. This is used as the classification. Then use the Hebb rule to find the weights by repeatedly presenting the input:target pairs ti find the weights. Check the system with training patterns that are similar but not identical to see if the system will still react with the correct classification. Dr. E.C. Kulasekere () Neural Networks 34 / 36
35 Missing and Mistaken Data Binary representation of inputs and targets (0 and 1 levels) Missing Data: Cannot be represented. Mistaken Data: A mistake of 1 is represented as 0 and vise versa. Bipolar representation of inputs and targets (-1 and 1 levels) Missing Data: this is represented by a 0. Mistaken Data: A mistake in +1 is represented as -1 and vise versa. In general a net can handle more missing components than wrong components. For input data the above translates to It s better not to guess. The reason is that if the value is zero the weight change during training may not get affected. Dr. E.C. Kulasekere () Neural Networks 35 / 36
36 Additional Comments ANN Training is completely ad hoc. Sometimes the result will not converge. The weight vector ca be multiplied by a positive number without changing the actions of the neurons. This normalization can eliminate rounding errors account for input variations etc. The value of η is significant for convergence as well as the rate of convergence. The order in which the training vectors are presented is also important. Randomizing is generally preferable. In multilayer networks, it is accepted that more than two layers are rarely used. Dr. E.C. Kulasekere () Neural Networks 36 / 36
Feed-Forward mapping networks KAIST 바이오및뇌공학과 정재승
Feed-Forward mapping networks KAIST 바이오및뇌공학과 정재승 How much energy do we need for brain functions? Information processing: Trade-off between energy consumption and wiring cost Trade-off between energy consumption
More information3 An Illustrative Example
Objectives An Illustrative Example Objectives - Theory and Examples -2 Problem Statement -2 Perceptron - Two-Input Case -4 Pattern Recognition Example -5 Hamming Network -8 Feedforward Layer -8 Recurrent
More informationdegrees of freedom and are able to adapt to the task they are supposed to do [Gupta].
1.3 Neural Networks 19 Neural Networks are large structured systems of equations. These systems have many degrees of freedom and are able to adapt to the task they are supposed to do [Gupta]. Two very
More informationMachine Learning and Pattern Recognition Logistic Regression
Machine Learning and Pattern Recognition Logistic Regression Course Lecturer:Amos J Storkey Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh Crichton Street,
More informationChapter 4: Artificial Neural Networks
Chapter 4: Artificial Neural Networks CS 536: Machine Learning Littman (Wu, TA) Administration icml-03: instructional Conference on Machine Learning http://www.cs.rutgers.edu/~mlittman/courses/ml03/icml03/
More informationLecture 8 February 4
ICS273A: Machine Learning Winter 2008 Lecture 8 February 4 Scribe: Carlos Agell (Student) Lecturer: Deva Ramanan 8.1 Neural Nets 8.1.1 Logistic Regression Recall the logistic function: g(x) = 1 1 + e θt
More informationNeural Networks and Support Vector Machines
INF5390 - Kunstig intelligens Neural Networks and Support Vector Machines Roar Fjellheim INF5390-13 Neural Networks and SVM 1 Outline Neural networks Perceptrons Neural networks Support vector machines
More informationLinear Threshold Units
Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear
More informationPredictive Dynamix Inc
Predictive Modeling Technology Predictive modeling is concerned with analyzing patterns and trends in historical and operational data in order to transform data into actionable decisions. This is accomplished
More informationLinear Algebra Notes
Linear Algebra Notes Chapter 19 KERNEL AND IMAGE OF A MATRIX Take an n m matrix a 11 a 12 a 1m a 21 a 22 a 2m a n1 a n2 a nm and think of it as a function A : R m R n The kernel of A is defined as Note
More informationArtificial neural networks
Artificial neural networks Now Neurons Neuron models Perceptron learning Multi-layer perceptrons Backpropagation 2 It all starts with a neuron 3 Some facts about human brain ~ 86 billion neurons ~ 10 15
More informationArtificial Neural Networks and Support Vector Machines. CS 486/686: Introduction to Artificial Intelligence
Artificial Neural Networks and Support Vector Machines CS 486/686: Introduction to Artificial Intelligence 1 Outline What is a Neural Network? - Perceptron learners - Multi-layer networks What is a Support
More informationAlgebra I Notes Relations and Functions Unit 03a
OBJECTIVES: F.IF.A.1 Understand the concept of a function and use function notation. Understand that a function from one set (called the domain) to another set (called the range) assigns to each element
More information3. INNER PRODUCT SPACES
. INNER PRODUCT SPACES.. Definition So far we have studied abstract vector spaces. These are a generalisation of the geometric spaces R and R. But these have more structure than just that of a vector space.
More informationBiological Neurons and Neural Networks, Artificial Neurons
Biological Neurons and Neural Networks, Artificial Neurons Neural Computation : Lecture 2 John A. Bullinaria, 2015 1. Organization of the Nervous System and Brain 2. Brains versus Computers: Some Numbers
More informationIntroduction to Machine Learning and Data Mining. Prof. Dr. Igor Trajkovski trajkovski@nyus.edu.mk
Introduction to Machine Learning and Data Mining Prof. Dr. Igor Trakovski trakovski@nyus.edu.mk Neural Networks 2 Neural Networks Analogy to biological neural systems, the most robust learning systems
More informationIntroduction to Machine Learning Using Python. Vikram Kamath
Introduction to Machine Learning Using Python Vikram Kamath Contents: 1. 2. 3. 4. 5. 6. 7. 8. 9. 10. Introduction/Definition Where and Why ML is used Types of Learning Supervised Learning Linear Regression
More informationDERIVATIVES AS MATRICES; CHAIN RULE
DERIVATIVES AS MATRICES; CHAIN RULE 1. Derivatives of Real-valued Functions Let s first consider functions f : R 2 R. Recall that if the partial derivatives of f exist at the point (x 0, y 0 ), then we
More informationAn Introduction to Neural Networks
An Introduction to Vincent Cheung Kevin Cannons Signal & Data Compression Laboratory Electrical & Computer Engineering University of Manitoba Winnipeg, Manitoba, Canada Advisor: Dr. W. Kinsner May 27,
More informationA Simple Feature Extraction Technique of a Pattern By Hopfield Network
A Simple Feature Extraction Technique of a Pattern By Hopfield Network A.Nag!, S. Biswas *, D. Sarkar *, P.P. Sarkar *, B. Gupta **! Academy of Technology, Hoogly - 722 *USIC, University of Kalyani, Kalyani
More informationRecurrent Neural Networks
Recurrent Neural Networks Neural Computation : Lecture 12 John A. Bullinaria, 2015 1. Recurrent Neural Network Architectures 2. State Space Models and Dynamical Systems 3. Backpropagation Through Time
More information6.2.8 Neural networks for data mining
6.2.8 Neural networks for data mining Walter Kosters 1 In many application areas neural networks are known to be valuable tools. This also holds for data mining. In this chapter we discuss the use of neural
More informationRole of Neural network in data mining
Role of Neural network in data mining Chitranjanjit kaur Associate Prof Guru Nanak College, Sukhchainana Phagwara,(GNDU) Punjab, India Pooja kapoor Associate Prof Swami Sarvanand Group Of Institutes Dinanagar(PTU)
More informationChapter 19. General Matrices. An n m matrix is an array. a 11 a 12 a 1m a 21 a 22 a 2m A = a n1 a n2 a nm. The matrix A has n row vectors
Chapter 9. General Matrices An n m matrix is an array a a a m a a a m... = [a ij]. a n a n a nm The matrix A has n row vectors and m column vectors row i (A) = [a i, a i,..., a im ] R m a j a j a nj col
More informationPATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical
More informationLecture 6. Artificial Neural Networks
Lecture 6 Artificial Neural Networks 1 1 Artificial Neural Networks In this note we provide an overview of the key concepts that have led to the emergence of Artificial Neural Networks as a major paradigm
More informationThis unit will lay the groundwork for later units where the students will extend this knowledge to quadratic and exponential functions.
Algebra I Overview View unit yearlong overview here Many of the concepts presented in Algebra I are progressions of concepts that were introduced in grades 6 through 8. The content presented in this course
More informationSystems of Linear Equations
Systems of Linear Equations Beifang Chen Systems of linear equations Linear systems A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where a, a,, a n and
More informationNeural Networks in Quantitative Finance
Neural Networks in Quantitative Finance Master Thesis submitted to Prof. Dr. Wolfgang Härdle Institute for Statistics and Econometrics CASE - Center for Applied Statistics and Economics Humboldt-Universität
More information5.3 The Cross Product in R 3
53 The Cross Product in R 3 Definition 531 Let u = [u 1, u 2, u 3 ] and v = [v 1, v 2, v 3 ] Then the vector given by [u 2 v 3 u 3 v 2, u 3 v 1 u 1 v 3, u 1 v 2 u 2 v 1 ] is called the cross product (or
More informationVector and Matrix Norms
Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a non-empty
More informationBinary Adders: Half Adders and Full Adders
Binary Adders: Half Adders and Full Adders In this set of slides, we present the two basic types of adders: 1. Half adders, and 2. Full adders. Each type of adder functions to add two binary bits. In order
More informationIntroduction to Artificial Neural Networks
POLYTECHNIC UNIVERSITY Department of Computer and Information Science Introduction to Artificial Neural Networks K. Ming Leung Abstract: A computing paradigm known as artificial neural network is introduced.
More information1 Determinants and the Solvability of Linear Systems
1 Determinants and the Solvability of Linear Systems In the last section we learned how to use Gaussian elimination to solve linear systems of n equations in n unknowns The section completely side-stepped
More informationStock Prediction using Artificial Neural Networks
Stock Prediction using Artificial Neural Networks Abhishek Kar (Y8021), Dept. of Computer Science and Engineering, IIT Kanpur Abstract In this work we present an Artificial Neural Network approach to predict
More information1. Classification problems
Neural and Evolutionary Computing. Lab 1: Classification problems Machine Learning test data repository Weka data mining platform Introduction Scilab 1. Classification problems The main aim of a classification
More informationSection 1.7 22 Continued
Section 1.5 23 A homogeneous equation is always consistent. TRUE - The trivial solution is always a solution. The equation Ax = 0 gives an explicit descriptions of its solution set. FALSE - The equation
More informationPerformance Evaluation of Artificial Neural. Networks for Spatial Data Analysis
Contemporary Engineering Sciences, Vol. 4, 2011, no. 4, 149-163 Performance Evaluation of Artificial Neural Networks for Spatial Data Analysis Akram A. Moustafa Department of Computer Science Al al-bayt
More informationNeural Computation - Assignment
Neural Computation - Assignment Analysing a Neural Network trained by Backpropagation AA SSt t aa t i iss i t i icc aa l l AA nn aa l lyy l ss i iss i oo f vv aa r i ioo i uu ss l lee l aa r nn i inn gg
More informationProgramming Exercise 3: Multi-class Classification and Neural Networks
Programming Exercise 3: Multi-class Classification and Neural Networks Machine Learning November 4, 2011 Introduction In this exercise, you will implement one-vs-all logistic regression and neural networks
More informationA Simple Introduction to Support Vector Machines
A Simple Introduction to Support Vector Machines Martin Law Lecture for CSE 802 Department of Computer Science and Engineering Michigan State University Outline A brief history of SVM Large-margin linear
More information/SOLUTIONS/ where a, b, c and d are positive constants. Study the stability of the equilibria of this system based on linearization.
echnische Universiteit Eindhoven Faculteit Elektrotechniek NIE-LINEAIRE SYSEMEN / NEURALE NEWERKEN (P6) gehouden op donderdag maart 7, van 9: tot : uur. Dit examenonderdeel bestaat uit 8 opgaven. /SOLUIONS/
More informationSection 1.1. Introduction to R n
The Calculus of Functions of Several Variables Section. Introduction to R n Calculus is the study of functional relationships and how related quantities change with each other. In your first exposure to
More informationAnalecta Vol. 8, No. 2 ISSN 2064-7964
EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,
More informationSelf Organizing Maps: Fundamentals
Self Organizing Maps: Fundamentals Introduction to Neural Networks : Lecture 16 John A. Bullinaria, 2004 1. What is a Self Organizing Map? 2. Topographic Maps 3. Setting up a Self Organizing Map 4. Kohonen
More informationEQUATIONS and INEQUALITIES
EQUATIONS and INEQUALITIES Linear Equations and Slope 1. Slope a. Calculate the slope of a line given two points b. Calculate the slope of a line parallel to a given line. c. Calculate the slope of a line
More informationAlgebra Unpacked Content For the new Common Core standards that will be effective in all North Carolina schools in the 2012-13 school year.
This document is designed to help North Carolina educators teach the Common Core (Standard Course of Study). NCDPI staff are continually updating and improving these tools to better serve teachers. Algebra
More informationSolving Simultaneous Equations and Matrices
Solving Simultaneous Equations and Matrices The following represents a systematic investigation for the steps used to solve two simultaneous linear equations in two unknowns. The motivation for considering
More informationOperation Count; Numerical Linear Algebra
10 Operation Count; Numerical Linear Algebra 10.1 Introduction Many computations are limited simply by the sheer number of required additions, multiplications, or function evaluations. If floating-point
More information15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
More informationThe Backpropagation Algorithm
7 The Backpropagation Algorithm 7. Learning as gradient descent We saw in the last chapter that multilayered networks are capable of computing a wider range of Boolean functions than networks with a single
More informationKEANSBURG SCHOOL DISTRICT KEANSBURG HIGH SCHOOL Mathematics Department. HSPA 10 Curriculum. September 2007
KEANSBURG HIGH SCHOOL Mathematics Department HSPA 10 Curriculum September 2007 Written by: Karen Egan Mathematics Supervisor: Ann Gagliardi 7 days Sample and Display Data (Chapter 1 pp. 4-47) Surveys and
More informationMATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +
More informationComponent Ordering in Independent Component Analysis Based on Data Power
Component Ordering in Independent Component Analysis Based on Data Power Anne Hendrikse Raymond Veldhuis University of Twente University of Twente Fac. EEMCS, Signals and Systems Group Fac. EEMCS, Signals
More informationNeural Networks algorithms and applications
Neural Networks algorithms and applications By Fiona Nielsen 4i 12/12-2001 Supervisor: Geert Rasmussen Niels Brock Business College 1 Introduction Neural Networks is a field of Artificial Intelligence
More informationThe Counterpropagation Network
214 The Counterpropagation Network The Counterpropagation Network " The Counterpropagation network (CPN) is the most recently developed of the models that we have discussed so far in this text. The CPN
More informationANN Based Fault Classifier and Fault Locator for Double Circuit Transmission Line
International Journal of Computer Sciences and Engineering Open Access Research Paper Volume-4, Special Issue-2, April 2016 E-ISSN: 2347-2693 ANN Based Fault Classifier and Fault Locator for Double Circuit
More informationChapter 9. Systems of Linear Equations
Chapter 9. Systems of Linear Equations 9.1. Solve Systems of Linear Equations by Graphing KYOTE Standards: CR 21; CA 13 In this section we discuss how to solve systems of two linear equations in two variables
More informationA fast modified constructive-covering algorithm for binary multi-layer neural networks
Neurocomputing 7 (26) 445 46 www.elsevier.com/locate/neucom A fast modified constructive-covering algorithm for binary multi-layer neural networks Di Wang a,, Narendra. S. Chaudhari b a School of Informatics,
More informationLeast-Squares Intersection of Lines
Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a
More informationNonlinear Iterative Partial Least Squares Method
Numerical Methods for Determining Principal Component Analysis Abstract Factors Béchu, S., Richard-Plouet, M., Fernandez, V., Walton, J., and Fairley, N. (2016) Developments in numerical treatments for
More informationSection 1.4. Lines, Planes, and Hyperplanes. The Calculus of Functions of Several Variables
The Calculus of Functions of Several Variables Section 1.4 Lines, Planes, Hyperplanes In this section we will add to our basic geometric understing of R n by studying lines planes. If we do this carefully,
More informationCORRELATED TO THE SOUTH CAROLINA COLLEGE AND CAREER-READY FOUNDATIONS IN ALGEBRA
We Can Early Learning Curriculum PreK Grades 8 12 INSIDE ALGEBRA, GRADES 8 12 CORRELATED TO THE SOUTH CAROLINA COLLEGE AND CAREER-READY FOUNDATIONS IN ALGEBRA April 2016 www.voyagersopris.com Mathematical
More informationSupport Vector Machines Explained
March 1, 2009 Support Vector Machines Explained Tristan Fletcher www.cs.ucl.ac.uk/staff/t.fletcher/ Introduction This document has been written in an attempt to make the Support Vector Machines (SVM),
More informationCreating, Solving, and Graphing Systems of Linear Equations and Linear Inequalities
Algebra 1, Quarter 2, Unit 2.1 Creating, Solving, and Graphing Systems of Linear Equations and Linear Inequalities Overview Number of instructional days: 15 (1 day = 45 60 minutes) Content to be learned
More informationMA 323 Geometric Modelling Course Notes: Day 02 Model Construction Problem
MA 323 Geometric Modelling Course Notes: Day 02 Model Construction Problem David L. Finn November 30th, 2004 In the next few days, we will introduce some of the basic problems in geometric modelling, and
More informationKarnaugh Maps. Circuit-wise, this leads to a minimal two-level implementation
Karnaugh Maps Applications of Boolean logic to circuit design The basic Boolean operations are AND, OR and NOT These operations can be combined to form complex expressions, which can also be directly translated
More informationClassification of Fingerprints. Sarat C. Dass Department of Statistics & Probability
Classification of Fingerprints Sarat C. Dass Department of Statistics & Probability Fingerprint Classification Fingerprint classification is a coarse level partitioning of a fingerprint database into smaller
More informationMetric Spaces. Chapter 7. 7.1. Metrics
Chapter 7 Metric Spaces A metric space is a set X that has a notion of the distance d(x, y) between every pair of points x, y X. The purpose of this chapter is to introduce metric spaces and give some
More informationLINEAR EQUATIONS IN TWO VARIABLES
66 MATHEMATICS CHAPTER 4 LINEAR EQUATIONS IN TWO VARIABLES The principal use of the Analytic Art is to bring Mathematical Problems to Equations and to exhibit those Equations in the most simple terms that
More informationData Mining using Artificial Neural Network Rules
Data Mining using Artificial Neural Network Rules Pushkar Shinde MCOERC, Nasik Abstract - Diabetes patients are increasing in number so it is necessary to predict, treat and diagnose the disease. Data
More informationHow To Use Neural Networks In Data Mining
International Journal of Electronics and Computer Science Engineering 1449 Available Online at www.ijecse.org ISSN- 2277-1956 Neural Networks in Data Mining Priyanka Gaur Department of Information and
More informationMachine Learning and Data Mining. Regression Problem. (adapted from) Prof. Alexander Ihler
Machine Learning and Data Mining Regression Problem (adapted from) Prof. Alexander Ihler Overview Regression Problem Definition and define parameters ϴ. Prediction using ϴ as parameters Measure the error
More informationECE 842 Report Implementation of Elliptic Curve Cryptography
ECE 842 Report Implementation of Elliptic Curve Cryptography Wei-Yang Lin December 15, 2004 Abstract The aim of this report is to illustrate the issues in implementing a practical elliptic curve cryptographic
More informationRepresenting Vector Fields Using Field Line Diagrams
Minds On Physics Activity FFá2 5 Representing Vector Fields Using Field Line Diagrams Purpose and Expected Outcome One way of representing vector fields is using arrows to indicate the strength and direction
More information1. True or False? A voltage level in the range 0 to 2 volts is interpreted as a binary 1.
File: chap04, Chapter 04 1. True or False? A voltage level in the range 0 to 2 volts is interpreted as a binary 1. 2. True or False? A gate is a device that accepts a single input signal and produces one
More informationAnalysis of Multilayer Neural Networks with Direct and Cross-Forward Connection
Analysis of Multilayer Neural Networks with Direct and Cross-Forward Connection Stanis law P laczek and Bijaya Adhikari Vistula University, Warsaw, Poland stanislaw.placzek@wp.pl,bijaya.adhikari1991@gmail.com
More informationA simple application of Artificial Neural Network to cloud classification
A simple application of Artificial Neural Network to cloud classification Tianle Yuan For AOSC 630 (by Prof. Kalnay) Introduction to Pattern Recognition (PR) Example1: visual separation between the character
More informationDecember 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS
December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B KITCHENS The equation 1 Lines in two-dimensional space (1) 2x y = 3 describes a line in two-dimensional space The coefficients of x and y in the equation
More informationNeural network software tool development: exploring programming language options
INEB- PSI Technical Report 2006-1 Neural network software tool development: exploring programming language options Alexandra Oliveira aao@fe.up.pt Supervisor: Professor Joaquim Marques de Sá June 2006
More informationLecture 2: August 29. Linear Programming (part I)
10-725: Convex Optimization Fall 2013 Lecture 2: August 29 Lecturer: Barnabás Póczos Scribes: Samrachana Adhikari, Mattia Ciollaro, Fabrizio Lecci Note: LaTeX template courtesy of UC Berkeley EECS dept.
More informationBig Data Analytics CSCI 4030
High dim. data Graph data Infinite data Machine learning Apps Locality sensitive hashing PageRank, SimRank Filtering data streams SVM Recommen der systems Clustering Community Detection Web advertising
More informationLecture 3: Linear methods for classification
Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,
More informationMATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.
MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column
More informationAUTOMATION OF ENERGY DEMAND FORECASTING. Sanzad Siddique, B.S.
AUTOMATION OF ENERGY DEMAND FORECASTING by Sanzad Siddique, B.S. A Thesis submitted to the Faculty of the Graduate School, Marquette University, in Partial Fulfillment of the Requirements for the Degree
More informationNeural Network Toolbox
Neural Network Toolbox A Tutorial for the Course Computational Intelligence http://www.igi.tugraz.at/lehre/ci Stefan Häusler Institute for Theoretical Computer Science Inffeldgasse 16b/I Abstract This
More informationMATHS LEVEL DESCRIPTORS
MATHS LEVEL DESCRIPTORS Number Level 3 Understand the place value of numbers up to thousands. Order numbers up to 9999. Round numbers to the nearest 10 or 100. Understand the number line below zero, and
More informationSELECTING NEURAL NETWORK ARCHITECTURE FOR INVESTMENT PROFITABILITY PREDICTIONS
UDC: 004.8 Original scientific paper SELECTING NEURAL NETWORK ARCHITECTURE FOR INVESTMENT PROFITABILITY PREDICTIONS Tonimir Kišasondi, Alen Lovren i University of Zagreb, Faculty of Organization and Informatics,
More informationSYSTEMS OF EQUATIONS AND MATRICES WITH THE TI-89. by Joseph Collison
SYSTEMS OF EQUATIONS AND MATRICES WITH THE TI-89 by Joseph Collison Copyright 2000 by Joseph Collison All rights reserved Reproduction or translation of any part of this work beyond that permitted by Sections
More informationTRAINING A LIMITED-INTERCONNECT, SYNTHETIC NEURAL IC
777 TRAINING A LIMITED-INTERCONNECT, SYNTHETIC NEURAL IC M.R. Walker. S. Haghighi. A. Afghan. and L.A. Akers Center for Solid State Electronics Research Arizona State University Tempe. AZ 85287-6206 mwalker@enuxha.eas.asu.edu
More informationLecture 1: Systems of Linear Equations
MTH Elementary Matrix Algebra Professor Chao Huang Department of Mathematics and Statistics Wright State University Lecture 1 Systems of Linear Equations ² Systems of two linear equations with two variables
More information8 Square matrices continued: Determinants
8 Square matrices continued: Determinants 8. Introduction Determinants give us important information about square matrices, and, as we ll soon see, are essential for the computation of eigenvalues. You
More informationNEURAL NETWORKS A Comprehensive Foundation
NEURAL NETWORKS A Comprehensive Foundation Second Edition Simon Haykin McMaster University Hamilton, Ontario, Canada Prentice Hall Prentice Hall Upper Saddle River; New Jersey 07458 Preface xii Acknowledgments
More informationMath Review. for the Quantitative Reasoning Measure of the GRE revised General Test
Math Review for the Quantitative Reasoning Measure of the GRE revised General Test www.ets.org Overview This Math Review will familiarize you with the mathematical skills and concepts that are important
More informationFigure 1.1 Vector A and Vector F
CHAPTER I VECTOR QUANTITIES Quantities are anything which can be measured, and stated with number. Quantities in physics are divided into two types; scalar and vector quantities. Scalar quantities have
More informationRotation Matrices and Homogeneous Transformations
Rotation Matrices and Homogeneous Transformations A coordinate frame in an n-dimensional space is defined by n mutually orthogonal unit vectors. In particular, for a two-dimensional (2D) space, i.e., n
More informationSupport Vector Machine (SVM)
Support Vector Machine (SVM) CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
More informationAlgebra I Vocabulary Cards
Algebra I Vocabulary Cards Table of Contents Expressions and Operations Natural Numbers Whole Numbers Integers Rational Numbers Irrational Numbers Real Numbers Absolute Value Order of Operations Expression
More informationNeural Networks in Data Mining
IOSR Journal of Engineering (IOSRJEN) ISSN (e): 2250-3021, ISSN (p): 2278-8719 Vol. 04, Issue 03 (March. 2014), V6 PP 01-06 www.iosrjen.org Neural Networks in Data Mining Ripundeep Singh Gill, Ashima Department
More informationa 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.
Chapter 1 LINEAR EQUATIONS 1.1 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,..., a n, b are given
More information