Recurrent Neural Networks A Brief Overview

Size: px
Start display at page:

Download "Recurrent Neural Networks A Brief Overview"

Transcription

1 Recurrent Neural Networks A Brief Overview Douglas Eck University of Montreal RNN Overview Oct p.1/33

2 RNNs versus FFNs Feed-Forward Networks (FFNs, left) can learn function mappings (FFN==>FIR filter) Recurrent Neural Networks (RNNs, right) use hidden layer as memory store to learn sequences (RNN==>IIR filter) RNNs can (in principle at least) exhiit virtually unlimited temporal dynamics RNN Overview Oct p.2/33

3 Several Methods SRN Simple Recurrent Network (Elman, 1990) BPTT Backpropagation Through Time (Rumelhart, Hinton & Williams, 1986) RTRL Real Time Recurrent Learning (Williams & Zipser, 1989) LSTM Long Short-Term Memory (Hochreiter & Schmidhuer, 1996) RNN Overview Oct p.3/33

4 SRN (Elman Net) Simple Recurrent Networks Hidden layer activations copied into a copy layer Cycles are eliminated, allowing use of standard ackpropagation RNN Overview Oct p.4/33

5 Some Oservations y Elman Some prolems change nature when expressed in time Example: Temporal XOR (Elman, 1990) [ ] RNN learns Frequency Detectors Error can give information aout temporal structure of input Increasing sequential dependencies does not necessarily make task harder Representation of time is task-dependent RNN Overview Oct p.5/33

6 SRN Strengths Easy to train Potential for complex and useful temporal dynamics Can induce hierarchical temporal structure Can learn, e.g., simple natural language grammar RNN Overview Oct p.6/33

7 SRN Shortcomings We need to compute weight changes at every timestep t0 < t t1: t1 t=t0+1 w ij(t) Thus we need to compute at every t: E(t) w ij = k U (t) y k (t) y k (t) w ij = k U e k(t) y k(t) w ij Since we know e k (t) we need only compute y k(t) w ij SRN Truncates this derivative Long-timescale (and short-timescale) dependencies etween error signals are lost RNN Overview Oct p.7/33

8 SRNs Generalized: BPTT Generalize SRN to rememer deeper into past Trick: unfold network to represent time spatially (one layer per discrete timestep) Still no cycles, allowing use of standard ackpropagation RNN Overview Oct p.8/33

9 BPTT( ) For each timestep t Current state of network and input pattern is added to history uffer (stores since time t=0) Error e k (t) is injected; ǫs and δs for times (t 0 < τ t) are computed: ǫ k (t) = e k (t) δ k (τ) = f k (s k(τ))ǫ k (τ) ǫ k (τ 1) = l inu w lkδ l (τ) Weight changes are computed as in standard BP: = t τ=t 0 +1 δ i(τ)x j (τ 1) E(t) w ij RNN Overview Oct p.9/33

10 Truncated/Epochwise BPTT When training data is in epochs, can limit size of history uffer to h, length of epoch When training data not in epochs, can nonetheless truncate gradient after h timesteps For n units and O(n 2 ) weights, epochwise BPTT has space complexity O(nh) and time complexity O(n 2 h) Compares favoraly to BPTT (sustitute L, length of input sequence, for h) RNN Overview Oct p.10/33

11 BPTT s Gradient Recall that BPTT( ) computes = t τ=t 0 +1 δ i(τ)x j (τ 1) E(t) w ij Thus errors at t take into account equally δ values from the entire history of computation Truncated/Epochwise BPTT cuts off the gradient: = t τ=t h δ i(τ)x j (τ 1) E(t) w ij If data is naturally organized in epochs, this is not a prolem If data is not organized in epochs, this is not a prolem either, provided h is ig enough RNN Overview Oct p.11/33

12 RTRL RTRL=Real Time Recurrent Learning Instead of unfolding network ackward in time, one can propagate error forward in time Compute directly p k ij (t) = y k(t) w ij p k ij (t + 1) = f k (s k(t)) l U w klp l ij (t) + δ ikx j (t) w ij = α E(t) w ij = α k U e k(t)p k ij (t) RNN Overview Oct p.12/33

13 RTRL vs BPTT RTRL saves from executing ackward dynamics Temporal credit assignment solved during forward pass RTRL is painfully slow: for n units and n 2 weights, O(n 2 ) space complexity and O(n 4 )(!) time complexity Because BPTT is faster, in general RTRL is only of theoretical interest RNN Overview Oct p.13/33

14 Training Paradigms Epochwise training: reset system at fixed stopping points Continual training: never reset system Epochwise training provides arrier for credit assignment Unnatural for many tasks Distinct from whether weights are atchwise or iteratively updated RNN Overview Oct p.14/33

15 Teacher Forcing Continually-trained networks can move far from desired trajectory and never return Especially true if network enters region where activations ecome saturated (gradients go to 0) Solution; replace during training actual output y k (t) with teacher signal d k (t). Necessary for certain prolems in BPTT/RTRL networks (e.g. generating square waves) RNN Overview Oct p.15/33

16 Puzzle: How Much is Enough? Recall that for Truncated BPTT, the true gradient is not eing computed How many steps do we need to get close enough? RNN Overview Oct p.16/33

17 Puzzle: How Much is Enough? Recall that for Truncated BPTT, the true gradient is not eing computed How many steps do we need to get close enough? Answer: certainly not more than 50; Proaly not more than 10(!) RNN Overview Oct p.17/33

18 Credit Assignment is Difficult In BPTT, error gets diluted with every susequent layer (credit assignment prolem): ǫ k (t) = e k (t) δ k (τ) = f k (s k(τ))ǫ k (τ) ǫ k (τ 1) = l inu w lkδ l (τ) The logistic sigmoid 1.0/(1.0 + (exp( x))) has maximum derivative of When w < 4.0, error is always<1.0 RNN Overview Oct p.18/33

19 Vanishing Gradients Bengio, Simard & Frasconi (1994) Rememering a it requires creation of attractor asin in state space Two cases: either system overly sensitive to noise or error gradient vanishes exponentially General prolem (HMMs suffer something similar) RNN Overview Oct p.19/33

20 Solutions and Alternatives Non-gradient learning algorithms (including gloal search) Expectation Maximization (EM) training (e.g. Bengio IO/HMM) Hyrid architectures to aid in preserving error signals RNN Overview Oct p.20/33

21 LSTM memory lock with single cell y[t] u[t] Hyrid recurrent neural network Make hidden units linear (derivative is then 1.0) Linear units unstale Place units in a memory lock protected y multiplicative gates RNN Overview Oct p.21/33

22 Inside an LSTM Memory Block g(net c ) s c h(s c ) y c y in y ϕ y out Gates are standard sigmoidal units Input gate y in protects linear unit s c from spurious inputs Forget gate y φ allows s c to empty own contents Output gate y out allows lock to take itself offline and ignore error RNN Overview Oct p.22/33

23 LSTM Learning For linear unit s c, a truncated RTRL approach is used (tales of partial derivatives) Everywhere else, standard ackpropagation is used Rationale: due to vanishing gradient prolem, errors would decay anyway RNN Overview Oct p.23/33

24 Properties of LSTM Very good at finding hierarchical structure Can induce nonlinear oscillation (for counting and timing) But error flow among locks truncated Difficult to train: weights into gates are sensitive RNN Overview Oct p.24/33

25 Formal Grammars LSTM solves Emedded Reer Grammar faster, more relialy and with smaller hidden layer than RTRL/BPTT LSTM solves CSL A n B n C n etter than RTRL/BPTT networks Trained on examples with n < 10 LSTM generalized to n > 1000 BPTT-trained networks generalize to n 18 in est case (Bodén & Wiles, 2001) a T S a a a a a a a a a a a c c c c c c c c c c c c T : input : target s y s y c 1 c 1 c2 c 2 RNN Overview Oct p.25/33

Recurrent Neural Networks

Recurrent Neural Networks Recurrent Neural Networks Neural Computation : Lecture 12 John A. Bullinaria, 2015 1. Recurrent Neural Network Architectures 2. State Space Models and Dynamical Systems 3. Backpropagation Through Time

More information

Introduction to Machine Learning and Data Mining. Prof. Dr. Igor Trajkovski trajkovski@nyus.edu.mk

Introduction to Machine Learning and Data Mining. Prof. Dr. Igor Trajkovski trajkovski@nyus.edu.mk Introduction to Machine Learning and Data Mining Prof. Dr. Igor Trakovski trakovski@nyus.edu.mk Neural Networks 2 Neural Networks Analogy to biological neural systems, the most robust learning systems

More information

LONG SHORT-TERM MEMORY Neural Computation 9(8):1735{1780, 1997 Sepp Hochreiter Fakultat fur Informatik Technische Universitat Munchen 80290 Munchen, Germany hochreit@informatik.tu-muenchen.de http://www7.informatik.tu-muenchen.de/~hochreit

More information

Sequence to Sequence Weather Forecasting with Long Short-Term Memory Recurrent Neural Networks

Sequence to Sequence Weather Forecasting with Long Short-Term Memory Recurrent Neural Networks Volume 143 - No.11, June 16 Sequence to Sequence Weather Forecasting with Long Short-Term Memory Recurrent Neural Networks Mohamed Akram Zaytar Research Student Department of Computer Engineering Faculty

More information

Neural Networks and Support Vector Machines

Neural Networks and Support Vector Machines INF5390 - Kunstig intelligens Neural Networks and Support Vector Machines Roar Fjellheim INF5390-13 Neural Networks and SVM 1 Outline Neural networks Perceptrons Neural networks Support vector machines

More information

NEURAL NETWORKS A Comprehensive Foundation

NEURAL NETWORKS A Comprehensive Foundation NEURAL NETWORKS A Comprehensive Foundation Second Edition Simon Haykin McMaster University Hamilton, Ontario, Canada Prentice Hall Prentice Hall Upper Saddle River; New Jersey 07458 Preface xii Acknowledgments

More information

SELECTING NEURAL NETWORK ARCHITECTURE FOR INVESTMENT PROFITABILITY PREDICTIONS

SELECTING NEURAL NETWORK ARCHITECTURE FOR INVESTMENT PROFITABILITY PREDICTIONS UDC: 004.8 Original scientific paper SELECTING NEURAL NETWORK ARCHITECTURE FOR INVESTMENT PROFITABILITY PREDICTIONS Tonimir Kišasondi, Alen Lovren i University of Zagreb, Faculty of Organization and Informatics,

More information

Performance Evaluation On Human Resource Management Of China S Commercial Banks Based On Improved Bp Neural Networks

Performance Evaluation On Human Resource Management Of China S Commercial Banks Based On Improved Bp Neural Networks Performance Evaluation On Human Resource Management Of China S *1 Honglei Zhang, 2 Wenshan Yuan, 1 Hua Jiang 1 School of Economics and Management, Hebei University of Engineering, Handan 056038, P. R.

More information

A generalized LSTM-like training algorithm for second-order recurrent neural networks

A generalized LSTM-like training algorithm for second-order recurrent neural networks A generalized LSTM-lie training algorithm for second-order recurrent neural networs Dere D. Monner, James A. Reggia Department of Computer Science, University of Maryland, College Par, MD 20742, USA Abstract

More information

Lecture 8 February 4

Lecture 8 February 4 ICS273A: Machine Learning Winter 2008 Lecture 8 February 4 Scribe: Carlos Agell (Student) Lecturer: Deva Ramanan 8.1 Neural Nets 8.1.1 Logistic Regression Recall the logistic function: g(x) = 1 1 + e θt

More information

PhD in Computer Science and Engineering Bologna, April 2016. Machine Learning. Marco Lippi. marco.lippi3@unibo.it. Marco Lippi Machine Learning 1 / 80

PhD in Computer Science and Engineering Bologna, April 2016. Machine Learning. Marco Lippi. marco.lippi3@unibo.it. Marco Lippi Machine Learning 1 / 80 PhD in Computer Science and Engineering Bologna, April 2016 Machine Learning Marco Lippi marco.lippi3@unibo.it Marco Lippi Machine Learning 1 / 80 Recurrent Neural Networks Marco Lippi Machine Learning

More information

Neural Network Design in Cloud Computing

Neural Network Design in Cloud Computing International Journal of Computer Trends and Technology- volume4issue2-2013 ABSTRACT: Neural Network Design in Cloud Computing B.Rajkumar #1,T.Gopikiran #2,S.Satyanarayana *3 #1,#2Department of Computer

More information

Long Short Term Memory Networks for Anomaly Detection in Time Series

Long Short Term Memory Networks for Anomaly Detection in Time Series Long Short Term Memory Networks for Anomaly Detection in Time Series Pankaj Malhotra1, Lovekesh Vig2, Gautam Shroff1, Puneet Agarwal1 1- TCS Research, Delhi, India 2- Jawaharlal Nehru University, New Delhi,

More information

SEMINAR OUTLINE. Introduction to Data Mining Using Artificial Neural Networks. Definitions of Neural Networks. Definitions of Neural Networks

SEMINAR OUTLINE. Introduction to Data Mining Using Artificial Neural Networks. Definitions of Neural Networks. Definitions of Neural Networks SEMINAR OUTLINE Introduction to Data Mining Using Artificial Neural Networks ISM 611 Dr. Hamid Nemati Introduction to and Characteristics of Neural Networks Comparison of Neural Networks to traditional

More information

Dynamic Factor Graphs for Time Series Modeling

Dynamic Factor Graphs for Time Series Modeling Dynamic Factor Graphs for Time Series Modeling Piotr Mirowski and Yann LeCun Courant Institute of Mathematical Sciences, New York University, 719 Broadway, New York, NY 10003 USA {mirowski,yann}@cs.nyu.edu

More information

Analecta Vol. 8, No. 2 ISSN 2064-7964

Analecta Vol. 8, No. 2 ISSN 2064-7964 EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,

More information

Artificial Neural Networks and Support Vector Machines. CS 486/686: Introduction to Artificial Intelligence

Artificial Neural Networks and Support Vector Machines. CS 486/686: Introduction to Artificial Intelligence Artificial Neural Networks and Support Vector Machines CS 486/686: Introduction to Artificial Intelligence 1 Outline What is a Neural Network? - Perceptron learners - Multi-layer networks What is a Support

More information

Forecasting of Economic Quantities using Fuzzy Autoregressive Model and Fuzzy Neural Network

Forecasting of Economic Quantities using Fuzzy Autoregressive Model and Fuzzy Neural Network Forecasting of Economic Quantities using Fuzzy Autoregressive Model and Fuzzy Neural Network Dušan Marček 1 Abstract Most models for the time series of stock prices have centered on autoregressive (AR)

More information

A tutorial on training recurrent neural networks, covering BPPT, RTRL, EKF and the "echo state network" approach

A tutorial on training recurrent neural networks, covering BPPT, RTRL, EKF and the echo state network approach A tutorial on training recurrent neural networks, covering BPPT, RTRL, EKF and the "echo state network" approach Herbert Jaeger Fraunhofer Institute for Autonomous Intelligent Systems (AIS) since 2003:

More information

Performance Evaluation of Artificial Neural. Networks for Spatial Data Analysis

Performance Evaluation of Artificial Neural. Networks for Spatial Data Analysis Contemporary Engineering Sciences, Vol. 4, 2011, no. 4, 149-163 Performance Evaluation of Artificial Neural Networks for Spatial Data Analysis Akram A. Moustafa Department of Computer Science Al al-bayt

More information

Feedforward Neural Networks and Backpropagation

Feedforward Neural Networks and Backpropagation Feedforward Neural Networks and Backpropagation Feedforward neural networks Architectural issues, computational capabilities Sigmoidal and radial basis functions Gradient-based learning and Backprogation

More information

Efficient online learning of a non-negative sparse autoencoder

Efficient online learning of a non-negative sparse autoencoder and Machine Learning. Bruges (Belgium), 28-30 April 2010, d-side publi., ISBN 2-93030-10-2. Efficient online learning of a non-negative sparse autoencoder Andre Lemme, R. Felix Reinhart and Jochen J. Steil

More information

Machine Learning and Data Mining -

Machine Learning and Data Mining - Machine Learning and Data Mining - Perceptron Neural Networks Nuno Cavalheiro Marques (nmm@di.fct.unl.pt) Spring Semester 2010/2011 MSc in Computer Science Multi Layer Perceptron Neurons and the Perceptron

More information

CHAPTER 5 PREDICTIVE MODELING STUDIES TO DETERMINE THE CONVEYING VELOCITY OF PARTS ON VIBRATORY FEEDER

CHAPTER 5 PREDICTIVE MODELING STUDIES TO DETERMINE THE CONVEYING VELOCITY OF PARTS ON VIBRATORY FEEDER 93 CHAPTER 5 PREDICTIVE MODELING STUDIES TO DETERMINE THE CONVEYING VELOCITY OF PARTS ON VIBRATORY FEEDER 5.1 INTRODUCTION The development of an active trap based feeder for handling brakeliners was discussed

More information

Chapter 4: Artificial Neural Networks

Chapter 4: Artificial Neural Networks Chapter 4: Artificial Neural Networks CS 536: Machine Learning Littman (Wu, TA) Administration icml-03: instructional Conference on Machine Learning http://www.cs.rutgers.edu/~mlittman/courses/ml03/icml03/

More information

Taking Inverse Graphics Seriously

Taking Inverse Graphics Seriously CSC2535: 2013 Advanced Machine Learning Taking Inverse Graphics Seriously Geoffrey Hinton Department of Computer Science University of Toronto The representation used by the neural nets that work best

More information

Feed-Forward mapping networks KAIST 바이오및뇌공학과 정재승

Feed-Forward mapping networks KAIST 바이오및뇌공학과 정재승 Feed-Forward mapping networks KAIST 바이오및뇌공학과 정재승 How much energy do we need for brain functions? Information processing: Trade-off between energy consumption and wiring cost Trade-off between energy consumption

More information

Machine Learning: Multi Layer Perceptrons

Machine Learning: Multi Layer Perceptrons Machine Learning: Multi Layer Perceptrons Prof. Dr. Martin Riedmiller Albert-Ludwigs-University Freiburg AG Maschinelles Lernen Machine Learning: Multi Layer Perceptrons p.1/61 Outline multi layer perceptrons

More information

COMBINED NEURAL NETWORKS FOR TIME SERIES ANALYSIS

COMBINED NEURAL NETWORKS FOR TIME SERIES ANALYSIS COMBINED NEURAL NETWORKS FOR TIME SERIES ANALYSIS Iris Ginzburg and David Horn School of Physics and Astronomy Raymond and Beverly Sackler Faculty of Exact Science Tel-Aviv University Tel-A viv 96678,

More information

TIME SERIES FORECASTING USING NEURAL NETWORKS

TIME SERIES FORECASTING USING NEURAL NETWORKS 1402 Challenges of the Knowledge Society. IT in Social Sciences TIME SERIES FORECASTING USING NEURAL NETWORKS BOGDAN OANCEA * ŞTEFAN CRISTIAN CIUCU ** Abstract Recent studies have shown the classification

More information

Supporting Online Material for

Supporting Online Material for www.sciencemag.org/cgi/content/full/313/5786/504/dc1 Supporting Online Material for Reducing the Dimensionality of Data with Neural Networks G. E. Hinton* and R. R. Salakhutdinov *To whom correspondence

More information

Application of Neural Network in User Authentication for Smart Home System

Application of Neural Network in User Authentication for Smart Home System Application of Neural Network in User Authentication for Smart Home System A. Joseph, D.B.L. Bong, D.A.A. Mat Abstract Security has been an important issue and concern in the smart home systems. Smart

More information

degrees of freedom and are able to adapt to the task they are supposed to do [Gupta].

degrees of freedom and are able to adapt to the task they are supposed to do [Gupta]. 1.3 Neural Networks 19 Neural Networks are large structured systems of equations. These systems have many degrees of freedom and are able to adapt to the task they are supposed to do [Gupta]. Two very

More information

Role of Neural network in data mining

Role of Neural network in data mining Role of Neural network in data mining Chitranjanjit kaur Associate Prof Guru Nanak College, Sukhchainana Phagwara,(GNDU) Punjab, India Pooja kapoor Associate Prof Swami Sarvanand Group Of Institutes Dinanagar(PTU)

More information

Lecture 6. Artificial Neural Networks

Lecture 6. Artificial Neural Networks Lecture 6 Artificial Neural Networks 1 1 Artificial Neural Networks In this note we provide an overview of the key concepts that have led to the emergence of Artificial Neural Networks as a major paradigm

More information

Multiple Layer Perceptron Training Using Genetic Algorithms

Multiple Layer Perceptron Training Using Genetic Algorithms Multiple Layer Perceptron Training Using Genetic Algorithms Udo Seiffert University of South Australia, Adelaide Knowledge-Based Intelligent Engineering Systems Centre (KES) Mawson Lakes, 5095, Adelaide,

More information

Keywords: Image complexity, PSNR, Levenberg-Marquardt, Multi-layer neural network.

Keywords: Image complexity, PSNR, Levenberg-Marquardt, Multi-layer neural network. Global Journal of Computer Science and Technology Volume 11 Issue 3 Version 1.0 Type: Double Blind Peer Reviewed International Research Journal Publisher: Global Journals Inc. (USA) Online ISSN: 0975-4172

More information

Back Propagation Neural Networks User Manual

Back Propagation Neural Networks User Manual Back Propagation Neural Networks User Manual Author: Lukáš Civín Library: BP_network.dll Runnable class: NeuralNetStart Document: Back Propagation Neural Networks Page 1/28 Content: 1 INTRODUCTION TO BACK-PROPAGATION

More information

Follow links Class Use and other Permissions. For more information, send email to: permissions@pupress.princeton.edu

Follow links Class Use and other Permissions. For more information, send email to: permissions@pupress.princeton.edu COPYRIGHT NOTICE: David A. Kendrick, P. Ruben Mercado, and Hans M. Amman: Computational Economics is published by Princeton University Press and copyrighted, 2006, by Princeton University Press. All rights

More information

Self Organizing Maps: Fundamentals

Self Organizing Maps: Fundamentals Self Organizing Maps: Fundamentals Introduction to Neural Networks : Lecture 16 John A. Bullinaria, 2004 1. What is a Self Organizing Map? 2. Topographic Maps 3. Setting up a Self Organizing Map 4. Kohonen

More information

An Introduction to Neural Networks

An Introduction to Neural Networks An Introduction to Vincent Cheung Kevin Cannons Signal & Data Compression Laboratory Electrical & Computer Engineering University of Manitoba Winnipeg, Manitoba, Canada Advisor: Dr. W. Kinsner May 27,

More information

Non-Linear Regression 2006-2008 Samuel L. Baker

Non-Linear Regression 2006-2008 Samuel L. Baker NON-LINEAR REGRESSION 1 Non-Linear Regression 2006-2008 Samuel L. Baker The linear least squares method that you have een using fits a straight line or a flat plane to a unch of data points. Sometimes

More information

A Neural Framework for Robot Motor Learning based on Memory Consolidation

A Neural Framework for Robot Motor Learning based on Memory Consolidation A Neural Framework for Robot Motor Learning based on Memory Consolidation Heni Ben Amor 1, Shuhei Ikemoto 2, Takashi Minato 2, Bernhard Jung 1 and Hiroshi Ishiguro 2 1 VR and Multimedia Group, TU Bergakademie

More information

Stochastic Weight Selection in Backpropagation Through Time

Stochastic Weight Selection in Backpropagation Through Time TECHNICAL UNIVERSITY IN KOSICE Faculty of Electrical Engineering and Informatics Department of Cybernetic and Artificial Intelligence Ing. Juraj Koščák Stochastic Weight Selection in Backpropagation Through

More information

EFFICIENT DATA PRE-PROCESSING FOR DATA MINING

EFFICIENT DATA PRE-PROCESSING FOR DATA MINING EFFICIENT DATA PRE-PROCESSING FOR DATA MINING USING NEURAL NETWORKS JothiKumar.R 1, Sivabalan.R.V 2 1 Research scholar, Noorul Islam University, Nagercoil, India Assistant Professor, Adhiparasakthi College

More information

Field Data Recovery in Tidal System Using Artificial Neural Networks (ANNs)

Field Data Recovery in Tidal System Using Artificial Neural Networks (ANNs) Field Data Recovery in Tidal System Using Artificial Neural Networks (ANNs) by Bernard B. Hsieh and Thad C. Pratt PURPOSE: The field data collection program consumes a major portion of a modeling budget.

More information

Deterministic Sampling-based Switching Kalman Filtering for Vehicle Tracking

Deterministic Sampling-based Switching Kalman Filtering for Vehicle Tracking Proceedings of the IEEE ITSC 2006 2006 IEEE Intelligent Transportation Systems Conference Toronto, Canada, September 17-20, 2006 WA4.1 Deterministic Sampling-based Switching Kalman Filtering for Vehicle

More information

LMS is a simple but powerful algorithm and can be implemented to take advantage of the Lattice FPGA architecture.

LMS is a simple but powerful algorithm and can be implemented to take advantage of the Lattice FPGA architecture. February 2012 Introduction Reference Design RD1031 Adaptive algorithms have become a mainstay in DSP. They are used in wide ranging applications including wireless channel estimation, radar guidance systems,

More information

Neural Networks and Back Propagation Algorithm

Neural Networks and Back Propagation Algorithm Neural Networks and Back Propagation Algorithm Mirza Cilimkovic Institute of Technology Blanchardstown Blanchardstown Road North Dublin 15 Ireland mirzac@gmail.com Abstract Neural Networks (NN) are important

More information

Open Access Research on Application of Neural Network in Computer Network Security Evaluation. Shujuan Jin *

Open Access Research on Application of Neural Network in Computer Network Security Evaluation. Shujuan Jin * Send Orders for Reprints to reprints@benthamscience.ae 766 The Open Electrical & Electronic Engineering Journal, 2014, 8, 766-771 Open Access Research on Application of Neural Network in Computer Network

More information

Neural network models: Foundations and applications to an audit decision problem

Neural network models: Foundations and applications to an audit decision problem Annals of Operations Research 75(1997)291 301 291 Neural network models: Foundations and applications to an audit decision problem Rebecca C. Wu Department of Accounting, College of Management, National

More information

Iranian J Env Health Sci Eng, 2004, Vol.1, No.2, pp.51-57. Application of Intelligent System for Water Treatment Plant Operation.

Iranian J Env Health Sci Eng, 2004, Vol.1, No.2, pp.51-57. Application of Intelligent System for Water Treatment Plant Operation. Iranian J Env Health Sci Eng, 2004, Vol.1, No.2, pp.51-57 Application of Intelligent System for Water Treatment Plant Operation *A Mirsepassi Dept. of Environmental Health Engineering, School of Public

More information

Predict Influencers in the Social Network

Predict Influencers in the Social Network Predict Influencers in the Social Network Ruishan Liu, Yang Zhao and Liuyu Zhou Email: rliu2, yzhao2, lyzhou@stanford.edu Department of Electrical Engineering, Stanford University Abstract Given two persons

More information

Programming Exercise 3: Multi-class Classification and Neural Networks

Programming Exercise 3: Multi-class Classification and Neural Networks Programming Exercise 3: Multi-class Classification and Neural Networks Machine Learning November 4, 2011 Introduction In this exercise, you will implement one-vs-all logistic regression and neural networks

More information

System Identification for Acoustic Comms.:

System Identification for Acoustic Comms.: System Identification for Acoustic Comms.: New Insights and Approaches for Tracking Sparse and Rapidly Fluctuating Channels Weichang Li and James Preisig Woods Hole Oceanographic Institution The demodulation

More information

Price Prediction of Share Market using Artificial Neural Network (ANN)

Price Prediction of Share Market using Artificial Neural Network (ANN) Prediction of Share Market using Artificial Neural Network (ANN) Zabir Haider Khan Department of CSE, SUST, Sylhet, Bangladesh Tasnim Sharmin Alin Department of CSE, SUST, Sylhet, Bangladesh Md. Akter

More information

Learning to Process Natural Language in Big Data Environment

Learning to Process Natural Language in Big Data Environment CCF ADL 2015 Nanchang Oct 11, 2015 Learning to Process Natural Language in Big Data Environment Hang Li Noah s Ark Lab Huawei Technologies Part 1: Deep Learning - Present and Future Talk Outline Overview

More information

Power Prediction Analysis using Artificial Neural Network in MS Excel

Power Prediction Analysis using Artificial Neural Network in MS Excel Power Prediction Analysis using Artificial Neural Network in MS Excel NURHASHINMAH MAHAMAD, MUHAMAD KAMAL B. MOHAMMED AMIN Electronic System Engineering Department Malaysia Japan International Institute

More information

Supply Chain Forecasting Model Using Computational Intelligence Techniques

Supply Chain Forecasting Model Using Computational Intelligence Techniques CMU.J.Nat.Sci Special Issue on Manufacturing Technology (2011) Vol.10(1) 19 Supply Chain Forecasting Model Using Computational Intelligence Techniques Wimalin S. Laosiritaworn Department of Industrial

More information

Learning Recurrent Neural Networks with Hessian-Free Optimization

Learning Recurrent Neural Networks with Hessian-Free Optimization James Martens Ilya Sutskever University of Toronto, Canada JMARTENS@CS.TORONTO.EDU ILYA@CS.UTORONTO.CA Abstract In this work we resolve the long-outstanding problem of how to effectively train recurrent

More information

Comparison of K-means and Backpropagation Data Mining Algorithms

Comparison of K-means and Backpropagation Data Mining Algorithms Comparison of K-means and Backpropagation Data Mining Algorithms Nitu Mathuriya, Dr. Ashish Bansal Abstract Data mining has got more and more mature as a field of basic research in computer science and

More information

THREE DIMENSIONAL REPRESENTATION OF AMINO ACID CHARAC- TERISTICS

THREE DIMENSIONAL REPRESENTATION OF AMINO ACID CHARAC- TERISTICS THREE DIMENSIONAL REPRESENTATION OF AMINO ACID CHARAC- TERISTICS O.U. Sezerman 1, R. Islamaj 2, E. Alpaydin 2 1 Laborotory of Computational Biology, Sabancı University, Istanbul, Turkey. 2 Computer Engineering

More information

Neural Networks for Machine Learning. Lecture 13a The ups and downs of backpropagation

Neural Networks for Machine Learning. Lecture 13a The ups and downs of backpropagation Neural Networks for Machine Learning Lecture 13a The ups and downs of backpropagation Geoffrey Hinton Nitish Srivastava, Kevin Swersky Tijmen Tieleman Abdel-rahman Mohamed A brief history of backpropagation

More information

Electroencephalography Analysis Using Neural Network and Support Vector Machine during Sleep

Electroencephalography Analysis Using Neural Network and Support Vector Machine during Sleep Engineering, 23, 5, 88-92 doi:.4236/eng.23.55b8 Published Online May 23 (http://www.scirp.org/journal/eng) Electroencephalography Analysis Using Neural Network and Support Vector Machine during Sleep JeeEun

More information

TRAINING A LIMITED-INTERCONNECT, SYNTHETIC NEURAL IC

TRAINING A LIMITED-INTERCONNECT, SYNTHETIC NEURAL IC 777 TRAINING A LIMITED-INTERCONNECT, SYNTHETIC NEURAL IC M.R. Walker. S. Haghighi. A. Afghan. and L.A. Akers Center for Solid State Electronics Research Arizona State University Tempe. AZ 85287-6206 mwalker@enuxha.eas.asu.edu

More information

Title : Analog Circuit for Sound Localization Applications

Title : Analog Circuit for Sound Localization Applications Title : Analog Circuit for Sound Localization Applications Author s Name : Saurabh Kumar Tiwary Brett Diamond Andrea Okerholm Contact Author : Saurabh Kumar Tiwary A-51 Amberson Plaza 5030 Center Avenue

More information

Binary Adders: Half Adders and Full Adders

Binary Adders: Half Adders and Full Adders Binary Adders: Half Adders and Full Adders In this set of slides, we present the two basic types of adders: 1. Half adders, and 2. Full adders. Each type of adder functions to add two binary bits. In order

More information

Simplified Machine Learning for CUDA. Umar Arshad @arshad_umar Arrayfire @arrayfire

Simplified Machine Learning for CUDA. Umar Arshad @arshad_umar Arrayfire @arrayfire Simplified Machine Learning for CUDA Umar Arshad @arshad_umar Arrayfire @arrayfire ArrayFire CUDA and OpenCL experts since 2007 Headquartered in Atlanta, GA In search for the best and the brightest Expert

More information

Magnetometer Realignment: Theory and Implementation

Magnetometer Realignment: Theory and Implementation Magnetometer Realignment: heory and Implementation William Premerlani, Octoer 16, 011 Prolem A magnetometer that is separately mounted from its IMU partner needs to e carefully aligned with the IMU in

More information

Neural network software tool development: exploring programming language options

Neural network software tool development: exploring programming language options INEB- PSI Technical Report 2006-1 Neural network software tool development: exploring programming language options Alexandra Oliveira aao@fe.up.pt Supervisor: Professor Joaquim Marques de Sá June 2006

More information

MIT Sloan School of Management

MIT Sloan School of Management MIT Sloan School of Management Working Paper 4347-02 February 2002 USE OF RECURRENT NEURAL NETWORKS FOR STRATEGIC DATA MINING OF SALES J. Shanmugasundarum, M.V. Nagendra-Prasad, S. Vadhavkar, A. Gupta

More information

Simple and efficient online algorithms for real world applications

Simple and efficient online algorithms for real world applications Simple and efficient online algorithms for real world applications Università degli Studi di Milano Milano, Italy Talk @ Centro de Visión por Computador Something about me PhD in Robotics at LIRA-Lab,

More information

Sanjeev Kumar. contribute

Sanjeev Kumar. contribute RESEARCH ISSUES IN DATAA MINING Sanjeev Kumar I.A.S.R.I., Library Avenue, Pusa, New Delhi-110012 sanjeevk@iasri.res.in 1. Introduction The field of data mining and knowledgee discovery is emerging as a

More information

University of Cambridge Engineering Part IIB Module 4F10: Statistical Pattern Processing Handout 8: Multi-Layer Perceptrons

University of Cambridge Engineering Part IIB Module 4F10: Statistical Pattern Processing Handout 8: Multi-Layer Perceptrons University of Cambridge Engineering Part IIB Module 4F0: Statistical Pattern Processing Handout 8: Multi-Layer Perceptrons x y (x) Inputs x 2 y (x) 2 Outputs x d First layer Second Output layer layer y

More information

Customer Relationship Management by Semi-Supervised Learning

Customer Relationship Management by Semi-Supervised Learning Middle-East Journal of Scientific Research 16 (5): 614-620, 2013 ISSN 1990-9233 IDOSI Pulications, 2013 DOI: 10.5829/idosi.mejsr.2013.16.05.930 Customer Relationship Management y Semi-Supervised Learning

More information

Stock Prediction using Artificial Neural Networks

Stock Prediction using Artificial Neural Networks Stock Prediction using Artificial Neural Networks Abhishek Kar (Y8021), Dept. of Computer Science and Engineering, IIT Kanpur Abstract In this work we present an Artificial Neural Network approach to predict

More information

Implementation of Neural Networks with Theano. http://deeplearning.net/tutorial/

Implementation of Neural Networks with Theano. http://deeplearning.net/tutorial/ Implementation of Neural Networks with Theano http://deeplearning.net/tutorial/ Feed Forward Neural Network (MLP) Hidden Layer Object Hidden Layer Object Hidden Layer Object Logistic Regression Object

More information

A New 128-bit Key Stream Cipher LEX

A New 128-bit Key Stream Cipher LEX A New 128-it Key Stream Cipher LEX Alex Biryukov Katholieke Universiteit Leuven, Dept. ESAT/SCD-COSIC, Kasteelpark Arenerg 10, B 3001 Heverlee, Belgium http://www.esat.kuleuven.ac.e/~airyuko/ Astract.

More information

The Backpropagation Algorithm

The Backpropagation Algorithm 7 The Backpropagation Algorithm 7. Learning as gradient descent We saw in the last chapter that multilayered networks are capable of computing a wider range of Boolean functions than networks with a single

More information

Impact of Feature Selection on the Performance of Wireless Intrusion Detection Systems

Impact of Feature Selection on the Performance of Wireless Intrusion Detection Systems 2009 International Conference on Computer Engineering and Applications IPCSIT vol.2 (2011) (2011) IACSIT Press, Singapore Impact of Feature Selection on the Performance of ireless Intrusion Detection Systems

More information

NEURAL NETWORKS FOR SENSOR VALIDATION AND PLANT MONITORING. B.R. Upadhyaya, E. Eryurek and G. Mathai

NEURAL NETWORKS FOR SENSOR VALIDATION AND PLANT MONITORING. B.R. Upadhyaya, E. Eryurek and G. Mathai NEURAL NETWORKS FOR SENSOR VALIDATION AND PLANT MONITORING B.R. Upadhyaya, E. Eryurek and G. Mathai The University of Tennessee Knoxville, Tennessee 37996-2300, U.S.A. OONF-900804 35 DE93 002127 ABSTRACT

More information

Anupam Tarsauliya Shoureya Kant Rahul Kala Researcher Researcher Researcher IIITM IIITM IIITM Gwalior Gwalior Gwalior

Anupam Tarsauliya Shoureya Kant Rahul Kala Researcher Researcher Researcher IIITM IIITM IIITM Gwalior Gwalior Gwalior Analysis of Artificial Neural Network for Financial Time Series Forecasting Anupam Tarsauliya Shoureya Kant Rahul Kala Researcher Researcher Researcher IIITM IIITM IIITM Gwalior Gwalior Gwalior Ritu Tiwari

More information

DLP Driven, Learning, Optical Neural Networks

DLP Driven, Learning, Optical Neural Networks 01001000100000110000001000001100 010010001000 DLP Driven, Learning, Optical Neural Networks Emmett Redd & A. Steven Younger Associate Professor SPRP506 Missouri State University EmmettRedd@MissouriState.edu

More information

Sequential Logic: Clocks, Registers, etc.

Sequential Logic: Clocks, Registers, etc. ENEE 245: igital Circuits & Systems Lab Lab 2 : Clocks, Registers, etc. ENEE 245: igital Circuits and Systems Laboratory Lab 2 Objectives The objectives of this laboratory are the following: To design

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Deep Learning Barnabás Póczos & Aarti Singh Credits Many of the pictures, results, and other materials are taken from: Ruslan Salakhutdinov Joshua Bengio Geoffrey

More information

School communication at its best!

School communication at its best! School communication at its est! Do you still use 19th century communication methods? School organization and communication Schools don t communicate in an up-to-date way The demands on oth schools and

More information

Supervised Learning (Big Data Analytics)

Supervised Learning (Big Data Analytics) Supervised Learning (Big Data Analytics) Vibhav Gogate Department of Computer Science The University of Texas at Dallas Practical advice Goal of Big Data Analytics Uncover patterns in Data. Can be used

More information

Scalable Intrusion Detection with Recurrent Neural Networks

Scalable Intrusion Detection with Recurrent Neural Networks Scalable Intrusion Detection with Recurrent Neural Networks Longy O. Anyanwu, M.S.; Ed.D. Dept. of Math and Computer Sc. Fort Hays State University Kansas, USA Email: loanyanwu@fhsu.edu Jared Keengwe Ph.D.

More information

DCMS DC MOTOR SYSTEM User Manual

DCMS DC MOTOR SYSTEM User Manual DCMS DC MOTOR SYSTEM User Manual release 1.3 March 3, 2011 Disclaimer The developers of the DC Motor System (hardware and software) have used their best efforts in the development. The developers make

More information

Data Mining Techniques Chapter 7: Artificial Neural Networks

Data Mining Techniques Chapter 7: Artificial Neural Networks Data Mining Techniques Chapter 7: Artificial Neural Networks Artificial Neural Networks.................................................. 2 Neural network example...................................................

More information

A Look Into the World of Reddit with Neural Networks

A Look Into the World of Reddit with Neural Networks A Look Into the World of Reddit with Neural Networks Jason Ting Institute of Computational and Mathematical Engineering Stanford University Stanford, CA 9435 jmting@stanford.edu Abstract Creating, placing,

More information

Stock Market Prediction System with Modular Neural Networks

Stock Market Prediction System with Modular Neural Networks Stock Market Prediction System with Modular Neural Networks Takashi Kimoto and Kazuo Asakawa Computer-based Systems Laboratory FUJTSU LABORATORES LTD., KAWASAK 1015 Kamikodanaka, Nakahara-Ku, Kawasaki

More information

Appendix 4 Simulation software for neuronal network models

Appendix 4 Simulation software for neuronal network models Appendix 4 Simulation software for neuronal network models D.1 Introduction This Appendix describes the Matlab software that has been made available with Cerebral Cortex: Principles of Operation (Rolls

More information

Spark: Cluster Computing with Working Sets

Spark: Cluster Computing with Working Sets Spark: Cluster Computing with Working Sets Outline Why? Mesos Resilient Distributed Dataset Spark & Scala Examples Uses Why? MapReduce deficiencies: Standard Dataflows are Acyclic Prevents Iterative Jobs

More information

Smart Home System Design based on Artificial Neural Networks

Smart Home System Design based on Artificial Neural Networks Smart Home System Design based on Artificial Neural Networks Amit Badlani, Surekha Bhanot Abstract Management and security of electric power plays a key role in economy and sustainable development The

More information

Systematicity in sentence processing with a recursive self-organizing neural network

Systematicity in sentence processing with a recursive self-organizing neural network Systematicity in sentence processing with a recursive self-organizing neural network Igor Farkaš 1 and Matthew W. Crocker 2 1- Dept. of Aplied Informatics, Comenius University Mlynská dolina, 842 48 Bratislava,

More information

129: Artificial Neural Networks. Ajith Abraham Oklahoma State University, Stillwater, OK, USA 1 INTRODUCTION TO ARTIFICIAL NEURAL NETWORKS

129: Artificial Neural Networks. Ajith Abraham Oklahoma State University, Stillwater, OK, USA 1 INTRODUCTION TO ARTIFICIAL NEURAL NETWORKS 129: Artificial Neural Networks Ajith Abraham Oklahoma State University, Stillwater, OK, USA 1 Introduction to Artificial Neural Networks 901 2 Neural Network Architectures 902 3 Neural Network Learning

More information

A bachelor of science degree in electrical engineering with a cumulative undergraduate GPA of at least 3.0 on a 4.0 scale

A bachelor of science degree in electrical engineering with a cumulative undergraduate GPA of at least 3.0 on a 4.0 scale What is the University of Florida EDGE Program? EDGE enables engineering professional, military members, and students worldwide to participate in courses, certificates, and degree programs from the UF

More information

Models of Cortical Maps II

Models of Cortical Maps II CN510: Principles and Methods of Cognitive and Neural Modeling Models of Cortical Maps II Lecture 19 Instructor: Anatoli Gorchetchnikov dy dt The Network of Grossberg (1976) Ay B y f (

More information

Temporal Difference Learning in the Tetris Game

Temporal Difference Learning in the Tetris Game Temporal Difference Learning in the Tetris Game Hans Pirnay, Slava Arabagi February 6, 2009 1 Introduction Learning to play the game Tetris has been a common challenge on a few past machine learning competitions.

More information