NEURAL NETWORK FUNDAMENTALS WITH GRAPHS, ALGORITHMS, AND APPLICATIONS



Similar documents
NEURAL NETWORKS A Comprehensive Foundation

Artificial Neural Networks and Support Vector Machines. CS 486/686: Introduction to Artificial Intelligence

Recurrent Neural Networks

Artificial neural networks

Introduction to Artificial Neural Networks

6.2.8 Neural networks for data mining

Feed-Forward mapping networks KAIST 바이오및뇌공학과 정재승

Introduction to Machine Learning and Data Mining. Prof. Dr. Igor Trajkovski

Software Project Management (Second Edition)

EFFICIENT DATA PRE-PROCESSING FOR DATA MINING

Neural Networks algorithms and applications

Method of Combining the Degrees of Similarity in Handwritten Signature Authentication Using Neural Networks

Data Mining Using Neural Networks: A Guide for Statisticians

Numerical Methods for Engineers

SEMINAR OUTLINE. Introduction to Data Mining Using Artificial Neural Networks. Definitions of Neural Networks. Definitions of Neural Networks

129: Artificial Neural Networks. Ajith Abraham Oklahoma State University, Stillwater, OK, USA 1 INTRODUCTION TO ARTIFICIAL NEURAL NETWORKS

DEC Networks and Architectures

Building VPNs. Nam-Kee Tan. With IPSec and MPLS. McGraw-Hill CCIE #4307 S&

Lecture 6. Artificial Neural Networks

Software and Hardware Solutions for Accurate Data and Profitable Operations. Miguel J. Donald J. Chmielewski Contributor. DuyQuang Nguyen Tanth

Neural Networks and Support Vector Machines

Chapter 4: Artificial Neural Networks

Neural Networks in Data Mining

QUANTITATIVE METHODS. for Decision Makers. Mik Wisniewski. Fifth Edition. FT Prentice Hall

Introduction to Machine Learning Using Python. Vikram Kamath

Neural Network and Its Application in IR

An Introduction to Artificial Neural Networks (ANN) - Methods, Abstraction, and Usage

Practical Applications of DATA MINING. Sang C Suh Texas A&M University Commerce JONES & BARTLETT LEARNING

A Simple Feature Extraction Technique of a Pattern By Hopfield Network

Computer Organization

How To Use Neural Networks In Data Mining

Univariate and Multivariate Methods PEARSON. Addison Wesley

Self Organizing Maps: Fundamentals

Neural Network Design in Cloud Computing

Biological Neurons and Neural Networks, Artificial Neurons

A Time Series ANN Approach for Weather Forecasting

PARALLEL PROGRAMMING

Preface. C++ Neural Networks and Fuzzy Logic:Preface. Table of Contents

NONLINEAR TIME SERIES ANALYSIS

An Introduction to Neural Networks

Compensating the Sales Force

Principles of Data Mining by Hand&Mannila&Smyth

AN APPLICATION OF TIME SERIES ANALYSIS FOR WEATHER FORECASTING

How to Implement Lean Manufacturing

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION

Machine Learning Introduction

Comparison of Supervised and Unsupervised Learning Classifiers for Travel Recommendations

CITY UNIVERSITY OF HONG KONG 香 港 城 市 大 學. Self-Organizing Map: Visualization and Data Handling 自 組 織 神 經 網 絡 : 可 視 化 和 數 據 處 理

Neural Networks: a replacement for Gaussian Processes?

An Introduction to Object-Oriented Programming with

Pricing and calibration in local volatility models via fast quantization

Computer Science MS Course Descriptions

ENTERPRISE RESOURCE PLANNING

Mining. Practical. Data. Monte F. Hancock, Jr. Chief Scientist, Celestech, Inc. CRC Press. Taylor & Francis Group

ADVANCES IN KNOWLEDGE DISCOVERY IN DATABASES

Data Mining. Concepts, Models, Methods, and Algorithms. 2nd Edition

Learning outcomes. Knowledge and understanding. Competence and skills

Data Mining Techniques Chapter 7: Artificial Neural Networks

Machine Learning. CUNY Graduate Center, Spring Professor Liang Huang.

Data Mining and Neural Networks in Stata

Chapter 12 Discovering New Knowledge Data Mining

An Artificial Neural Networks-Based on-line Monitoring Odor Sensing System

VISUALIZING DATA POWER VIEW. with MICROSOFT. Brian Larson. Mark Davis Dan English Paui Purington. Mc Grauu. Sydney Toronto

TRAINING A LIMITED-INTERCONNECT, SYNTHETIC NEURAL IC

The Process. Improvement. Handbook. A Blueprint for Managing Change and. Increasing Organizational Performance. Tristan Boutros.

ARTIFICIAL INTELLIGENCE METHODS IN EARLY MANUFACTURING TIME ESTIMATION

Neural network software tool development: exploring programming language options

Learning is a very general term denoting the way in which agents:

ADVANCED COMPUTER ARCHITECTURE: Parallelism, Scalability, Programmability

Role of Neural network in data mining

Comparison of Supervised and Unsupervised Learning Algorithms for Pattern Classification

Chapter ML:XI (continued)

MCSE Windows 98 Study Guide

Machine Learning: Overview

Learning to Process Natural Language in Big Data Environment

NEUROMATHEMATICS: DEVELOPMENT TENDENCIES. 1. Which tasks are adequate of neurocomputers?

Machine Learning and Data Mining -

Feedforward Neural Networks and Backpropagation

Data Mining: Concepts and Techniques. Jiawei Han. Micheline Kamber. Simon Fräser University К MORGAN KAUFMANN PUBLISHERS. AN IMPRINT OF Elsevier

CRYPTOGRAPHY AND NETWORK SECURITY

SUCCESSFUL PREDICTION OF HORSE RACING RESULTS USING A NEURAL NETWORK

CSC384 Intro to Artificial Intelligence

Study Guide. Professional vsphere 4. VCP VMware Certified. (ExamVCP4IO) Robert Schmidt. IVIC GratAf Hill

Distinguished Professor George Washington University. Graw Hill

ARTIFICIAL NEURAL NETWORKS FOR DATA MINING

SELECTING NEURAL NETWORK ARCHITECTURE FOR INVESTMENT PROFITABILITY PREDICTIONS

Detection. Perspective. Network Anomaly. Bhattacharyya. Jugal. A Machine Learning »C) Dhruba Kumar. Kumar KaKta. CRC Press J Taylor & Francis Croup

Appendix 4 Simulation software for neuronal network models

MANAGING QUEUE STABILITY USING ART2 IN ACTIVE QUEUE MANAGEMENT FOR CONGESTION CONTROL

Oracle Backup & Recovery

Fundamentals of THERMAL-FLUID SCIENCES

Load balancing in a heterogeneous computer system by self-organizing Kohonen network

The Research of Data Mining Based on Neural Networks

3 An Illustrative Example

ABSTRACT The World MINING R. Vasudevan. Trichy. Page 9. usage mining. basic. processing. Web usage mining. Web. useful information

Transcription:

NEURAL NETWORK FUNDAMENTALS WITH GRAPHS, ALGORITHMS, AND APPLICATIONS N. K. Bose HRB-Systems Professor of Electrical Engineering The Pennsylvania State University, University Park P. Liang Associate Professor of Electrical Engineering University of California, Riverside McGraw-Hill, Inc. New York St. Louis San Francisco Auckland Bogota Caracas Lisbon London Madrid Mexico City Milan Montreal New Delhi San Juan Singapore Sydney Tokyo Toronto

CONTENTS List of Figures List of Tables Preface List of Acronyms Glossary of Notations xiv xix xxi xxix xxxii I Fundamentals 1 Basics of Neuroscience and Artificial Neuron Models 3 1.1 The Brain as a Neural Network 5 1.2 Basic Properties of Neurons 9-1:2.1 Structure of a Neuron 9 1.2.2 Dendritic Tree 10 1.2.3 Action Potential and Its Propagation 11 1.2.4 Synapses 14 1.2.5 Connection Patterns between Neurons 18 1.2.6 An Example: A Motion Detection Neuron 19 1.3 Neuron Models 21 1.3.1 McCulloch-Pitts Model 21 1.3.2 Neuron Models with Continuous Transfer Characteristics 24 1.3.3 Other Neuron Models 25 1.4 Conclusions and Suggestions 28 Problems 31 2 Graphs 34 2.1 Terminology and Preliminaries 35 2.2 Special Types of Graphs 38 2.3 Directed Graphs 41 ix

X CONTENTS 2.4 Matrix Representation of Graphs 44 2.4.1 Adjacency Matrix 44 2.4.2 Interconnection Matrix 44 2.5 Topological Invariants 45 2.5.1 Euler and Schlaefli Invariants 46 2.5.2 Genus 47 2.5.3 Thickness 49 2.5.4 Some Other Topological Invariants 51 2.6 Voronoi Diagrams and Deiaunay Tessellation 53 2.7 Conclusions and Suggestions 55 Problems 57 3 Algorithms 61 3.1 Computational Complexity: P- and NP-Complete Problems 62 3.2 Shortest-Path and Max-Flow Min-Cut Problems 66 3.2.1 Dijkstra's Shortest-Path Algorithm 66 3.2.2 Max-Flow Min-Cut Algorithm 68 3.3 Interconnection and Routing Algorithms 77 3.3.1 Problem Formulation 77 3.3.2 Minimal Spanning Tree (MST) Algorithms 79 3.3.3 Minimal Fermat Tree (MFT) Problem 82 3.3.4 Traveling Salesperson (TS) Problem 82 3.3.5 Steiner Minimal Tree (SMT) 83 3.4 Placement and Partitioning 92 3.4.1 Placement 92 3.4.2 Partitioning 96 3.5 Parallel Computation 97 3.6 Associative Memory 99 3.6.1 The Linear Associator: Solution by Hebbian Rule 100 3.6.2 The Linear Associator: Solution by Generalized Inverse 101 3.6.3 Implementation of Associative Memory 102 3.7 Conclusions 106 Problems 108 II Feedforward Networks 4 Perceptrons and the LMS Algorithm 119 4.1 Rosenblatt's Perceptron 120 4.1.1 Definitions 122 4.1.2 Linear Separability of Training Patterns 124 4.1.3 Perceptron Learning Algorithms 129 4.1.4 Derivation of the Perceptron Algorithm as Gradient Descent 134 4.1.5 The Perceptron Convergence Theorem 136 4.2 The Widrow-Hoff LMS Algorithm 138 4.3 Order of a Predicate and a Perceptron 142 4.4 Conclusions and Suggestions 147 Problems 148

CONTENTS XI 5 Multilayer Networks 155 5.1 Exact and Approximate Representation Using Feedforward Networks 156 5.1.1 Exact Representation: Kolmogorov's Theorem and Its Consequences 156 5.1.2 Approximate Representations 159 5.2 Fixed Multilayer Feedforward Network Training by Backpropagation 162 5.2.1 Implementation Considerations for Backpropagation 176 5.2.2 Variants of BPA 177 5.2.3 Temporal Signal Recognition and Prediction 179 5.3 Structural Training of Multilayer Feedforward Networks 181 5.3.1 Algorithm for Design Based on VoD 183 5.3.2 Robustness and Size Issues 189 5.4 Unsupervised and Reinforcement Learning 192 5.4.1 Principal Component Analysis Networks 193 5.4.2 Self-Organization in a Perceptual Network 197 5.4.3 Reinforcement Learning 201 5.5 The Probabilistic Neural Network 204 5.6 Conclusions and Suggestions 209 Problems 212 6 Complexity of Learning Using Feedforward Networks 219 6.1 Learnability in ANN 219 6.1.1 The Problem of Loading 221 6.1.2 Using an Appropriate Network to Get Around Intractability 228 6.2 Generalizability of Learning 231 6.2.1 VC Dimension and Generalization 232 \ "6.2.2 Sufficient Conditions for Valid Generalization ' in Feedforward Networks 237 6.2.3 Necessary Conditions for Valid Generalization in Feedforward Networks 238 6.2.4 Discussions and Ways to Improve Generalization 240 6.3 Space Complexity of Feedforward Networks 245 6.3.1 Order of a Function and the Complexity of a Network 247 6.3.2 High Connectivity in Analog Neural Computations 248 6.4 Summary and Discussion 250 Problems 252 7 Adaptive-Structure Networks 254 7.1 Growth Algorithms 255 7.1.1 The Upstart Algorithm 257 7.1.2 Learning by Divide and Conquer 259 7.1.3 Other Growth Algorithms 265

XH CONTENTS 7.2 Networks with Nonlinear Synapses and Nonlinear Synaptic Contacts 270 7.2.1 Quasi-Polynomial Synapses and Product Synaptic Contacts 273 7.2.2 Generalization of Learning and Hardware Considerations 276 7.3 Conclusions and Suggestions 278 Problems 281 III Recurrent Networks 8 Symmetric and Asymmetric Recurrent Networks 287 8.1 Symmetric Hopfield Networks and Associative Memory 289 8.1.1 Convergence Proofs 292 8.1.2 Computation in a Network and Minimum Cuts in a Graph 294 8.1.3 Capacity and Spurious Memory 300 8.1.4 Correlated Patterns 304 8.1.5 Hopfield Networks with Variations in the Connection Weights 306 8.1.6 Bidirectional Associative Memory 307 8.2 Symmetric Networks with Analog Units 310 8.2.1 Analog Hopfield Networks 310 8.2.2 Convergence Proof 314 8.2.3 Relation between Stable States of Discrete and Analog Hopfield Networks 315 8.2.4 Cellular Neural Networks 316 8.3 Seeking the Global Minimum: Simulated Annealing 318 8.3.1 Simulated Annealing in Optimization 319 8.3.2 Stochastic Networks: Applying Simulated Annealing to Hopfield Networks 323 8.4 A Learning Algorithm for the Boltzmann Machine 324 8.4.1 Learning the Underlying Structure of an Environment 324 8.4.2 The Learning Procedure 328 8.4.3 Mean Field Theory and the Deterministic Boltzmann Machine 330 8.5 Asymmetric Recurrent Networks 331 8.5.1 Phase Transition from Stationary to Chaotic 331 8.5.2 Spatial and Temporal Patterns 332 8.5.3 Learning in Asymmetric Networks: Recurrent Backpropagation 335 8.6 Summary and Discussion 340 Problems 341 9 Competitive Learning and Self-Organizing Networks 343 9.1 Unsupervised Competitive Learning 344 9.1.1 Two Phases of Competitive Learning 346

CONTENTS XUl 9.2 9.3 9.4 9.5 9.1.2 Using a Competitive Learning Network for Associative Memory Adaptive Resonant Networks 9.2.1 The ART1 Clustering Algorithm 9.2.2 The ART1 Network Self-Organizing Feature Maps 9.3.1 The Kohonen Map 9.3.2 Analysis of Kohonen Maps 9.3.3 Adaptive and Learning Vector Quantization 9.3.4 Two-Dimensional Topographic Maps 9.3.5 A Multilayer Self-Organizing Feature Map Hybrid Learning 9.4.1 Counterpropagation Network 9.4.2 Regularizing Networks and Radial Basis Functions Summary and Discussion Problems 349 352 352 355 360 361 366 375 381 385 391 391 395 398 401 IV Applications of Neural Networks 10 Neural Network Approaches to Solving Hard Problems 407 10.1 The Traveling Salesperson Problem 409 10.2 Multitarget Tracking 412 10.3 Time Series Prediction 415 10.4 Talking Network and Phonetic Typewriter 417 10.4.1 Speech Generation 417 10.4.2 Speech Recognition 418 10.5 Autonomous Vehicle Navigation 422 10.6 Handwritten Digit Recognition 425 10.7 Image Compression by a Multilayer Feedforward { - - Structure Trained through Backpropagation 430 I 10.8 Character Retrieval Using the Discrete Hopfield Network 435 10.9 Visual Processing Networks 437 10.10 Conclusion and Discussion 443 References 447 Appendix A Basis of Gradient-Based Optimization Methods 463 A.I The Gradient Descent Method 464 A.2 Newton's Method 467 A.3 The Conjugate Gradient Method 468 A.4 Constrained Optimization 469 Bibliography 470 Index 471