Online Evolution of Deep Convolutional Network for Vision-Based Reinforcement Learning
|
|
|
- Jemima Franklin
- 10 years ago
- Views:
Transcription
1 Online Evolution of Deep Convolutional Network for Vision-Based Reinforcement Learning Jan Koutník, Jürgen Schmidhuber, and Faustino Gomez IDSIA, USI-SUPSI Galleria 2 Manno-Lugano, CH 6928 {hkou,juergen,tino}@idsia.ch Abstract. Dealing with high-dimensional input spaces, like visual input, is a challenging task for reinforcement learning (RL). Neuroevolution (NE), used for continuous RL problems, has to either reduce the problem dimensionality by (1) compressing the representation of the neural network controllers or (2) employing a pre-processor (compressor) that transforms the high-dimensional raw inputs into low-dimensional features. In this paper we extend the approach in [16]. The Max-Pooling Convolutional Neural Network (MPCNN) compressor is evolved online, maximizing the distances between normalized feature vectors computed from the images collected by the recurrent neural network (RNN) controllers during their evaluation in the environment. These two interleaved evolutionary searches are used to find MPCNN compressors and RNN controllers that drive a race car in the TORCS racing simulator using only visual input. Keywords: deep learning; neuroevolution; vision-based TORCS; reinforcement learning; computer games 1 Introduction Most approaches to scaling neuroevolution to tasks that require large networks, such as those processing video input, have focused on indirect encodings where relatively small neural network descriptions are transformed via a complex mapping into networks of arbitrary size [4, 7, 10, 13, 15, 23]. A different approach to dealing with high-dimensional input which has been studied in the context of single-agent RL (i.e. TD [25], policy gradients [24], etc.), is to combine action learning with an unsupervised learning (UL) preprocessor or compressor which provides a lower-dimensional feature vector that the agent receives as input instead of the raw observation [5, 8, 12, 17, 19 21]. The UL compressor is trained on the high-dimensional observations generated by the learning agent s actions, that the agent then uses as a state representation to learn a value function. In [3], the first combination of UL and evolutionary reinforcement learning was introduced where a single UL module is trained on data generated by
2 2 Koutník et al. entire population as it interacts with environment (which would normally be discarded) to build a representation that allows evolution to search in a relatively low-dimensional feature space. This approach attacks the problem of high-dimensionality from the opposite direction compared to indirect encoding: instead of compressing large networks into short genomes, the inputs are compressed so that smaller networks can be used. In [16], we scaled up this Unsupervised Learning Evolutionary Reinforcement Learning (UL-ERL) approach to the challenging reinforcement learning problem of driving a car in the TORCS simulator using vision from the driver s perspective as input. The high-dimensional images were compressed down to just three features by a Max-Pooling Convolutional Neural Network (MPCNN; [2,22]) that allowed an extremely small (only 33 weights) recurrent neural network controller to be evolved to drive the car successfully. The MPCNN was itself trained separately off-line using images collected previously while driving the car manually around the training track. In this paper, the feature learning and control learning are interleaved as in [3]. Both the MPCNN, acting as the sensory preprocessor (compressor), and the recurrent neural network controllers are evolved simultaneously in separate populations, with the images used to train the MPCNNs being taken from the driving trials of the evolving controllers rather than being collected manually a priori. The next section describes the MPCNN architecture that is used to compress the high-dimensional vision inputs. Section 3 covers our method the UL-ERL framework applied to visual TORCS race car driving domain. Section 4 presents the experiments in the TORCS race car driving, which are discussed in section 5. 2 Max-Pooling Convolutional Neural Networks Convolution Neural Networks [6, 18] are deep hierarchical networks that have recently become the state-of-the-art in image classification due to the advent of fast implementations on graphics card multiprocessors (GPUs) [1]. CNNs have two parts: (1) a deep feature detector consisting of alternating convolutional and down-sampling layers, and (2) a classifier that receives the output of the final layer of the feature detector. Each convolutional layer l, has a bank of m l n l filters, F l, where m l is the number input maps (images), I l, to the layer, and n l is the number of output maps (inputs to the next layer). The i-th output map is computed by: I l+1 i m l = σ Ij l Fij l, i = 1..n l, l = 1, 3, 5..., j=1 where is the convolution operator, Fij l is the i-th filter for the j-th map and σ is a non-linear squashing function (e.g. sigmoid). Note that l is always odd because of the subsampling layers between each of the convolutional layers.
3 Online Evolution of Deep CNN for Vision-Based RL 3 The downsampling layers reduce resolution of each map. Each map is partitioned into non-overlapping blocks and a value from each block is used in the output map. The max-pooling operation used here subsamples the map by simply taking the maximum value in the block as the output, making these networks Max-Pooling CNNs [2, 22]. The stacked alternating layers transform the input into progressively lower dimensional abstract representations that are then classified, typically using a standard feed-forward neural network that has an output neuron for each class. The entire network is usually trained using a large training set of class-labeled images via backpropagation [18]. Figure 2 illustrates the particular MPCNN architecture used in this paper. It should be clear from the figure that each convolution layer is just a matrix product where the matrix entries are filters and convolution is used instead of multiplication. 3 Method: Online MPC-RNN The Online MPC-RNN method for evolving the controller and the compressor at the same time is overviewed in figure 1. The controller is evolved in the usual way (i.e. neuroevolution; [26]), but instead of accessing the observations directly, it receives feature vectors of much lower dimensionality provided by the unsupervised compressor, which is also evolved. The compressor is trained on observations (images) generated by the actions taken by candidate controllers as they interact with the environment. Here, unlike in [16] where the compressor was pre-trained off-line, the loop is closed by simultaneously evolving the compressor using the latest observations (images) of the environment that are provoked by the actions taken by the evolving controllers. For the compressor, a deep max-pooling convolutional neural network (MPCNN) is used. These networks are normally trained to perform image classification through supervised learning using enormous training sets. Of course, this requires a priori knowledge of what constitutes a class. In a general RL setting, we may not know how the space of images should be partitioned. For example, how many classes should there be for images perceived from the first-person perspective while driving the TORCS car? We could study the domain in detail to construct a training set that could then be used to train the MPCNN with backpropagation. However, we do not want the learning system to rely on taskspecific domain knowledge, so, instead, the MPCNN is evolved without using a labeled training set. A set of k images is collected from the environment, and then MPCNNs are evolved to maximize the fitness: where D is a list of all Euclidean distances, f k = min(d) + mean(d), (1) d i,j = f i f j, i > j,
4 4 Koutnı k et al. Fig. 1. Overview of the online TORCS controller evolution in the online UL-ERL scheme. At each time-step, a raw grayscale (saturation plane) pixel image, taken from the driver s perspective, is transformed to features by the Max-Pooling Convolutional Neural Network (MPCNN) compressor. The features are used by the candidate RNN controller to drive the car by steering, accelerating and braking. Two evolutionary algorithms are interleaved to simultaneously optimize both the controller and compressor. The current implementation uses the CoSyNE algorithm: CoSyNEk for the MPCNN, CoSyNEc for the RNN controllers, using the best MPCNN found by CoSyNEk. between k normalized feature vectors {f1... fk } generated from k images in the training set by the MPCNN encoded in the genome. This fitness function forces the evolving MPCNNs to output feature vectors that are spread out in feature space, so that when the final, evolved MPCNN processes images for the evolving controllers, it will provide enough discriminative power to allow them to take correct actions. 4 Visual TORCS Experiments The goal of the task is to evolve a recurrent neural network controller and MPCNN compressor that can drive the car around a race track. The visual TORCS environment is based on TORCS version The simulator had to be modified to provide images as input to the controllers (a detailed description of modifications is provided in [16]). The most important changes involve decrease of the control frequency from 50 Hz to 5 Hz, and removal of the GO sequence from the beginning of each race.
5 Online Evolution of Deep CNN for Vision-Based RL 5 Fig. 2. Example Max-Pooling Convolutional Neural Network (MPCNN) with 8 layers alternating between convolution (C) and downsampling (MP; using max-pooling). The first layer convolves the input pixel image with a bank of filters producing 10 maps of size 63 63, that are down-sampled to by MP layer 2. Layer 3 convolves each of these 10 maps with a filter, sums the results and passes them through the nonlinear function f, producing 10 maps of pixels each, and so on until the input image is transformed to just 3 features that are passed to the RNN controller, see Figure Setup The MPCNN compressors and RNN controllers are evolved in a coupled system, alternating every four generations between two separate CoSyNE [9] algorithms, denoted CoSyNE k for the MPCNNs, and CoSyNE c for the RNN controllers. The process starts by first evolving the controllers. In each fitness evaluation, the candidate controller is tested in two trials, one on the track shown in figure 1, and one on its mirror image. A trial consists of placing the car at the starting line and driving it for 25 s of simulated time, resulting in a maximum of 125 time-steps at the 5 Hz control frequency. At each control step a raw pixel image (saturation plane only), taken from the driver s perspective is passed through an initially random MPCNN compressor which generates a 3-dimensional feature vector that is fed into a simple recurrent neural network (SRN) with 3 hidden neurons, and 3 output neurons (33 total weights). The first two outputs, o 1, o 2, are averaged, (o 1 + o 2 )/2, to provide the steering signal ( 1 = full left lock, 1 = full right lock), and the third neuron, o 3, controls the brake and throttle ( 1 = full brake, 1 = full throttle). All neurons use sigmoidal activation functions. The fitness of the controllers use by CoSyNE c is computed by: f c = d 3m v max 100c, (2) 5 where d is the distance along the track axis measured from the starting line, v max is maximum speed, m is the cumulative damage, and c is the sum of squares of
6 6 Koutník et al Generation Fig. 3. Evolving MPCNN features. Each plot shows the feature vectors for each of the 40 training images collected at the given generation of CoSyNE k on the unit sphere. Initially (generation 0), the features are clustered together. After just a few generations spread out so that the MPCNN discriminates more clearly between the images. The features stabilize after generation 50. the control signal differences, divided by the number of control variables, 3, and the number simulation control steps, T : c = 1 3T 3 i T [o i (t) o i (t 1)] 2. (3) t The maximum speed component in equation (2) forces the controllers to accelerate and brake efficiently, while the damage component favors controllers that drive safely, and c encourages smoother driving. Fitness scores roughly correspond to the distance traveled along the race track axis. Each individual is evaluated both on the track and its mirror image to prevent the RNN from blindly memorizing the track without using the visual input. The original track starts with a left turn, while the mirrored track starts with a right turn, forcing the network to use the visual input to distinguish between tracks. The final fitness score is the minimum of the two track scores (equation 2). After four generations of CoSyNE c, CoSyNE k starts. First, a population of MPCNNs with 8 layers, alternating between convolution and max-pooling operations (see figure 2 and Table 1) is initialized with random kernel weights uniformly distributed between 1.5 and 1.5). The MPCNNs are then evolved for four generations using 40 randomly selected images from the previous CoSyNE c phase to compute the fitness function f k (equation 1). Each pixel image is processed by a candidate MPCNN by convolving it with each of the 20 filters in layer 1, to produce feature maps, each of which is reduced down to by the first max-pooling layer (2). These 20 features are convolved again by layer 3, max-pooled in layer 4, and so on, until the image is reduced down to just 3 1-dimensional features which are fed to the controller. This architecture has a total of 3583 kernel coefficients (weights). All MPCNN layers used scaled tanh transfer function.
7 Online Evolution of Deep CNN for Vision-Based RL 7 Layer Type m n #maps map size 1 C MP C MP C MP C MP Table 1. MPCNN topology. The table summarizes the MPCNN architecture used; type of a layer, where C is for convolutional, MP for max-pooling, dimensions of the filter bank m and n, number of output maps and their resolution. Fig. 4. Evolved visual features. Each inset shows the image at a particular point on the track and the 3-dimensional feature vector produced by the MPCNN after its evolution finished. At the end of the four generations, the best compressor in the population becomes the compressor for the controllers in the next round of CoSyNE c. CoSyNE k and CoSyNE c continue to alternate until a sufficiently fit controller is found, with 20% of the training images being replaced by new ones selected at random in each iteration of CoSyNE k. Both MPCNN compressors and controllers are directly encoded into real-valued genomes, and the population size of both CoSyNEs was 100, with a mutation rate of Results We report just 2 runs because a single run of 300 generations of both CoSyNE c and CoSyNE k takes almost 80 hours on an 8-core machine 1 (running 8 evaluations in parallel). The fitness, f c, reached at generation 289 for run no. 1 AMD FX core, 16 GB RAM, nvidia GTX-570
8 8 Koutník et al. controller d [m] v max [km/h] olethros bt berniw tita inferno visual RNN [14] MPC-RNN [16] Online MPC-RNN Table 2. Maximum distance, d, in meters and maximum speed, v max, in kilometers per hour achieved by hand-coded controllers that come with TORCS which enjoy access to the state variables (the five upper table entries), a million-weight RNN controller that drives using pre-processed pixel images as input, evolved indirectly in the Fourier domain, MPC-RNN, where the MPCNN is trained offline from manually collected images [16] and the Online MPC-RNN agent that collects the MPCNN training images automatically. 1 and at generation 145 for run no. 2, where a generation refers to one generation of each CoSyNE. Table 2 compares the distance travelled and maximum speed of the best controller with the offline-evolved MPCNN controller [16], a large RNN controller evolved in frequency domain [14], and the hand-coded controllers that come with the TORCS package. The performance of Online MPC-RNN is not as good as its offline variant, but the controllers still approach a fitness of 500, which allows them to complete a lap and continue driving without crashing. The controllers with the pre-trained MPCNN drive slightly better because of possibly two reasons (1) the MPCNN compressor, that improves online does not reach the optimum and (2) as the compressor evolves together with the RNN controller, the weights of the controller have to be updated after each compressor change due to different features that it provides. The hand-coded controllers are much better since they enjoy an access to the car telemetry and features like distances to the track edges. Figure 3 shows the evolution of the MPCNN feature vectors for each of the 40 images in the training set, in one of the two runs. As the features evolve they very quickly move away from each other in the feature space. While simply pushing the feature vectors apart is no guarantee of achieving maximally informative compressed representations, this simple, unsupervised training procedure provides enough discriminative power in practice to get the car safely across the finish line. 5 Discussion The results show that it is not necessary to undertake the complicated procedure of collecting the images manually by driving the car and training the compressor beforehand. The MPCNN compressor, trained online from images gathered is
9 Online Evolution of Deep CNN for Vision-Based RL 9 good enough for a small RNN with 33 weights to be evolved efficiently to solve the driving control task. Another approach would be to combine the controllers and compressors within a single directly-encoded genome, but one can expect this to run even longer. Also, instead of using an MPCNN with fitness f k, the collected images could be used to train autoencoders [11] that would be forced to generated suitable features. The presented framework is more general than just a TORCS driving system. It remains to be seen whether the agents can be plugged into some other environment, collect the images on the fly and train the controllers to perform the desired task. The TORCS control signal is only 3-dimensional, future experiments will apply Online MPC-RNN to higher-dimensional action spaces like the 41-DOF icub humanoid, to perform manipulation tasks using vision. Acknowledgments This research was supported by Swiss National Science Foundation grant #138219: Theory and Practice of Reinforcement Learning 2, and EU FP7 project: NAnoSCaleEngineering for Novel Computation using Evolution (NASCENCE), grant agreement # References 1. D. C. Ciresan, U. Meier, L. M. Gambardella, and J. Schmidhuber. Deep big simple neural nets for handwritten digit recognition. Neural Computation, 22(12): , D. C. Ciresan, U. Meier, J. Masci, L. M. Gambardella, and J. Schmidhuber. Flexible, high performance convolutional neural networks for image classification. In Proceedings of the International Joint Conference on Artificial Intelligence (IJ- CAI), pages , G. Cuccu, M. Luciw, J. Schmidhuber, and F. Gomez. Intrinsically motivated evolutionary search for vision-based reinforcement learning. In Proceedings of the IEEE Conference on Development and Learning, and Epigenetic Robotics, D. B. D Ambrosio and K. O. Stanley. A novel generative encoding for exploiting neural network sensor and output geometry. In Proceedings of the 9th Conference on Genetic and Evolutionary Computation, (GECCO), pages , New York, USA, ACM. 5. F. Fernández and D. Borrajo. Two steps reinforcement learning. International Journal of Intelligent Systems, 23(2): , K. Fukushima. Neocognitron: A self-organizing neural network for a mechanism of pattern recognition unaffected by shift in position. Biological Cybernetics, 36(4): , J. Gauci and K. Stanley. Generating large-scale neural networks through discovering geometric regularities. In Proceedings of the Conference on Genetic and Evolutionary Computation, (GECCO), pages ACM, 2007.
10 10 Koutník et al. 8. L. Gisslén, M. Luciw, V. Graziano, and J. Schmidhuber. Sequential Constant Size Compressors and Reinforcement Learning. In Proceedings of the Fourth Conference on Artificial General Intelligence, F. J. Gomez, J. Schmidhuber, and R. Miikkulainen. Accelerated neural evolution through cooperatively coevolved synapses. Journal of Machine Learning Research, 9(May): , F. Gruau. Cellular encoding of genetic neural networks. Technical Report RR-92-21, Ecole Normale Superieure de Lyon, Institut IMAG, Lyon, France, G. Hinton and R. Salakhutdinov. Reducing the dimensionality of data with neural networks. Science, 313(5786): , S. R. Jodogne and J. H. Piater. Closed-loop learning of visual control policies. Journal of Artificial Intelligence Research, 28: , H. Kitano. Designing neural networks using genetic algorithms with graph generation system. Complex Systems, 4: , J. Koutník, G. Cuccu, J. Schmidhuber, and F. Gomez. Evolving large-scale neural networks for vision-based reinforcement learning. In Proceedings of the Genetic and Evolutionary Computation Conference (GECCO), Amsterdam, J. Koutník, F. Gomez, and J. Schmidhuber. Evolving neural networks in compressed weight space. In Proceedings of the Conference on Genetic and Evolutionary Computation (GECCO), J. Koutník, J. Schmidhuber, and F. Gomez. Evolving deep unsupervised convolutional networks for vision-based reinforcement learning. In Proceedings of the 2014 Genetic and Evolutionary Computation Conference (GECCO). ACM Press, S. Lange and M. Riedmiller. Deep auto-encoder neural networks in reinforcement learning. In International Joint Conference on Neural Networks (IJCNN), Barcelona, Spain, Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11): , R. Legenstein, N. Wilbert, and L. Wiskott. Reinforcement Learning on Slow Features of High-Dimensional Input Streams. PLoS Computational Biology, 6(8), D. Pierce and B. Kuipers. Map learning with uninterpreted sensors and effectors. Artificial Intelligence, 92: , M. Riedmiller, S. Lange, and A. Voigtlaender. Autonomous reinforcement learning on raw visual input data in a real world application. In Proceedings of the International Joint Conference on Neural Networks (IJCNN), pages 1 8, Brisbane, Australia, D. Scherer, A. Müller, and S. Behnke. Evaluation of pooling operations in convolutional architectures for object recognition. In Proceedings of the International Conference on Artificial Neural Networks, ICANN, J. Schmidhuber. Discovering neural nets with low Kolmogorov complexity and high generalization capability. Neural Networks, 10(5): , R. S. Sutton, D. A. McAllester, S. P. Singh, and Y. Mansour. Policy gradient methods for reinforcement learning with function approximation. In Advances in Neural Information Processing Systems 12 (NIPS), pages , G. Tesauro. Practical issues in temporal difference learning. In D. S. Lippman, J. E. Moody, and D. S. Touretzky, editors, Advances in Neural Information Processing Systems 4 (NIPS), pages Morgan Kaufmann, X. Yao. Evolving artificial neural networks. Proceedings of the IEEE, 87(9): , 1999.
Multi-Column Deep Neural Network for Traffic Sign Classification
Multi-Column Deep Neural Network for Traffic Sign Classification Dan Cireşan, Ueli Meier, Jonathan Masci and Jürgen Schmidhuber IDSIA - USI - SUPSI Galleria 2, Manno - Lugano 6928, Switzerland Abstract
Introduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU-10701 Deep Learning Barnabás Póczos & Aarti Singh Credits Many of the pictures, results, and other materials are taken from: Ruslan Salakhutdinov Joshua Bengio Geoffrey
Artificial Neural Networks and Support Vector Machines. CS 486/686: Introduction to Artificial Intelligence
Artificial Neural Networks and Support Vector Machines CS 486/686: Introduction to Artificial Intelligence 1 Outline What is a Neural Network? - Perceptron learners - Multi-layer networks What is a Support
3D Object Recognition using Convolutional Neural Networks with Transfer Learning between Input Channels
3D Object Recognition using Convolutional Neural Networks with Transfer Learning between Input Channels Luís A. Alexandre Department of Informatics and Instituto de Telecomunicações Univ. Beira Interior,
Transfer Learning for Latin and Chinese Characters with Deep Neural Networks
Transfer Learning for Latin and Chinese Characters with Deep Neural Networks Dan C. Cireşan IDSIA USI-SUPSI Manno, Switzerland, 6928 Email: [email protected] Ueli Meier IDSIA USI-SUPSI Manno, Switzerland, 6928
Steven C.H. Hoi School of Information Systems Singapore Management University Email: [email protected]
Steven C.H. Hoi School of Information Systems Singapore Management University Email: [email protected] Introduction http://stevenhoi.org/ Finance Recommender Systems Cyber Security Machine Learning Visual
EFFICIENT DATA PRE-PROCESSING FOR DATA MINING
EFFICIENT DATA PRE-PROCESSING FOR DATA MINING USING NEURAL NETWORKS JothiKumar.R 1, Sivabalan.R.V 2 1 Research scholar, Noorul Islam University, Nagercoil, India Assistant Professor, Adhiparasakthi College
Learning is a very general term denoting the way in which agents:
What is learning? Learning is a very general term denoting the way in which agents: Acquire and organize knowledge (by building, modifying and organizing internal representations of some external reality);
Novelty Detection in image recognition using IRF Neural Networks properties
Novelty Detection in image recognition using IRF Neural Networks properties Philippe Smagghe, Jean-Luc Buessler, Jean-Philippe Urban Université de Haute-Alsace MIPS 4, rue des Frères Lumière, 68093 Mulhouse,
Taking Inverse Graphics Seriously
CSC2535: 2013 Advanced Machine Learning Taking Inverse Graphics Seriously Geoffrey Hinton Department of Computer Science University of Toronto The representation used by the neural nets that work best
Hybrid Evolution of Heterogeneous Neural Networks
Hybrid Evolution of Heterogeneous Neural Networks 01001110 01100101 01110101 01110010 01101111 01101110 01101111 01110110 01100001 00100000 01110011 01101011 01110101 01110000 01101001 01101110 01100001
Flexible, High Performance Convolutional Neural Networks for Image Classification
Proceedings of the Twenty-Second International Joint Conference on Artificial Intelligence Flexible, High Performance Convolutional Neural Networks for Image Classification Dan C. Cireşan, Ueli Meier,
How To Find Out If Novelty Search Is A Good Way To Find A Novel Solution
When Novelty is Not Enough Giuseppe Cuccu and Faustino Gomez IDSIA Galleria 2, 6928 Manno-Lugano, CH Email: {giuse,tino}@idsia.ch, WWW: http://www.idsia.ch/{~giuse,~tino} Abstract. The idea of evolving
Lecture 6. Artificial Neural Networks
Lecture 6 Artificial Neural Networks 1 1 Artificial Neural Networks In this note we provide an overview of the key concepts that have led to the emergence of Artificial Neural Networks as a major paradigm
Handwritten Digit Recognition with a Back-Propagation Network
396 Le Cun, Boser, Denker, Henderson, Howard, Hubbard and Jackel Handwritten Digit Recognition with a Back-Propagation Network Y. Le Cun, B. Boser, J. S. Denker, D. Henderson, R. E. Howard, W. Hubbard,
Neural Networks and Back Propagation Algorithm
Neural Networks and Back Propagation Algorithm Mirza Cilimkovic Institute of Technology Blanchardstown Blanchardstown Road North Dublin 15 Ireland [email protected] Abstract Neural Networks (NN) are important
Image Classification for Dogs and Cats
Image Classification for Dogs and Cats Bang Liu, Yan Liu Department of Electrical and Computer Engineering {bang3,yan10}@ualberta.ca Kai Zhou Department of Computing Science [email protected] Abstract
Comparison of K-means and Backpropagation Data Mining Algorithms
Comparison of K-means and Backpropagation Data Mining Algorithms Nitu Mathuriya, Dr. Ashish Bansal Abstract Data mining has got more and more mature as a field of basic research in computer science and
Efficient online learning of a non-negative sparse autoencoder
and Machine Learning. Bruges (Belgium), 28-30 April 2010, d-side publi., ISBN 2-93030-10-2. Efficient online learning of a non-negative sparse autoencoder Andre Lemme, R. Felix Reinhart and Jochen J. Steil
Sense Making in an IOT World: Sensor Data Analysis with Deep Learning
Sense Making in an IOT World: Sensor Data Analysis with Deep Learning Natalia Vassilieva, PhD Senior Research Manager GTC 2016 Deep learning proof points as of today Vision Speech Text Other Search & information
SUCCESSFUL PREDICTION OF HORSE RACING RESULTS USING A NEURAL NETWORK
SUCCESSFUL PREDICTION OF HORSE RACING RESULTS USING A NEURAL NETWORK N M Allinson and D Merritt 1 Introduction This contribution has two main sections. The first discusses some aspects of multilayer perceptrons,
6.2.8 Neural networks for data mining
6.2.8 Neural networks for data mining Walter Kosters 1 In many application areas neural networks are known to be valuable tools. This also holds for data mining. In this chapter we discuss the use of neural
Data Mining and Neural Networks in Stata
Data Mining and Neural Networks in Stata 2 nd Italian Stata Users Group Meeting Milano, 10 October 2005 Mario Lucchini e Maurizo Pisati Università di Milano-Bicocca [email protected] [email protected]
Analecta Vol. 8, No. 2 ISSN 2064-7964
EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,
Supporting Online Material for
www.sciencemag.org/cgi/content/full/313/5786/504/dc1 Supporting Online Material for Reducing the Dimensionality of Data with Neural Networks G. E. Hinton* and R. R. Salakhutdinov *To whom correspondence
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic
Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets
Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets Macario O. Cordel II and Arnulfo P. Azcarraga College of Computer Studies *Corresponding Author: [email protected]
Chapter 4: Artificial Neural Networks
Chapter 4: Artificial Neural Networks CS 536: Machine Learning Littman (Wu, TA) Administration icml-03: instructional Conference on Machine Learning http://www.cs.rutgers.edu/~mlittman/courses/ml03/icml03/
Recurrent Neural Networks
Recurrent Neural Networks Neural Computation : Lecture 12 John A. Bullinaria, 2015 1. Recurrent Neural Network Architectures 2. State Space Models and Dynamical Systems 3. Backpropagation Through Time
The Role of Size Normalization on the Recognition Rate of Handwritten Numerals
The Role of Size Normalization on the Recognition Rate of Handwritten Numerals Chun Lei He, Ping Zhang, Jianxiong Dong, Ching Y. Suen, Tien D. Bui Centre for Pattern Recognition and Machine Intelligence,
Machine Learning. 01 - Introduction
Machine Learning 01 - Introduction Machine learning course One lecture (Wednesday, 9:30, 346) and one exercise (Monday, 17:15, 203). Oral exam, 20 minutes, 5 credit points. Some basic mathematical knowledge
Introduction to Machine Learning and Data Mining. Prof. Dr. Igor Trajkovski [email protected]
Introduction to Machine Learning and Data Mining Prof. Dr. Igor Trakovski [email protected] Neural Networks 2 Neural Networks Analogy to biological neural systems, the most robust learning systems
Applying Deep Learning to Car Data Logging (CDL) and Driver Assessor (DA) October 22-Oct-15
Applying Deep Learning to Car Data Logging (CDL) and Driver Assessor (DA) October 22-Oct-15 GENIVI is a registered trademark of the GENIVI Alliance in the USA and other countries Copyright GENIVI Alliance
International Journal of Computer Science Trends and Technology (IJCST) Volume 2 Issue 3, May-Jun 2014
RESEARCH ARTICLE OPEN ACCESS A Survey of Data Mining: Concepts with Applications and its Future Scope Dr. Zubair Khan 1, Ashish Kumar 2, Sunny Kumar 3 M.Tech Research Scholar 2. Department of Computer
Selecting Receptive Fields in Deep Networks
Selecting Receptive Fields in Deep Networks Adam Coates Department of Computer Science Stanford University Stanford, CA 94305 [email protected] Andrew Y. Ng Department of Computer Science Stanford
Network Machine Learning Research Group. Intended status: Informational October 19, 2015 Expires: April 21, 2016
Network Machine Learning Research Group S. Jiang Internet-Draft Huawei Technologies Co., Ltd Intended status: Informational October 19, 2015 Expires: April 21, 2016 Abstract Network Machine Learning draft-jiang-nmlrg-network-machine-learning-00
Data, Measurements, Features
Data, Measurements, Features Middle East Technical University Dep. of Computer Engineering 2009 compiled by V. Atalay What do you think of when someone says Data? We might abstract the idea that data are
Load balancing in a heterogeneous computer system by self-organizing Kohonen network
Bull. Nov. Comp. Center, Comp. Science, 25 (2006), 69 74 c 2006 NCC Publisher Load balancing in a heterogeneous computer system by self-organizing Kohonen network Mikhail S. Tarkov, Yakov S. Bezrukov Abstract.
Using Data Mining for Mobile Communication Clustering and Characterization
Using Data Mining for Mobile Communication Clustering and Characterization A. Bascacov *, C. Cernazanu ** and M. Marcu ** * Lasting Software, Timisoara, Romania ** Politehnica University of Timisoara/Computer
Neural Networks for Machine Learning. Lecture 13a The ups and downs of backpropagation
Neural Networks for Machine Learning Lecture 13a The ups and downs of backpropagation Geoffrey Hinton Nitish Srivastava, Kevin Swersky Tijmen Tieleman Abdel-rahman Mohamed A brief history of backpropagation
14.10.2014. Overview. Swarms in nature. Fish, birds, ants, termites, Introduction to swarm intelligence principles Particle Swarm Optimization (PSO)
Overview Kyrre Glette kyrrehg@ifi INF3490 Swarm Intelligence Particle Swarm Optimization Introduction to swarm intelligence principles Particle Swarm Optimization (PSO) 3 Swarms in nature Fish, birds,
CS 1699: Intro to Computer Vision. Deep Learning. Prof. Adriana Kovashka University of Pittsburgh December 1, 2015
CS 1699: Intro to Computer Vision Deep Learning Prof. Adriana Kovashka University of Pittsburgh December 1, 2015 Today: Deep neural networks Background Architectures and basic operations Applications Visualizing
ARTIFICIAL INTELLIGENCE METHODS IN EARLY MANUFACTURING TIME ESTIMATION
1 ARTIFICIAL INTELLIGENCE METHODS IN EARLY MANUFACTURING TIME ESTIMATION B. Mikó PhD, Z-Form Tool Manufacturing and Application Ltd H-1082. Budapest, Asztalos S. u 4. Tel: (1) 477 1016, e-mail: [email protected]
Learning to Process Natural Language in Big Data Environment
CCF ADL 2015 Nanchang Oct 11, 2015 Learning to Process Natural Language in Big Data Environment Hang Li Noah s Ark Lab Huawei Technologies Part 1: Deep Learning - Present and Future Talk Outline Overview
An Early Attempt at Applying Deep Reinforcement Learning to the Game 2048
An Early Attempt at Applying Deep Reinforcement Learning to the Game 2048 Hong Gui, Tinghan Wei, Ching-Bo Huang, I-Chen Wu 1 1 Department of Computer Science, National Chiao Tung University, Hsinchu, Taiwan
Machine Learning CS 6830. Lecture 01. Razvan C. Bunescu School of Electrical Engineering and Computer Science [email protected]
Machine Learning CS 6830 Razvan C. Bunescu School of Electrical Engineering and Computer Science [email protected] What is Learning? Merriam-Webster: learn = to acquire knowledge, understanding, or skill
Programming Exercise 3: Multi-class Classification and Neural Networks
Programming Exercise 3: Multi-class Classification and Neural Networks Machine Learning November 4, 2011 Introduction In this exercise, you will implement one-vs-all logistic regression and neural networks
Chapter 2 The Research on Fault Diagnosis of Building Electrical System Based on RBF Neural Network
Chapter 2 The Research on Fault Diagnosis of Building Electrical System Based on RBF Neural Network Qian Wu, Yahui Wang, Long Zhang and Li Shen Abstract Building electrical system fault diagnosis is the
A Simple Feature Extraction Technique of a Pattern By Hopfield Network
A Simple Feature Extraction Technique of a Pattern By Hopfield Network A.Nag!, S. Biswas *, D. Sarkar *, P.P. Sarkar *, B. Gupta **! Academy of Technology, Hoogly - 722 *USIC, University of Kalyani, Kalyani
PLAANN as a Classification Tool for Customer Intelligence in Banking
PLAANN as a Classification Tool for Customer Intelligence in Banking EUNITE World Competition in domain of Intelligent Technologies The Research Report Ireneusz Czarnowski and Piotr Jedrzejowicz Department
ImageNet Classification with Deep Convolutional Neural Networks
ImageNet Classification with Deep Convolutional Neural Networks Alex Krizhevsky University of Toronto [email protected] Ilya Sutskever University of Toronto [email protected] Geoffrey E. Hinton University
Unsupervised Learning of Invariant Feature Hierarchies with Applications to Object Recognition
Unsupervised Learning of Invariant Feature Hierarchies with Applications to Object Recognition Marc Aurelio Ranzato, Fu-Jie Huang, Y-Lan Boureau, Yann LeCun Courant Institute of Mathematical Sciences,
NEURAL NETWORKS A Comprehensive Foundation
NEURAL NETWORKS A Comprehensive Foundation Second Edition Simon Haykin McMaster University Hamilton, Ontario, Canada Prentice Hall Prentice Hall Upper Saddle River; New Jersey 07458 Preface xii Acknowledgments
Lecture 6: Classification & Localization. boris. [email protected]
Lecture 6: Classification & Localization boris. [email protected] 1 Agenda ILSVRC 2014 Overfeat: integrated classification, localization, and detection Classification with Localization Detection. 2 ILSVRC-2014
NAVIGATING SCIENTIFIC LITERATURE A HOLISTIC PERSPECTIVE. Venu Govindaraju
NAVIGATING SCIENTIFIC LITERATURE A HOLISTIC PERSPECTIVE Venu Govindaraju BIOMETRICS DOCUMENT ANALYSIS PATTERN RECOGNITION 8/24/2015 ICDAR- 2015 2 Towards a Globally Optimal Approach for Learning Deep Unsupervised
Stochastic Pooling for Regularization of Deep Convolutional Neural Networks
Stochastic Pooling for Regularization of Deep Convolutional Neural Networks Matthew D. Zeiler Department of Computer Science Courant Institute, New York University [email protected] Rob Fergus Department
D A T A M I N I N G C L A S S I F I C A T I O N
D A T A M I N I N G C L A S S I F I C A T I O N FABRICIO VOZNIKA LEO NARDO VIA NA INTRODUCTION Nowadays there is huge amount of data being collected and stored in databases everywhere across the globe.
NEURAL NETWORK FUNDAMENTALS WITH GRAPHS, ALGORITHMS, AND APPLICATIONS
NEURAL NETWORK FUNDAMENTALS WITH GRAPHS, ALGORITHMS, AND APPLICATIONS N. K. Bose HRB-Systems Professor of Electrical Engineering The Pennsylvania State University, University Park P. Liang Associate Professor
Support Vector Machines with Clustering for Training with Very Large Datasets
Support Vector Machines with Clustering for Training with Very Large Datasets Theodoros Evgeniou Technology Management INSEAD Bd de Constance, Fontainebleau 77300, France [email protected] Massimiliano
Non-negative Matrix Factorization (NMF) in Semi-supervised Learning Reducing Dimension and Maintaining Meaning
Non-negative Matrix Factorization (NMF) in Semi-supervised Learning Reducing Dimension and Maintaining Meaning SAMSI 10 May 2013 Outline Introduction to NMF Applications Motivations NMF as a middle step
Applications of Deep Learning to the GEOINT mission. June 2015
Applications of Deep Learning to the GEOINT mission June 2015 Overview Motivation Deep Learning Recap GEOINT applications: Imagery exploitation OSINT exploitation Geospatial and activity based analytics
A Novel Binary Particle Swarm Optimization
Proceedings of the 5th Mediterranean Conference on T33- A Novel Binary Particle Swarm Optimization Motaba Ahmadieh Khanesar, Member, IEEE, Mohammad Teshnehlab and Mahdi Aliyari Shoorehdeli K. N. Toosi
HybrID: A Hybridization of Indirect and Direct Encodings for Evolutionary Computation
HybrID: A Hybridization of Indirect and Direct Encodings for Evolutionary Computation Jeff Clune 1, Benjamin E. Beckmann 1, Robert T. Pennock 1,2 and Charles Ofria 1 1 Department of Computer Science and
Introduction to Artificial Neural Networks
POLYTECHNIC UNIVERSITY Department of Computer and Information Science Introduction to Artificial Neural Networks K. Ming Leung Abstract: A computing paradigm known as artificial neural network is introduced.
ARTIFICIAL INTELLIGENCE (CSCU9YE) LECTURE 6: MACHINE LEARNING 2: UNSUPERVISED LEARNING (CLUSTERING)
ARTIFICIAL INTELLIGENCE (CSCU9YE) LECTURE 6: MACHINE LEARNING 2: UNSUPERVISED LEARNING (CLUSTERING) Gabriela Ochoa http://www.cs.stir.ac.uk/~goc/ OUTLINE Preliminaries Classification and Clustering Applications
Advanced Ensemble Strategies for Polynomial Models
Advanced Ensemble Strategies for Polynomial Models Pavel Kordík 1, Jan Černý 2 1 Dept. of Computer Science, Faculty of Information Technology, Czech Technical University in Prague, 2 Dept. of Computer
CS Master Level Courses and Areas COURSE DESCRIPTIONS. CSCI 521 Real-Time Systems. CSCI 522 High Performance Computing
CS Master Level Courses and Areas The graduate courses offered may change over time, in response to new developments in computer science and the interests of faculty and students; the list of graduate
Using Convolutional Neural Networks for Image Recognition
Using Convolutional Neural Networks for Image Recognition By Samer Hijazi, Rishi Kumar, and Chris Rowen, IP Group, Cadence Convolutional neural networks (CNNs) are widely used in pattern- and image-recognition
Applying Deep Learning to Enhance Momentum Trading Strategies in Stocks
This version: December 12, 2013 Applying Deep Learning to Enhance Momentum Trading Strategies in Stocks Lawrence Takeuchi * Yu-Ying (Albert) Lee [email protected] [email protected] Abstract We
Open Access Research on Application of Neural Network in Computer Network Security Evaluation. Shujuan Jin *
Send Orders for Reprints to [email protected] 766 The Open Electrical & Electronic Engineering Journal, 2014, 8, 766-771 Open Access Research on Application of Neural Network in Computer Network
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
arxiv:1312.6034v2 [cs.cv] 19 Apr 2014
Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps arxiv:1312.6034v2 [cs.cv] 19 Apr 2014 Karen Simonyan Andrea Vedaldi Andrew Zisserman Visual Geometry Group,
How To Use Neural Networks In Data Mining
International Journal of Electronics and Computer Science Engineering 1449 Available Online at www.ijecse.org ISSN- 2277-1956 Neural Networks in Data Mining Priyanka Gaur Department of Information and
Simplified Machine Learning for CUDA. Umar Arshad @arshad_umar Arrayfire @arrayfire
Simplified Machine Learning for CUDA Umar Arshad @arshad_umar Arrayfire @arrayfire ArrayFire CUDA and OpenCL experts since 2007 Headquartered in Atlanta, GA In search for the best and the brightest Expert
TRAINING A LIMITED-INTERCONNECT, SYNTHETIC NEURAL IC
777 TRAINING A LIMITED-INTERCONNECT, SYNTHETIC NEURAL IC M.R. Walker. S. Haghighi. A. Afghan. and L.A. Akers Center for Solid State Electronics Research Arizona State University Tempe. AZ 85287-6206 [email protected]
Classifying Manipulation Primitives from Visual Data
Classifying Manipulation Primitives from Visual Data Sandy Huang and Dylan Hadfield-Menell Abstract One approach to learning from demonstrations in robotics is to make use of a classifier to predict if
An Introduction to Neural Networks
An Introduction to Vincent Cheung Kevin Cannons Signal & Data Compression Laboratory Electrical & Computer Engineering University of Manitoba Winnipeg, Manitoba, Canada Advisor: Dr. W. Kinsner May 27,
Self Organizing Maps: Fundamentals
Self Organizing Maps: Fundamentals Introduction to Neural Networks : Lecture 16 John A. Bullinaria, 2004 1. What is a Self Organizing Map? 2. Topographic Maps 3. Setting up a Self Organizing Map 4. Kohonen
Role of Neural network in data mining
Role of Neural network in data mining Chitranjanjit kaur Associate Prof Guru Nanak College, Sukhchainana Phagwara,(GNDU) Punjab, India Pooja kapoor Associate Prof Swami Sarvanand Group Of Institutes Dinanagar(PTU)
A Systemic Artificial Intelligence (AI) Approach to Difficult Text Analytics Tasks
A Systemic Artificial Intelligence (AI) Approach to Difficult Text Analytics Tasks Text Analytics World, Boston, 2013 Lars Hard, CTO Agenda Difficult text analytics tasks Feature extraction Bio-inspired
Automatic Labeling of Lane Markings for Autonomous Vehicles
Automatic Labeling of Lane Markings for Autonomous Vehicles Jeffrey Kiske Stanford University 450 Serra Mall, Stanford, CA 94305 [email protected] 1. Introduction As autonomous vehicles become more popular,
Neural Network and Genetic Algorithm Based Trading Systems. Donn S. Fishbein, MD, PhD Neuroquant.com
Neural Network and Genetic Algorithm Based Trading Systems Donn S. Fishbein, MD, PhD Neuroquant.com Consider the challenge of constructing a financial market trading system using commonly available technical
Predict Influencers in the Social Network
Predict Influencers in the Social Network Ruishan Liu, Yang Zhao and Liuyu Zhou Email: rliu2, yzhao2, [email protected] Department of Electrical Engineering, Stanford University Abstract Given two persons
Categorical Data Visualization and Clustering Using Subjective Factors
Categorical Data Visualization and Clustering Using Subjective Factors Chia-Hui Chang and Zhi-Kai Ding Department of Computer Science and Information Engineering, National Central University, Chung-Li,
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set Amhmed A. Bhih School of Electrical and Electronic Engineering Princy Johnson School of Electrical and Electronic Engineering Martin
Efficient Reward Functions for Adaptive Multi-Rover Systems
Efficient Reward Functions for Adaptive Multi-Rover Systems Kagan Tumer 1 and Adrian Agogino 2 1 NASA Ames Research Center Mail Stop 269-4 Moffet Field, CA 94035 Phone:650-604-4940 Fax:650-604-4036 [email protected]
Visualization by Linear Projections as Information Retrieval
Visualization by Linear Projections as Information Retrieval Jaakko Peltonen Helsinki University of Technology, Department of Information and Computer Science, P. O. Box 5400, FI-0015 TKK, Finland [email protected]
NEUROEVOLUTION OF AUTO-TEACHING ARCHITECTURES
NEUROEVOLUTION OF AUTO-TEACHING ARCHITECTURES EDWARD ROBINSON & JOHN A. BULLINARIA School of Computer Science, University of Birmingham Edgbaston, Birmingham, B15 2TT, UK [email protected] This
An Overview of Knowledge Discovery Database and Data mining Techniques
An Overview of Knowledge Discovery Database and Data mining Techniques Priyadharsini.C 1, Dr. Antony Selvadoss Thanamani 2 M.Phil, Department of Computer Science, NGM College, Pollachi, Coimbatore, Tamilnadu,
A Learning Based Method for Super-Resolution of Low Resolution Images
A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 [email protected] Abstract The main objective of this project is the study of a learning based method
International Journal of Computer Science Trends and Technology (IJCST) Volume 3 Issue 3, May-June 2015
RESEARCH ARTICLE OPEN ACCESS Data Mining Technology for Efficient Network Security Management Ankit Naik [1], S.W. Ahmad [2] Student [1], Assistant Professor [2] Department of Computer Science and Engineering
Incremental PCA: An Alternative Approach for Novelty Detection
Incremental PCA: An Alternative Approach for Detection Hugo Vieira Neto and Ulrich Nehmzow Department of Computer Science University of Essex Wivenhoe Park Colchester CO4 3SQ {hvieir, udfn}@essex.ac.uk
Training Methods for Adaptive Boosting of Neural Networks for Character Recognition
Submission to NIPS*97, Category: Algorithms & Architectures, Preferred: Oral Training Methods for Adaptive Boosting of Neural Networks for Character Recognition Holger Schwenk Dept. IRO Université de Montréal
Visualization of large data sets using MDS combined with LVQ.
Visualization of large data sets using MDS combined with LVQ. Antoine Naud and Włodzisław Duch Department of Informatics, Nicholas Copernicus University, Grudziądzka 5, 87-100 Toruń, Poland. www.phys.uni.torun.pl/kmk
REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING
REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING Ms.PALLAVI CHOUDEKAR Ajay Kumar Garg Engineering College, Department of electrical and electronics Ms.SAYANTI BANERJEE Ajay Kumar Garg Engineering
Effect of Using Neural Networks in GA-Based School Timetabling
Effect of Using Neural Networks in GA-Based School Timetabling JANIS ZUTERS Department of Computer Science University of Latvia Raina bulv. 19, Riga, LV-1050 LATVIA [email protected] Abstract: - The school
