Training a Self-Organizing distributed on a PVM network

Size: px
Start display at page:

Download "Training a Self-Organizing distributed on a PVM network"

Transcription

1 Training a Self-Organizing Map distributed on a PVM network Nuno Bandeira Dep.Informatics, New University of Lisbon, Quinta da Torre 85 MONTE DA CAPARICA, PORTUGAL nb@di.fct.unl.pt Victor Jose Lobo Fernando Moura-Pires li. New University of Dep.Informatics, New University Lisbon/Portuguese Naval of Lisbon, Quinta da Tort-e 85 Academy, Escola Naval, Alfeite, MONTE DA CAPARICA, 8 ALMADA, PORTUGAL PORTUGAL vsl@di.fct.unl.pt finp@di.fct.unl.pt Abstract A distributed version of Kohonen s Self Organizing Map (SOM) algorithm is presented, and it s implementation on a Parallel Virtual Machine (PVM) network running on Intel based PCs with Microsoft Windows 95 is described. Numerical results are given for d#?erent work loads. Keywords: SOM, distributed algorithms, parallel algorithms, neural networks 1 - Introduction Although neural networks are intrinsically parallel algorithms, they are not easily implemented on distributed architectures because the strong interactions between neurons impose a very high communication overhead (neural networks are also called connectionist models). One of the neural models that has beenimplemented with more success onto parallel architectures is Kohonen s SOM [ 1][][3], because it requires very little communication between neurons. However these implementations have traditionally used parallel machines that tend to be expensive and non-standard []. Over the last few years, a system called Parallel Virtual Machine (PVM) [4][6], has been developed that enables a programmer to use networked computers (running different operating systems such as UNIX and MS- Windows 95) in a manner very similar to a single UNIX machine, using common languages such as C. Thanks to PVM, existing computer networks, no matter how heterogeneous, can easily be programmed. Moreover, simple PCs running MS-Windows (which abound in most organizations), can be put to work during otherwise unproductive times, such as nights and weekends. In this paper we describe our experience in trying to use the University s PC computer laboratories to train SOM networks. Although we intend to use these maps to classify sound, we used random vectors in these --I experiments. - Distributed SOM algorithm A thorough discussion of Kohonen s algorithm can be found in [5], but the core training algorithm may be described as follows: 1. For a given training pattern x: Calculate the distance of each neuron to the training pattern x (Calculation phase) 1. - Find the neuron with smaller distance, and call it the winner W (Voting phase) Change the network neurons with a function G, which depends on the learning rate a, the distance d to W (in the output plane), and the neighborhood function F. Due to the nature of the neighborhood function, only the neurons closer to W (in the output space) will be changed. (Update phase). Update the learning rate a and the neighborhood function F according to some rule 3. Repeat steps 1 and for the next training pattern, until some stopping criteria is reached. Many different distributed versions of Kohonen s SOM are possible, each being more adequate for a certain machine architecture. For implementing in PVM, we think the most adequate is the following: O l/98 $ IEEE 457

2 Given: Np processors (PVM hosts) A coordinator process C (within Np) Nt training patterns Xi N, neurons forming a SOM 1 - Assign the neurons. Assign the N, neurons to the processors, in such a way that each processor receives approximately the same number of neurons &/NJ., and that for any given area of the SOM the neurons are evenly distributed amongst the processors. - Set up the training set. Send all the N, training patterns to all the processors, together with the initial training parameters. 3 - Train the neurons. For each set, do the following: 3.1 Calculate the local (calculation phase) pattern xi in the training winner neuron in each processor 3. Send all local winner neuron coordinates and distances to the pattern x to the coordinator. At the coordinator, select the global winner, and send it s coordinates to all processors (voting phase) 3.3 Update the neurons at each processor, according to the update function F and update the network training parameters (update phase) 4 - Repeat step 3 until the stopping criteria is met 5 - Send all neurons back to the coordinator. A graphical representation of the algorithm messages involved is given in Figure 1. showing the The initial and final phases of the algorithm (represented in white in Figure 1) are executed only once, and thus have very little influence in the overall performance. Most of the time is spent in the main loop (represented in gray in Figure I), which iterates through the three main phases: calculation of the distances, voting for the global winner, and updating the neurons. The calculation phase of the algorithm (fmding the winner) is inherently parallel, and it s computational load can be spread evenly across the network if each processor has roughly the same number of neurons. The voting phase is the only one that requires communication ( and synchronization ) between processors, because the global winner must be known to all for the algorithm to proceed. If there was no coordinator, each processor would have to send information about it s local winner to all other processors. While PVM does support a broadcast mechanism, this would translate to a multicast at the data link level, thus originating N&VP-I) messages per iteration. With a coordinator, each process sends,only one message to the coordinator, and it in return sends only one message back, thus originating only (?V$) messages. Furthermore, the coordinator can piggyback additional information on the return message, that can be used to select the next training pattern, change the training parameters, etc. Coordinator Send start order Receive neurons Clients Receive neurons Figure 1 - Message exchange in distributed SOM During the update phase, each processor will have to calculate the distance between it s local neurons and the global winner (in the output plane), and then update only the neurons in p the winner s neighborhood. The computational load will be evenly distributed only if all processors have roughly the same number of neurons in this neighborhood. Thus it is very important to assign the neurons evenly during step 1 of the distributed algorithm. When the neighborhood radius is large, the computational load will easily be distributed. However, when the neighborhood radius is small, it is difficult to guarantee that all processors will have the same number of neurons, and thus some processors will have to wait for others. It must be pointed out that in this case, the number of neurons to update will be small, so the difference in processing time am-ongst the processors will be small. 3 - Experimental results During our experiments, we used networks of up to 1 PCs. Each was a Digital Venturis FX, with a Pentium running at loomhz, with 16Mb of RAM (55ISb cache). The computers where connected with a coaxial cable, 458

3 using lobase level protocol (1 Megabits per second), and TCP/IP as the level 3 protocol. The computers where running the MS-Windows 95 operating system, WPVM. [6] and Microsoft LAN Manager peer-to-peer network clients and servers. The computers had all screen-savers and anti-virus checkers disabled and where not running any other software during these tests. As the speedup obtained by using the distributed SOM depends critically on the amount of processing required before each synchronization, we used very large pattern vectors, with 14 features, and then varied the number of neurons on the map. We used square maps with 5x5, 1x1,x, and 4x4 neurons. Although square maps tend to slow convergence, we used them because they have a shorter boundary than rectangular maps (for the same number of neurons) and thus are less affected by discontinuities of the neighborhood function on those boundaries. This discontinuity effect would also affect the smaller maps more then the large ones if we used the same initial radius in all tests. To avoid this, we used an initial neighborhood function radius equal to each map s side. So as to make the radius decrease smoothly, we force it to decrease only 1 unit each time the whole training set patterns are presented. The map with 5x5 neurons will thus have only 5xh$ iterations of the patterns, while the one with 4x4 will have 4OxN,. So as not to make the simulation too long for larger maps, we use fewer training patterns for these maps. In the end, each map requires exactly 4 times more calculations then the previous one. Figure - Diagonal neuron distribution The number of patterns used for training does not influence the performance or speedup of the algorithm, so we use only 1 patterns for the 5x5 map, 6 for the 1x1, 3 for the x and 15 for the 4x4. While this number of patterns (and iterations) would be far too small for a useful classification or clustering, it is sufficient to prove that the system works reliably. The number of training patterns in the smaller maps has to be greater than for the larger maps, because otherwise the time intervals would be too small to be measured reliably. During these simulations we distributed the neurons amongst the processors in such a way that each processor has one or more diagonal lines of neurons. This distribution, although not ideal, provides a reasonable equilibrium of neurons by processors for rectangular neighborhoods (see Figure - every uncut neighborhood has the same neuron load per processor at each radius). The execution times where measured within the programs (with a call to a system timer), so as to measure only the time spent on the iterations (step 3 of the distributed algorithm, and shaded area in Figure 1). The numerical results are presented in Table 1 and in Figure 3, we can see a graph of the absolute execution times, while in Figure 4 we can see the relative times, that is, the speedup. tn g 5 Q i 3, No of machines Figure 3 - Absolute execution times ~(VdCOON \ \ No of machines + 5x5 -..-@b- 1x1 - Ref. Figure 4 - Relative execution time (l=time on a single machine) 459

4 Size of map No of PCs 5x5 1x1 x 4x4 r L Conclusions t Table 1 - Execution times (in seconds) The results presented confirm the claim that the SOM can be efficiently distributed on an ordinary computer network. However, depending on the work load, we may obtain overwhelming gains (as in the 4x4 map), moderate but consistent gains (as in the x map), or even high losses (as in the 5x5 map). There are a couple of extremely high running times corresponding to the 4x4 map running in a single machine or in a group of two. After that there is a sudden break and then the running times decrease smoothly. This initial peak is due to the machine configuration we are working on, namely because we have 16 Mb of RAM and a 4x4 map takes up to 13Mb RAM, forcing the operating system to use the disk as swap-space. We used machines with this configuration because we needed to have a reasonable pool of highly similar computers to achieve fair comparisons, and these where the ones available. Nevertheless we consider this as an advantage instead of a shortcoming, since these are the most common machines around any office or university lab, and allowed us to expose another very important fact when using distributed processing - the efficient use of each machine s memory. Using our distribution model, you can take advantage of each machine s local memory along with the corresponding processing power, effectively avoiding a RAM/HardDisk swapping situation which terribly slows down the SOM processing. In the more general case, the total execution time will decrease smoothly (except sometimes for the transition from 1 to machines), and then start to increase slowly. It s not reasonable to expect an unlimited gains as you throw more and more machines into the pool because of the increase in network load. When distributing a process, there will be a minimum overload on the total running time, due to the network. In our tests this can be seen by observing that the highest jumps upwards happen when there is a switch from a single machine to two machines. The distribution is profitable only when there is a significant workload to be distributed, thereby overcoming this minimum network * overload. After that initial step, all other machine extensions walk along an almost smooth curve, reflecting the converging equilibrium between each machine s designated workload and the network overload due to an increasing number of traded messages. The overload due to the network will increase linearly with the number of machines (each additional machine will be responsible for new messages), until the network starts to saturate due to collisions and/or sending queues. The workload per machine, on the other hand, will decrease hyperbolically, so at a certain point a minimum execution time will be reached, and adding a new machine will not improve the overall performance. We intend to concentrate our future work in three main areas. One of them is finding a better algorithm for distributing the neurons amongst processors. The ideal distribution will be a function of the radius of the neighborhood function and the number of processors, and genetic algorithms have been proposed to fmd good distributions. Another interesting problem is trying to fme tune the measurements of time spent calculating, transmitting, and waiting, during the iterations. On a PC running MS- Windows, this is rather tricky due to the lack of suitable tools. Finally, we intent to test and improve the algorithm for heterogeneous networks. This poses interesting problems of load balancing, and even more so if we attempt to do dynamic balancing during operation. Acknowledgments We would like to thank the Department de Engenharia Informatica da Universidade de Coimbra and the Escola Superior de Tecnologia e Gest%o do Instituto Politecnico de Viana do Castelo the for the development and availability of WPVM., the PVM implementation used in this work, and also the Prodep, FLAD and INVOTAN for the financial support to Nuno Bandeira (Prodep) and Victor Lobo (FLAD and INVOTAN). 46

5 References [I] Kohonen,T, et al.; Bibliography on Self-Organizing Map (SOM) and Learning Vector Quantization (LVQ), Helsinki 1997, available at ftp://cochlea.hut.fi [] Przytula, K.W, Prasanna, V.K.; Parallel Digital Implementations of Neural Networks, Prentice-Hall, [3] Ultsch, A, Siemon, H.P; Kohonen Networks on Transputers: Implementation and Animation, Proceedings of the International Neural Network Conference (INNC), July 199. [4] Geist, A., et al.; PVM: Parallel Virtual Machine - A User s guide and Tutorial for Networked Parallel Computing, The MIT Press, 1994 [5] Kohonen, Tuevo; Self-Organizing Maps, Springer- Verlag, 1995 [6] Alves, A, et al.; WPVM Manual, supplied with WPVM. package 461

Load Balancing on a Non-dedicated Heterogeneous Network of Workstations

Load Balancing on a Non-dedicated Heterogeneous Network of Workstations Load Balancing on a Non-dedicated Heterogeneous Network of Workstations Dr. Maurice Eggen Nathan Franklin Department of Computer Science Trinity University San Antonio, Texas 78212 Dr. Roger Eggen Department

More information

Load balancing in a heterogeneous computer system by self-organizing Kohonen network

Load balancing in a heterogeneous computer system by self-organizing Kohonen network Bull. Nov. Comp. Center, Comp. Science, 25 (2006), 69 74 c 2006 NCC Publisher Load balancing in a heterogeneous computer system by self-organizing Kohonen network Mikhail S. Tarkov, Yakov S. Bezrukov Abstract.

More information

ultra fast SOM using CUDA

ultra fast SOM using CUDA ultra fast SOM using CUDA SOM (Self-Organizing Map) is one of the most popular artificial neural network algorithms in the unsupervised learning category. Sijo Mathew Preetha Joy Sibi Rajendra Manoj A

More information

Rodrigo Fernandes de Mello, Evgueni Dodonov, José Augusto Andrade Filho

Rodrigo Fernandes de Mello, Evgueni Dodonov, José Augusto Andrade Filho Middleware for High Performance Computing Rodrigo Fernandes de Mello, Evgueni Dodonov, José Augusto Andrade Filho University of São Paulo São Carlos, Brazil {mello, eugeni, augustoa}@icmc.usp.br Outline

More information

Chapter 12: Multiprocessor Architectures. Lesson 01: Performance characteristics of Multiprocessor Architectures and Speedup

Chapter 12: Multiprocessor Architectures. Lesson 01: Performance characteristics of Multiprocessor Architectures and Speedup Chapter 12: Multiprocessor Architectures Lesson 01: Performance characteristics of Multiprocessor Architectures and Speedup Objective Be familiar with basic multiprocessor architectures and be able to

More information

CS Master Level Courses and Areas COURSE DESCRIPTIONS. CSCI 521 Real-Time Systems. CSCI 522 High Performance Computing

CS Master Level Courses and Areas COURSE DESCRIPTIONS. CSCI 521 Real-Time Systems. CSCI 522 High Performance Computing CS Master Level Courses and Areas The graduate courses offered may change over time, in response to new developments in computer science and the interests of faculty and students; the list of graduate

More information

Overlapping Data Transfer With Application Execution on Clusters

Overlapping Data Transfer With Application Execution on Clusters Overlapping Data Transfer With Application Execution on Clusters Karen L. Reid and Michael Stumm reid@cs.toronto.edu stumm@eecg.toronto.edu Department of Computer Science Department of Electrical and Computer

More information

Visualization of Breast Cancer Data by SOM Component Planes

Visualization of Breast Cancer Data by SOM Component Planes International Journal of Science and Technology Volume 3 No. 2, February, 2014 Visualization of Breast Cancer Data by SOM Component Planes P.Venkatesan. 1, M.Mullai 2 1 Department of Statistics,NIRT(Indian

More information

Well, now you have a Language Lab! www.easylab.es

Well, now you have a Language Lab! www.easylab.es Do you have a computer lab? Well, now you have a Language Lab! What exactly is easylab? easylab can transform any ordinary computer laboratory into a cuttingedge Language Laboratory. easylab is a system

More information

Capacity Estimation for Linux Workloads

Capacity Estimation for Linux Workloads Capacity Estimation for Linux Workloads Session L985 David Boyes Sine Nomine Associates 1 Agenda General Capacity Planning Issues Virtual Machine History and Value Unique Capacity Issues in Virtual Machines

More information

Local Area Networks: Software

Local Area Networks: Software School of Business Eastern Illinois University Local Area Networks: Software (Week 8, Thursday 3/1/2007) Abdou Illia, Spring 2007 Learning Objectives 2 Identify main functions of operating systems Describe

More information

white paper Capacity and Scaling of Microsoft Terminal Server on the Unisys ES7000/600 Unisys Systems & Technology Modeling and Measurement

white paper Capacity and Scaling of Microsoft Terminal Server on the Unisys ES7000/600 Unisys Systems & Technology Modeling and Measurement white paper Capacity and Scaling of Microsoft Terminal Server on the Unisys ES7000/600 Unisys Systems & Technology Modeling and Measurement 2 This technical white paper has been written for IT professionals

More information

Distributed Systems LEEC (2005/06 2º Sem.)

Distributed Systems LEEC (2005/06 2º Sem.) Distributed Systems LEEC (2005/06 2º Sem.) Introduction João Paulo Carvalho Universidade Técnica de Lisboa / Instituto Superior Técnico Outline Definition of a Distributed System Goals Connecting Users

More information

Reliable Systolic Computing through Redundancy

Reliable Systolic Computing through Redundancy Reliable Systolic Computing through Redundancy Kunio Okuda 1, Siang Wun Song 1, and Marcos Tatsuo Yamamoto 1 Universidade de São Paulo, Brazil, {kunio,song,mty}@ime.usp.br, http://www.ime.usp.br/ song/

More information

A Simultaneous Solution for General Linear Equations on a Ring or Hierarchical Cluster

A Simultaneous Solution for General Linear Equations on a Ring or Hierarchical Cluster Acta Technica Jaurinensis Vol. 3. No. 1. 010 A Simultaneous Solution for General Linear Equations on a Ring or Hierarchical Cluster G. Molnárka, N. Varjasi Széchenyi István University Győr, Hungary, H-906

More information

Parallel Ray Tracing using MPI: A Dynamic Load-balancing Approach

Parallel Ray Tracing using MPI: A Dynamic Load-balancing Approach Parallel Ray Tracing using MPI: A Dynamic Load-balancing Approach S. M. Ashraful Kadir 1 and Tazrian Khan 2 1 Scientific Computing, Royal Institute of Technology (KTH), Stockholm, Sweden smakadir@csc.kth.se,

More information

Distributed Dynamic Load Balancing for Iterative-Stencil Applications

Distributed Dynamic Load Balancing for Iterative-Stencil Applications Distributed Dynamic Load Balancing for Iterative-Stencil Applications G. Dethier 1, P. Marchot 2 and P.A. de Marneffe 1 1 EECS Department, University of Liege, Belgium 2 Chemical Engineering Department,

More information

Models of Cortical Maps II

Models of Cortical Maps II CN510: Principles and Methods of Cognitive and Neural Modeling Models of Cortical Maps II Lecture 19 Instructor: Anatoli Gorchetchnikov dy dt The Network of Grossberg (1976) Ay B y f (

More information

Web Server (Step 1) Processes request and sends query to SQL server via ADO/OLEDB. Web Server (Step 2) Creates HTML page dynamically from record set

Web Server (Step 1) Processes request and sends query to SQL server via ADO/OLEDB. Web Server (Step 2) Creates HTML page dynamically from record set Dawn CF Performance Considerations Dawn CF key processes Request (http) Web Server (Step 1) Processes request and sends query to SQL server via ADO/OLEDB. Query (SQL) SQL Server Queries Database & returns

More information

6.2.8 Neural networks for data mining

6.2.8 Neural networks for data mining 6.2.8 Neural networks for data mining Walter Kosters 1 In many application areas neural networks are known to be valuable tools. This also holds for data mining. In this chapter we discuss the use of neural

More information

Grid e-services for Multi-Layer SOM Neural Network Simulation

Grid e-services for Multi-Layer SOM Neural Network Simulation Grid e-services for Multi-Layer SOM Neural Network Simulation,, Rui Silva Faculdade de Engenharia 4760-108 V. N. Famalicão, Portugal {rml,rsilva}@fam.ulusiada.pt 2007 Outline Overview Multi-Layer SOM Background

More information

pc resource monitoring and performance advisor

pc resource monitoring and performance advisor pc resource monitoring and performance advisor application note www.hp.com/go/desktops Overview HP Toptools is a modular web-based device management tool that provides dynamic information about HP hardware

More information

A Performance Study of Load Balancing Strategies for Approximate String Matching on an MPI Heterogeneous System Environment

A Performance Study of Load Balancing Strategies for Approximate String Matching on an MPI Heterogeneous System Environment A Performance Study of Load Balancing Strategies for Approximate String Matching on an MPI Heterogeneous System Environment Panagiotis D. Michailidis and Konstantinos G. Margaritis Parallel and Distributed

More information

Enhance student s learning with an aid of simulation software to understand Computer Networking Undergraduate courses.

Enhance student s learning with an aid of simulation software to understand Computer Networking Undergraduate courses. Enhance student s learning with an aid of simulation software to understand Computer Networking Undergraduate courses. Veeramani Shanmugam s.veeramani@curtin.edu.my Lenin Gopal lenin@curtin.edu.my Zeya

More information

Processor Capacity Reserves: An Abstraction for Managing Processor Usage

Processor Capacity Reserves: An Abstraction for Managing Processor Usage Processor Capacity Reserves: An Abstraction for Managing Processor Usage Clifford W. Mercer, Stefan Savage, and Hideyuki Tokuda School of Computer Science Carnegie Mellon University Pittsburgh, Pennsylvania

More information

Windows Server Performance Monitoring

Windows Server Performance Monitoring Spot server problems before they are noticed The system s really slow today! How often have you heard that? Finding the solution isn t so easy. The obvious questions to ask are why is it running slowly

More information

Attunity RepliWeb Event Driven Jobs

Attunity RepliWeb Event Driven Jobs Attunity RepliWeb Event Driven Jobs Software Version 5.2 June 25, 2012 RepliWeb, Inc., 6441 Lyons Road, Coconut Creek, FL 33073 Tel: (954) 946-2274, Fax: (954) 337-6424 E-mail: info@repliweb.com, Support:

More information

GPU File System Encryption Kartik Kulkarni and Eugene Linkov

GPU File System Encryption Kartik Kulkarni and Eugene Linkov GPU File System Encryption Kartik Kulkarni and Eugene Linkov 5/10/2012 SUMMARY. We implemented a file system that encrypts and decrypts files. The implementation uses the AES algorithm computed through

More information

Building an Inexpensive Parallel Computer

Building an Inexpensive Parallel Computer Res. Lett. Inf. Math. Sci., (2000) 1, 113-118 Available online at http://www.massey.ac.nz/~wwiims/rlims/ Building an Inexpensive Parallel Computer Lutz Grosz and Andre Barczak I.I.M.S., Massey University

More information

White Paper Perceived Performance Tuning a system for what really matters

White Paper Perceived Performance Tuning a system for what really matters TMurgent Technologies White Paper Perceived Performance Tuning a system for what really matters September 18, 2003 White Paper: Perceived Performance 1/7 TMurgent Technologies Introduction The purpose

More information

AN AUTOMATED DATA THROUGHPUT MEASUREMENT OVER A 10/100MBPS LAN

AN AUTOMATED DATA THROUGHPUT MEASUREMENT OVER A 10/100MBPS LAN AN AUTOMATED DATA THROUGHPUT MEASUREMENT OVER A 10/100MBPS LAN John B. Oladosu* & Abigail B. Adetunji Computer Science & Engineering Department, Ladoke Akintola University of Technology, Ogbomoso, Nigeria

More information

Design Issues in a Bare PC Web Server

Design Issues in a Bare PC Web Server Design Issues in a Bare PC Web Server Long He, Ramesh K. Karne, Alexander L. Wijesinha, Sandeep Girumala, and Gholam H. Khaksari Department of Computer & Information Sciences, Towson University, 78 York

More information

Agenda. Distributed System Structures. Why Distributed Systems? Motivation

Agenda. Distributed System Structures. Why Distributed Systems? Motivation Agenda Distributed System Structures CSCI 444/544 Operating Systems Fall 2008 Motivation Network structure Fundamental network services Sockets and ports Client/server model Remote Procedure Call (RPC)

More information

Sla Aware Load Balancing Algorithm Using Join-Idle Queue for Virtual Machines in Cloud Computing

Sla Aware Load Balancing Algorithm Using Join-Idle Queue for Virtual Machines in Cloud Computing Sla Aware Load Balancing Using Join-Idle Queue for Virtual Machines in Cloud Computing Mehak Choudhary M.Tech Student [CSE], Dept. of CSE, SKIET, Kurukshetra University, Haryana, India ABSTRACT: Cloud

More information

LOAD BALANCING STRATEGY BASED ON CLOUD PARTITIONING CONCEPT

LOAD BALANCING STRATEGY BASED ON CLOUD PARTITIONING CONCEPT Journal homepage: www.mjret.in ISSN:2348-6953 LOAD BALANCING STRATEGY BASED ON CLOUD PARTITIONING CONCEPT Ms. Shilpa D.More 1, Prof. Arti Mohanpurkar 2 1,2 Department of computer Engineering DYPSOET, Pune,India

More information

A user friendly toolbox for exploratory data analysis of underwater sound

A user friendly toolbox for exploratory data analysis of underwater sound 061215-064 1 A user friendly toolbox for exploratory data analysis of underwater sound Fernando J. Pires 1 and Victor Lobo 2, Member, IEEE Abstract The underwater acoustics research group at the Portuguese

More information

Muse Server Sizing. 18 June 2012. Document Version 0.0.1.9 Muse 2.7.0.0

Muse Server Sizing. 18 June 2012. Document Version 0.0.1.9 Muse 2.7.0.0 Muse Server Sizing 18 June 2012 Document Version 0.0.1.9 Muse 2.7.0.0 Notice No part of this publication may be reproduced stored in a retrieval system, or transmitted, in any form or by any means, without

More information

Parallelism and Cloud Computing

Parallelism and Cloud Computing Parallelism and Cloud Computing Kai Shen Parallel Computing Parallel computing: Process sub tasks simultaneously so that work can be completed faster. For instances: divide the work of matrix multiplication

More information

Elemental functions: Writing data-parallel code in C/C++ using Intel Cilk Plus

Elemental functions: Writing data-parallel code in C/C++ using Intel Cilk Plus Elemental functions: Writing data-parallel code in C/C++ using Intel Cilk Plus A simple C/C++ language extension construct for data parallel operations Robert Geva robert.geva@intel.com Introduction Intel

More information

Parallel & Distributed Optimization. Based on Mark Schmidt s slides

Parallel & Distributed Optimization. Based on Mark Schmidt s slides Parallel & Distributed Optimization Based on Mark Schmidt s slides Motivation behind using parallel & Distributed optimization Performance Computational throughput have increased exponentially in linear

More information

Computer Network. Interconnected collection of autonomous computers that are able to exchange information

Computer Network. Interconnected collection of autonomous computers that are able to exchange information Introduction Computer Network. Interconnected collection of autonomous computers that are able to exchange information No master/slave relationship between the computers in the network Data Communications.

More information

A Review of Customized Dynamic Load Balancing for a Network of Workstations

A Review of Customized Dynamic Load Balancing for a Network of Workstations A Review of Customized Dynamic Load Balancing for a Network of Workstations Taken from work done by: Mohammed Javeed Zaki, Wei Li, Srinivasan Parthasarathy Computer Science Department, University of Rochester

More information

Efficient Load Balancing using VM Migration by QEMU-KVM

Efficient Load Balancing using VM Migration by QEMU-KVM International Journal of Computer Science and Telecommunications [Volume 5, Issue 8, August 2014] 49 ISSN 2047-3338 Efficient Load Balancing using VM Migration by QEMU-KVM Sharang Telkikar 1, Shreyas Talele

More information

Keywords: Dynamic Load Balancing, Process Migration, Load Indices, Threshold Level, Response Time, Process Age.

Keywords: Dynamic Load Balancing, Process Migration, Load Indices, Threshold Level, Response Time, Process Age. Volume 3, Issue 10, October 2013 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Load Measurement

More information

How To Test For Performance And Scalability On A Server With A Multi-Core Computer (For A Large Server)

How To Test For Performance And Scalability On A Server With A Multi-Core Computer (For A Large Server) Scalability Results Select the right hardware configuration for your organization to optimize performance Table of Contents Introduction... 1 Scalability... 2 Definition... 2 CPU and Memory Usage... 2

More information

DYNAMIC LOAD BALANCING APPLICATIONS ON A HETEROGENEOUS UNIX/NT CLUSTER

DYNAMIC LOAD BALANCING APPLICATIONS ON A HETEROGENEOUS UNIX/NT CLUSTER European Congress on Computational Methods in Applied Sciences and Engineering ECCOMAS 2000 Barcelona, 11-14 September 2000 ECCOMAS DYNAMIC LOAD BALANCING APPLICATIONS ON A HETEROGENEOUS UNIX/NT CLUSTER

More information

Understanding the Benefits of IBM SPSS Statistics Server

Understanding the Benefits of IBM SPSS Statistics Server IBM SPSS Statistics Server Understanding the Benefits of IBM SPSS Statistics Server Contents: 1 Introduction 2 Performance 101: Understanding the drivers of better performance 3 Why performance is faster

More information

CHAPTER 1 INTRODUCTION

CHAPTER 1 INTRODUCTION 1 CHAPTER 1 INTRODUCTION 1.1 MOTIVATION OF RESEARCH Multicore processors have two or more execution cores (processors) implemented on a single chip having their own set of execution and architectural recourses.

More information

LCMON Network Traffic Analysis

LCMON Network Traffic Analysis LCMON Network Traffic Analysis Adam Black Centre for Advanced Internet Architectures, Technical Report 79A Swinburne University of Technology Melbourne, Australia adamblack@swin.edu.au Abstract The Swinburne

More information

Control 2004, University of Bath, UK, September 2004

Control 2004, University of Bath, UK, September 2004 Control, University of Bath, UK, September ID- IMPACT OF DEPENDENCY AND LOAD BALANCING IN MULTITHREADING REAL-TIME CONTROL ALGORITHMS M A Hossain and M O Tokhi Department of Computing, The University of

More information

Neural network software tool development: exploring programming language options

Neural network software tool development: exploring programming language options INEB- PSI Technical Report 2006-1 Neural network software tool development: exploring programming language options Alexandra Oliveira aao@fe.up.pt Supervisor: Professor Joaquim Marques de Sá June 2006

More information

An Oracle White Paper November 2010. Oracle Real Application Clusters One Node: The Always On Single-Instance Database

An Oracle White Paper November 2010. Oracle Real Application Clusters One Node: The Always On Single-Instance Database An Oracle White Paper November 2010 Oracle Real Application Clusters One Node: The Always On Single-Instance Database Executive Summary... 1 Oracle Real Application Clusters One Node Overview... 1 Always

More information

Bootstrapping Big Data

Bootstrapping Big Data Bootstrapping Big Data Ariel Kleiner Ameet Talwalkar Purnamrita Sarkar Michael I. Jordan Computer Science Division University of California, Berkeley {akleiner, ameet, psarkar, jordan}@eecs.berkeley.edu

More information

EFFICIENT DATA PRE-PROCESSING FOR DATA MINING

EFFICIENT DATA PRE-PROCESSING FOR DATA MINING EFFICIENT DATA PRE-PROCESSING FOR DATA MINING USING NEURAL NETWORKS JothiKumar.R 1, Sivabalan.R.V 2 1 Research scholar, Noorul Islam University, Nagercoil, India Assistant Professor, Adhiparasakthi College

More information

SERVER CLUSTERING TECHNOLOGY & CONCEPT

SERVER CLUSTERING TECHNOLOGY & CONCEPT SERVER CLUSTERING TECHNOLOGY & CONCEPT M00383937, Computer Network, Middlesex University, E mail: vaibhav.mathur2007@gmail.com Abstract Server Cluster is one of the clustering technologies; it is use for

More information

Self Organizing Maps: Fundamentals

Self Organizing Maps: Fundamentals Self Organizing Maps: Fundamentals Introduction to Neural Networks : Lecture 16 John A. Bullinaria, 2004 1. What is a Self Organizing Map? 2. Topographic Maps 3. Setting up a Self Organizing Map 4. Kohonen

More information

SCHEDULING IN CLOUD COMPUTING

SCHEDULING IN CLOUD COMPUTING SCHEDULING IN CLOUD COMPUTING Lipsa Tripathy, Rasmi Ranjan Patra CSA,CPGS,OUAT,Bhubaneswar,Odisha Abstract Cloud computing is an emerging technology. It process huge amount of data so scheduling mechanism

More information

Efficient DNS based Load Balancing for Bursty Web Application Traffic

Efficient DNS based Load Balancing for Bursty Web Application Traffic ISSN Volume 1, No.1, September October 2012 International Journal of Science the and Internet. Applied However, Information this trend leads Technology to sudden burst of Available Online at http://warse.org/pdfs/ijmcis01112012.pdf

More information

MANAGING QUEUE STABILITY USING ART2 IN ACTIVE QUEUE MANAGEMENT FOR CONGESTION CONTROL

MANAGING QUEUE STABILITY USING ART2 IN ACTIVE QUEUE MANAGEMENT FOR CONGESTION CONTROL MANAGING QUEUE STABILITY USING ART2 IN ACTIVE QUEUE MANAGEMENT FOR CONGESTION CONTROL G. Maria Priscilla 1 and C. P. Sumathi 2 1 S.N.R. Sons College (Autonomous), Coimbatore, India 2 SDNB Vaishnav College

More information

USING SELF-ORGANISING MAPS FOR ANOMALOUS BEHAVIOUR DETECTION IN A COMPUTER FORENSIC INVESTIGATION

USING SELF-ORGANISING MAPS FOR ANOMALOUS BEHAVIOUR DETECTION IN A COMPUTER FORENSIC INVESTIGATION USING SELF-ORGANISING MAPS FOR ANOMALOUS BEHAVIOUR DETECTION IN A COMPUTER FORENSIC INVESTIGATION B.K.L. Fei, J.H.P. Eloff, M.S. Olivier, H.M. Tillwick and H.S. Venter Information and Computer Security

More information

XTM Web 2.0 Enterprise Architecture Hardware Implementation Guidelines. A.Zydroń 18 April 2009. Page 1 of 12

XTM Web 2.0 Enterprise Architecture Hardware Implementation Guidelines. A.Zydroń 18 April 2009. Page 1 of 12 XTM Web 2.0 Enterprise Architecture Hardware Implementation Guidelines A.Zydroń 18 April 2009 Page 1 of 12 1. Introduction...3 2. XTM Database...4 3. JVM and Tomcat considerations...5 4. XTM Engine...5

More information

The software beyond the climatic. Environment simulation

The software beyond the climatic. Environment simulation Spirale 2 The software beyond the climatic Environment simulation Spirale 2... Take it easy! Spirale 2 is a new software, based on a reliable system (Windows NT) and if its performances are surprising,

More information

Dynamic load balancing of parallel cellular automata

Dynamic load balancing of parallel cellular automata Dynamic load balancing of parallel cellular automata Marc Mazzariol, Benoit A. Gennart, Roger D. Hersch Ecole Polytechnique Fédérale de Lausanne, EPFL * ABSTRACT We are interested in running in parallel

More information

Operating Systems for Parallel Processing Assistent Lecturer Alecu Felician Economic Informatics Department Academy of Economic Studies Bucharest

Operating Systems for Parallel Processing Assistent Lecturer Alecu Felician Economic Informatics Department Academy of Economic Studies Bucharest Operating Systems for Parallel Processing Assistent Lecturer Alecu Felician Economic Informatics Department Academy of Economic Studies Bucharest 1. Introduction Few years ago, parallel computers could

More information

A Comparison of Distributed Systems: ChorusOS and Amoeba

A Comparison of Distributed Systems: ChorusOS and Amoeba A Comparison of Distributed Systems: ChorusOS and Amoeba Angelo Bertolli Prepared for MSIT 610 on October 27, 2004 University of Maryland University College Adelphi, Maryland United States of America Abstract.

More information

Load Balance Strategies for DEVS Approximated Parallel and Distributed Discrete-Event Simulations

Load Balance Strategies for DEVS Approximated Parallel and Distributed Discrete-Event Simulations Load Balance Strategies for DEVS Approximated Parallel and Distributed Discrete-Event Simulations Alonso Inostrosa-Psijas, Roberto Solar, Verónica Gil-Costa and Mauricio Marín Universidad de Santiago,

More information

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays Red Hat Performance Engineering Version 1.0 August 2013 1801 Varsity Drive Raleigh NC

More information

Simplest Scalable Architecture

Simplest Scalable Architecture Simplest Scalable Architecture NOW Network Of Workstations Many types of Clusters (form HP s Dr. Bruce J. Walker) High Performance Clusters Beowulf; 1000 nodes; parallel programs; MPI Load-leveling Clusters

More information

Introduction 1 Performance on Hosted Server 1. Benchmarks 2. System Requirements 7 Load Balancing 7

Introduction 1 Performance on Hosted Server 1. Benchmarks 2. System Requirements 7 Load Balancing 7 Introduction 1 Performance on Hosted Server 1 Figure 1: Real World Performance 1 Benchmarks 2 System configuration used for benchmarks 2 Figure 2a: New tickets per minute on E5440 processors 3 Figure 2b:

More information

ASSEMBLY PROGRAMMING ON A VIRTUAL COMPUTER

ASSEMBLY PROGRAMMING ON A VIRTUAL COMPUTER ASSEMBLY PROGRAMMING ON A VIRTUAL COMPUTER Pierre A. von Kaenel Mathematics and Computer Science Department Skidmore College Saratoga Springs, NY 12866 (518) 580-5292 pvonk@skidmore.edu ABSTRACT This paper

More information

Ready Time Observations

Ready Time Observations VMWARE PERFORMANCE STUDY VMware ESX Server 3 Ready Time Observations VMware ESX Server is a thin software layer designed to multiplex hardware resources efficiently among virtual machines running unmodified

More information

Task Scheduling in Hadoop

Task Scheduling in Hadoop Task Scheduling in Hadoop Sagar Mamdapure Munira Ginwala Neha Papat SAE,Kondhwa SAE,Kondhwa SAE,Kondhwa Abstract Hadoop is widely used for storing large datasets and processing them efficiently under distributed

More information

VDI FIT and VDI UX: Composite Metrics Track Good, Fair, Poor Desktop Performance

VDI FIT and VDI UX: Composite Metrics Track Good, Fair, Poor Desktop Performance VDI FIT and VDI UX: Composite Metrics Track Good, Fair, Poor Desktop Performance Key indicators and classification capabilities in Stratusphere FIT and Stratusphere UX Whitepaper INTRODUCTION This whitepaper

More information

Scaling from 1 PC to a super computer using Mascot

Scaling from 1 PC to a super computer using Mascot Scaling from 1 PC to a super computer using Mascot 1 Mascot - options for high throughput When is more processing power required? Options: Multiple Mascot installations Using a large Unix server Clusters

More information

Understanding the Performance of an X550 11-User Environment

Understanding the Performance of an X550 11-User Environment Understanding the Performance of an X550 11-User Environment Overview NComputing's desktop virtualization technology enables significantly lower computing costs by letting multiple users share a single

More information

Microsoft Windows Server 2003 with Internet Information Services (IIS) 6.0 vs. Linux Competitive Web Server Performance Comparison

Microsoft Windows Server 2003 with Internet Information Services (IIS) 6.0 vs. Linux Competitive Web Server Performance Comparison April 23 11 Aviation Parkway, Suite 4 Morrisville, NC 2756 919-38-28 Fax 919-38-2899 32 B Lakeside Drive Foster City, CA 9444 65-513-8 Fax 65-513-899 www.veritest.com info@veritest.com Microsoft Windows

More information

Real Time Network Server Monitoring using Smartphone with Dynamic Load Balancing

Real Time Network Server Monitoring using Smartphone with Dynamic Load Balancing www.ijcsi.org 227 Real Time Network Server Monitoring using Smartphone with Dynamic Load Balancing Dhuha Basheer Abdullah 1, Zeena Abdulgafar Thanoon 2, 1 Computer Science Department, Mosul University,

More information

SIP Registration Stress Test

SIP Registration Stress Test SIP Registration Stress Test Miroslav Voznak and Jan Rozhon Department of Telecommunications VSB Technical University of Ostrava 17. listopadu 15/2172, 708 33 Ostrava Poruba CZECH REPUBLIC miroslav.voznak@vsb.cz,

More information

Chapter 4: Artificial Neural Networks

Chapter 4: Artificial Neural Networks Chapter 4: Artificial Neural Networks CS 536: Machine Learning Littman (Wu, TA) Administration icml-03: instructional Conference on Machine Learning http://www.cs.rutgers.edu/~mlittman/courses/ml03/icml03/

More information

Recommended hardware system configurations for ANSYS users

Recommended hardware system configurations for ANSYS users Recommended hardware system configurations for ANSYS users The purpose of this document is to recommend system configurations that will deliver high performance for ANSYS users across the entire range

More information

Preserving Message Integrity in Dynamic Process Migration

Preserving Message Integrity in Dynamic Process Migration Preserving Message Integrity in Dynamic Process Migration E. Heymann, F. Tinetti, E. Luque Universidad Autónoma de Barcelona Departamento de Informática 8193 - Bellaterra, Barcelona, Spain e-mail: e.heymann@cc.uab.es

More information

Experiments on the local load balancing algorithms; part 1

Experiments on the local load balancing algorithms; part 1 Experiments on the local load balancing algorithms; part 1 Ştefan Măruşter Institute e-austria Timisoara West University of Timişoara, Romania maruster@info.uvt.ro Abstract. In this paper the influence

More information

Infor Web UI Sizing and Deployment for a Thin Client Solution

Infor Web UI Sizing and Deployment for a Thin Client Solution Infor Web UI Sizing and Deployment for a Thin Client Solution Copyright 2012 Infor Important Notices The material contained in this publication (including any supplementary information) constitutes and

More information

Analysis of an Artificial Hormone System (Extended abstract)

Analysis of an Artificial Hormone System (Extended abstract) c 2013. This is the author s version of the work. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purpose or for creating

More information

Visualization of large data sets using MDS combined with LVQ.

Visualization of large data sets using MDS combined with LVQ. Visualization of large data sets using MDS combined with LVQ. Antoine Naud and Włodzisław Duch Department of Informatics, Nicholas Copernicus University, Grudziądzka 5, 87-100 Toruń, Poland. www.phys.uni.torun.pl/kmk

More information

MAS 200 Supported Platform Matrix

MAS 200 Supported Platform Matrix The information in this document applies to MAS 200 Level 3.60 and higher. Best Software only supports the current MAS 200 level and one major level back. As of the Level 3.70 release, MAS 200 Levels 3.51

More information

Performance analysis of a Linux based FTP server

Performance analysis of a Linux based FTP server Performance analysis of a Linux based FTP server A Thesis Submitted in Partial Fulfillment of the Requirements for the Degree of Master of Technology by Anand Srivastava to the Department of Computer Science

More information

Parallel Computing for Option Pricing Based on the Backward Stochastic Differential Equation

Parallel Computing for Option Pricing Based on the Backward Stochastic Differential Equation Parallel Computing for Option Pricing Based on the Backward Stochastic Differential Equation Ying Peng, Bin Gong, Hui Liu, and Yanxin Zhang School of Computer Science and Technology, Shandong University,

More information

Load Distribution on a Linux Cluster using Load Balancing

Load Distribution on a Linux Cluster using Load Balancing Load Distribution on a Linux Cluster using Load Balancing Aravind Elango M. Mohammed Safiq Undergraduate Students of Engg. Dept. of Computer Science and Engg. PSG College of Technology India Abstract:

More information

LVQ Plug-In Algorithm for SQL Server

LVQ Plug-In Algorithm for SQL Server LVQ Plug-In Algorithm for SQL Server Licínia Pedro Monteiro Instituto Superior Técnico licinia.monteiro@tagus.ist.utl.pt I. Executive Summary In this Resume we describe a new functionality implemented

More information

High Performance Cluster Support for NLB on Window

High Performance Cluster Support for NLB on Window High Performance Cluster Support for NLB on Window [1]Arvind Rathi, [2] Kirti, [3] Neelam [1]M.Tech Student, Department of CSE, GITM, Gurgaon Haryana (India) arvindrathi88@gmail.com [2]Asst. Professor,

More information

Chapter 18: Database System Architectures. Centralized Systems

Chapter 18: Database System Architectures. Centralized Systems Chapter 18: Database System Architectures! Centralized Systems! Client--Server Systems! Parallel Systems! Distributed Systems! Network Types 18.1 Centralized Systems! Run on a single computer system and

More information

Introduction to Virtual Machines

Introduction to Virtual Machines Introduction to Virtual Machines Introduction Abstraction and interfaces Virtualization Computer system architecture Process virtual machines System virtual machines 1 Abstraction Mechanism to manage complexity

More information

Week 1 out-of-class notes, discussions and sample problems

Week 1 out-of-class notes, discussions and sample problems Week 1 out-of-class notes, discussions and sample problems Although we will primarily concentrate on RISC processors as found in some desktop/laptop computers, here we take a look at the varying types

More information

Dell Compellent Storage Center SAN & VMware View 1,000 Desktop Reference Architecture. Dell Compellent Product Specialist Team

Dell Compellent Storage Center SAN & VMware View 1,000 Desktop Reference Architecture. Dell Compellent Product Specialist Team Dell Compellent Storage Center SAN & VMware View 1,000 Desktop Reference Architecture Dell Compellent Product Specialist Team THIS WHITE PAPER IS FOR INFORMATIONAL PURPOSES ONLY, AND MAY CONTAIN TYPOGRAPHICAL

More information

Performance Modeling and Analysis of a Database Server with Write-Heavy Workload

Performance Modeling and Analysis of a Database Server with Write-Heavy Workload Performance Modeling and Analysis of a Database Server with Write-Heavy Workload Manfred Dellkrantz, Maria Kihl 2, and Anders Robertsson Department of Automatic Control, Lund University 2 Department of

More information

Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets

Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets Macario O. Cordel II and Arnulfo P. Azcarraga College of Computer Studies *Corresponding Author: macario.cordel@dlsu.edu.ph

More information

A Guide to Getting Started with Successful Load Testing

A Guide to Getting Started with Successful Load Testing Ingenieurbüro David Fischer AG A Company of the Apica Group http://www.proxy-sniffer.com A Guide to Getting Started with Successful Load Testing English Edition 2007 All Rights Reserved Table of Contents

More information

Learning Vector Quantization: generalization ability and dynamics of competing prototypes

Learning Vector Quantization: generalization ability and dynamics of competing prototypes Learning Vector Quantization: generalization ability and dynamics of competing prototypes Aree Witoelar 1, Michael Biehl 1, and Barbara Hammer 2 1 University of Groningen, Mathematics and Computing Science

More information

Intel Cloud Builders Guide to Cloud Design and Deployment on Intel Platforms

Intel Cloud Builders Guide to Cloud Design and Deployment on Intel Platforms Intel Cloud Builders Guide Intel Xeon Processor-based Servers RES Virtual Desktop Extender Intel Cloud Builders Guide to Cloud Design and Deployment on Intel Platforms Client Aware Cloud with RES Virtual

More information