Training a Self-Organizing distributed on a PVM network
|
|
- Gerald Anthony
- 8 years ago
- Views:
Transcription
1 Training a Self-Organizing Map distributed on a PVM network Nuno Bandeira Dep.Informatics, New University of Lisbon, Quinta da Torre 85 MONTE DA CAPARICA, PORTUGAL nb@di.fct.unl.pt Victor Jose Lobo Fernando Moura-Pires li. New University of Dep.Informatics, New University Lisbon/Portuguese Naval of Lisbon, Quinta da Tort-e 85 Academy, Escola Naval, Alfeite, MONTE DA CAPARICA, 8 ALMADA, PORTUGAL PORTUGAL vsl@di.fct.unl.pt finp@di.fct.unl.pt Abstract A distributed version of Kohonen s Self Organizing Map (SOM) algorithm is presented, and it s implementation on a Parallel Virtual Machine (PVM) network running on Intel based PCs with Microsoft Windows 95 is described. Numerical results are given for d#?erent work loads. Keywords: SOM, distributed algorithms, parallel algorithms, neural networks 1 - Introduction Although neural networks are intrinsically parallel algorithms, they are not easily implemented on distributed architectures because the strong interactions between neurons impose a very high communication overhead (neural networks are also called connectionist models). One of the neural models that has beenimplemented with more success onto parallel architectures is Kohonen s SOM [ 1][][3], because it requires very little communication between neurons. However these implementations have traditionally used parallel machines that tend to be expensive and non-standard []. Over the last few years, a system called Parallel Virtual Machine (PVM) [4][6], has been developed that enables a programmer to use networked computers (running different operating systems such as UNIX and MS- Windows 95) in a manner very similar to a single UNIX machine, using common languages such as C. Thanks to PVM, existing computer networks, no matter how heterogeneous, can easily be programmed. Moreover, simple PCs running MS-Windows (which abound in most organizations), can be put to work during otherwise unproductive times, such as nights and weekends. In this paper we describe our experience in trying to use the University s PC computer laboratories to train SOM networks. Although we intend to use these maps to classify sound, we used random vectors in these --I experiments. - Distributed SOM algorithm A thorough discussion of Kohonen s algorithm can be found in [5], but the core training algorithm may be described as follows: 1. For a given training pattern x: Calculate the distance of each neuron to the training pattern x (Calculation phase) 1. - Find the neuron with smaller distance, and call it the winner W (Voting phase) Change the network neurons with a function G, which depends on the learning rate a, the distance d to W (in the output plane), and the neighborhood function F. Due to the nature of the neighborhood function, only the neurons closer to W (in the output space) will be changed. (Update phase). Update the learning rate a and the neighborhood function F according to some rule 3. Repeat steps 1 and for the next training pattern, until some stopping criteria is reached. Many different distributed versions of Kohonen s SOM are possible, each being more adequate for a certain machine architecture. For implementing in PVM, we think the most adequate is the following: O l/98 $ IEEE 457
2 Given: Np processors (PVM hosts) A coordinator process C (within Np) Nt training patterns Xi N, neurons forming a SOM 1 - Assign the neurons. Assign the N, neurons to the processors, in such a way that each processor receives approximately the same number of neurons &/NJ., and that for any given area of the SOM the neurons are evenly distributed amongst the processors. - Set up the training set. Send all the N, training patterns to all the processors, together with the initial training parameters. 3 - Train the neurons. For each set, do the following: 3.1 Calculate the local (calculation phase) pattern xi in the training winner neuron in each processor 3. Send all local winner neuron coordinates and distances to the pattern x to the coordinator. At the coordinator, select the global winner, and send it s coordinates to all processors (voting phase) 3.3 Update the neurons at each processor, according to the update function F and update the network training parameters (update phase) 4 - Repeat step 3 until the stopping criteria is met 5 - Send all neurons back to the coordinator. A graphical representation of the algorithm messages involved is given in Figure 1. showing the The initial and final phases of the algorithm (represented in white in Figure 1) are executed only once, and thus have very little influence in the overall performance. Most of the time is spent in the main loop (represented in gray in Figure I), which iterates through the three main phases: calculation of the distances, voting for the global winner, and updating the neurons. The calculation phase of the algorithm (fmding the winner) is inherently parallel, and it s computational load can be spread evenly across the network if each processor has roughly the same number of neurons. The voting phase is the only one that requires communication ( and synchronization ) between processors, because the global winner must be known to all for the algorithm to proceed. If there was no coordinator, each processor would have to send information about it s local winner to all other processors. While PVM does support a broadcast mechanism, this would translate to a multicast at the data link level, thus originating N&VP-I) messages per iteration. With a coordinator, each process sends,only one message to the coordinator, and it in return sends only one message back, thus originating only (?V$) messages. Furthermore, the coordinator can piggyback additional information on the return message, that can be used to select the next training pattern, change the training parameters, etc. Coordinator Send start order Receive neurons Clients Receive neurons Figure 1 - Message exchange in distributed SOM During the update phase, each processor will have to calculate the distance between it s local neurons and the global winner (in the output plane), and then update only the neurons in p the winner s neighborhood. The computational load will be evenly distributed only if all processors have roughly the same number of neurons in this neighborhood. Thus it is very important to assign the neurons evenly during step 1 of the distributed algorithm. When the neighborhood radius is large, the computational load will easily be distributed. However, when the neighborhood radius is small, it is difficult to guarantee that all processors will have the same number of neurons, and thus some processors will have to wait for others. It must be pointed out that in this case, the number of neurons to update will be small, so the difference in processing time am-ongst the processors will be small. 3 - Experimental results During our experiments, we used networks of up to 1 PCs. Each was a Digital Venturis FX, with a Pentium running at loomhz, with 16Mb of RAM (55ISb cache). The computers where connected with a coaxial cable, 458
3 using lobase level protocol (1 Megabits per second), and TCP/IP as the level 3 protocol. The computers where running the MS-Windows 95 operating system, WPVM. [6] and Microsoft LAN Manager peer-to-peer network clients and servers. The computers had all screen-savers and anti-virus checkers disabled and where not running any other software during these tests. As the speedup obtained by using the distributed SOM depends critically on the amount of processing required before each synchronization, we used very large pattern vectors, with 14 features, and then varied the number of neurons on the map. We used square maps with 5x5, 1x1,x, and 4x4 neurons. Although square maps tend to slow convergence, we used them because they have a shorter boundary than rectangular maps (for the same number of neurons) and thus are less affected by discontinuities of the neighborhood function on those boundaries. This discontinuity effect would also affect the smaller maps more then the large ones if we used the same initial radius in all tests. To avoid this, we used an initial neighborhood function radius equal to each map s side. So as to make the radius decrease smoothly, we force it to decrease only 1 unit each time the whole training set patterns are presented. The map with 5x5 neurons will thus have only 5xh$ iterations of the patterns, while the one with 4x4 will have 4OxN,. So as not to make the simulation too long for larger maps, we use fewer training patterns for these maps. In the end, each map requires exactly 4 times more calculations then the previous one. Figure - Diagonal neuron distribution The number of patterns used for training does not influence the performance or speedup of the algorithm, so we use only 1 patterns for the 5x5 map, 6 for the 1x1, 3 for the x and 15 for the 4x4. While this number of patterns (and iterations) would be far too small for a useful classification or clustering, it is sufficient to prove that the system works reliably. The number of training patterns in the smaller maps has to be greater than for the larger maps, because otherwise the time intervals would be too small to be measured reliably. During these simulations we distributed the neurons amongst the processors in such a way that each processor has one or more diagonal lines of neurons. This distribution, although not ideal, provides a reasonable equilibrium of neurons by processors for rectangular neighborhoods (see Figure - every uncut neighborhood has the same neuron load per processor at each radius). The execution times where measured within the programs (with a call to a system timer), so as to measure only the time spent on the iterations (step 3 of the distributed algorithm, and shaded area in Figure 1). The numerical results are presented in Table 1 and in Figure 3, we can see a graph of the absolute execution times, while in Figure 4 we can see the relative times, that is, the speedup. tn g 5 Q i 3, No of machines Figure 3 - Absolute execution times ~(VdCOON \ \ No of machines + 5x5 -..-@b- 1x1 - Ref. Figure 4 - Relative execution time (l=time on a single machine) 459
4 Size of map No of PCs 5x5 1x1 x 4x4 r L Conclusions t Table 1 - Execution times (in seconds) The results presented confirm the claim that the SOM can be efficiently distributed on an ordinary computer network. However, depending on the work load, we may obtain overwhelming gains (as in the 4x4 map), moderate but consistent gains (as in the x map), or even high losses (as in the 5x5 map). There are a couple of extremely high running times corresponding to the 4x4 map running in a single machine or in a group of two. After that there is a sudden break and then the running times decrease smoothly. This initial peak is due to the machine configuration we are working on, namely because we have 16 Mb of RAM and a 4x4 map takes up to 13Mb RAM, forcing the operating system to use the disk as swap-space. We used machines with this configuration because we needed to have a reasonable pool of highly similar computers to achieve fair comparisons, and these where the ones available. Nevertheless we consider this as an advantage instead of a shortcoming, since these are the most common machines around any office or university lab, and allowed us to expose another very important fact when using distributed processing - the efficient use of each machine s memory. Using our distribution model, you can take advantage of each machine s local memory along with the corresponding processing power, effectively avoiding a RAM/HardDisk swapping situation which terribly slows down the SOM processing. In the more general case, the total execution time will decrease smoothly (except sometimes for the transition from 1 to machines), and then start to increase slowly. It s not reasonable to expect an unlimited gains as you throw more and more machines into the pool because of the increase in network load. When distributing a process, there will be a minimum overload on the total running time, due to the network. In our tests this can be seen by observing that the highest jumps upwards happen when there is a switch from a single machine to two machines. The distribution is profitable only when there is a significant workload to be distributed, thereby overcoming this minimum network * overload. After that initial step, all other machine extensions walk along an almost smooth curve, reflecting the converging equilibrium between each machine s designated workload and the network overload due to an increasing number of traded messages. The overload due to the network will increase linearly with the number of machines (each additional machine will be responsible for new messages), until the network starts to saturate due to collisions and/or sending queues. The workload per machine, on the other hand, will decrease hyperbolically, so at a certain point a minimum execution time will be reached, and adding a new machine will not improve the overall performance. We intend to concentrate our future work in three main areas. One of them is finding a better algorithm for distributing the neurons amongst processors. The ideal distribution will be a function of the radius of the neighborhood function and the number of processors, and genetic algorithms have been proposed to fmd good distributions. Another interesting problem is trying to fme tune the measurements of time spent calculating, transmitting, and waiting, during the iterations. On a PC running MS- Windows, this is rather tricky due to the lack of suitable tools. Finally, we intent to test and improve the algorithm for heterogeneous networks. This poses interesting problems of load balancing, and even more so if we attempt to do dynamic balancing during operation. Acknowledgments We would like to thank the Department de Engenharia Informatica da Universidade de Coimbra and the Escola Superior de Tecnologia e Gest%o do Instituto Politecnico de Viana do Castelo the for the development and availability of WPVM., the PVM implementation used in this work, and also the Prodep, FLAD and INVOTAN for the financial support to Nuno Bandeira (Prodep) and Victor Lobo (FLAD and INVOTAN). 46
5 References [I] Kohonen,T, et al.; Bibliography on Self-Organizing Map (SOM) and Learning Vector Quantization (LVQ), Helsinki 1997, available at ftp://cochlea.hut.fi [] Przytula, K.W, Prasanna, V.K.; Parallel Digital Implementations of Neural Networks, Prentice-Hall, [3] Ultsch, A, Siemon, H.P; Kohonen Networks on Transputers: Implementation and Animation, Proceedings of the International Neural Network Conference (INNC), July 199. [4] Geist, A., et al.; PVM: Parallel Virtual Machine - A User s guide and Tutorial for Networked Parallel Computing, The MIT Press, 1994 [5] Kohonen, Tuevo; Self-Organizing Maps, Springer- Verlag, 1995 [6] Alves, A, et al.; WPVM Manual, supplied with WPVM. package 461
Load Balancing on a Non-dedicated Heterogeneous Network of Workstations
Load Balancing on a Non-dedicated Heterogeneous Network of Workstations Dr. Maurice Eggen Nathan Franklin Department of Computer Science Trinity University San Antonio, Texas 78212 Dr. Roger Eggen Department
More informationLoad balancing in a heterogeneous computer system by self-organizing Kohonen network
Bull. Nov. Comp. Center, Comp. Science, 25 (2006), 69 74 c 2006 NCC Publisher Load balancing in a heterogeneous computer system by self-organizing Kohonen network Mikhail S. Tarkov, Yakov S. Bezrukov Abstract.
More informationultra fast SOM using CUDA
ultra fast SOM using CUDA SOM (Self-Organizing Map) is one of the most popular artificial neural network algorithms in the unsupervised learning category. Sijo Mathew Preetha Joy Sibi Rajendra Manoj A
More informationRodrigo Fernandes de Mello, Evgueni Dodonov, José Augusto Andrade Filho
Middleware for High Performance Computing Rodrigo Fernandes de Mello, Evgueni Dodonov, José Augusto Andrade Filho University of São Paulo São Carlos, Brazil {mello, eugeni, augustoa}@icmc.usp.br Outline
More informationChapter 12: Multiprocessor Architectures. Lesson 01: Performance characteristics of Multiprocessor Architectures and Speedup
Chapter 12: Multiprocessor Architectures Lesson 01: Performance characteristics of Multiprocessor Architectures and Speedup Objective Be familiar with basic multiprocessor architectures and be able to
More informationCS Master Level Courses and Areas COURSE DESCRIPTIONS. CSCI 521 Real-Time Systems. CSCI 522 High Performance Computing
CS Master Level Courses and Areas The graduate courses offered may change over time, in response to new developments in computer science and the interests of faculty and students; the list of graduate
More informationOverlapping Data Transfer With Application Execution on Clusters
Overlapping Data Transfer With Application Execution on Clusters Karen L. Reid and Michael Stumm reid@cs.toronto.edu stumm@eecg.toronto.edu Department of Computer Science Department of Electrical and Computer
More informationVisualization of Breast Cancer Data by SOM Component Planes
International Journal of Science and Technology Volume 3 No. 2, February, 2014 Visualization of Breast Cancer Data by SOM Component Planes P.Venkatesan. 1, M.Mullai 2 1 Department of Statistics,NIRT(Indian
More informationWell, now you have a Language Lab! www.easylab.es
Do you have a computer lab? Well, now you have a Language Lab! What exactly is easylab? easylab can transform any ordinary computer laboratory into a cuttingedge Language Laboratory. easylab is a system
More informationCapacity Estimation for Linux Workloads
Capacity Estimation for Linux Workloads Session L985 David Boyes Sine Nomine Associates 1 Agenda General Capacity Planning Issues Virtual Machine History and Value Unique Capacity Issues in Virtual Machines
More informationLocal Area Networks: Software
School of Business Eastern Illinois University Local Area Networks: Software (Week 8, Thursday 3/1/2007) Abdou Illia, Spring 2007 Learning Objectives 2 Identify main functions of operating systems Describe
More informationwhite paper Capacity and Scaling of Microsoft Terminal Server on the Unisys ES7000/600 Unisys Systems & Technology Modeling and Measurement
white paper Capacity and Scaling of Microsoft Terminal Server on the Unisys ES7000/600 Unisys Systems & Technology Modeling and Measurement 2 This technical white paper has been written for IT professionals
More informationDistributed Systems LEEC (2005/06 2º Sem.)
Distributed Systems LEEC (2005/06 2º Sem.) Introduction João Paulo Carvalho Universidade Técnica de Lisboa / Instituto Superior Técnico Outline Definition of a Distributed System Goals Connecting Users
More informationReliable Systolic Computing through Redundancy
Reliable Systolic Computing through Redundancy Kunio Okuda 1, Siang Wun Song 1, and Marcos Tatsuo Yamamoto 1 Universidade de São Paulo, Brazil, {kunio,song,mty}@ime.usp.br, http://www.ime.usp.br/ song/
More informationA Simultaneous Solution for General Linear Equations on a Ring or Hierarchical Cluster
Acta Technica Jaurinensis Vol. 3. No. 1. 010 A Simultaneous Solution for General Linear Equations on a Ring or Hierarchical Cluster G. Molnárka, N. Varjasi Széchenyi István University Győr, Hungary, H-906
More informationParallel Ray Tracing using MPI: A Dynamic Load-balancing Approach
Parallel Ray Tracing using MPI: A Dynamic Load-balancing Approach S. M. Ashraful Kadir 1 and Tazrian Khan 2 1 Scientific Computing, Royal Institute of Technology (KTH), Stockholm, Sweden smakadir@csc.kth.se,
More informationDistributed Dynamic Load Balancing for Iterative-Stencil Applications
Distributed Dynamic Load Balancing for Iterative-Stencil Applications G. Dethier 1, P. Marchot 2 and P.A. de Marneffe 1 1 EECS Department, University of Liege, Belgium 2 Chemical Engineering Department,
More informationModels of Cortical Maps II
CN510: Principles and Methods of Cognitive and Neural Modeling Models of Cortical Maps II Lecture 19 Instructor: Anatoli Gorchetchnikov dy dt The Network of Grossberg (1976) Ay B y f (
More informationWeb Server (Step 1) Processes request and sends query to SQL server via ADO/OLEDB. Web Server (Step 2) Creates HTML page dynamically from record set
Dawn CF Performance Considerations Dawn CF key processes Request (http) Web Server (Step 1) Processes request and sends query to SQL server via ADO/OLEDB. Query (SQL) SQL Server Queries Database & returns
More information6.2.8 Neural networks for data mining
6.2.8 Neural networks for data mining Walter Kosters 1 In many application areas neural networks are known to be valuable tools. This also holds for data mining. In this chapter we discuss the use of neural
More informationGrid e-services for Multi-Layer SOM Neural Network Simulation
Grid e-services for Multi-Layer SOM Neural Network Simulation,, Rui Silva Faculdade de Engenharia 4760-108 V. N. Famalicão, Portugal {rml,rsilva}@fam.ulusiada.pt 2007 Outline Overview Multi-Layer SOM Background
More informationpc resource monitoring and performance advisor
pc resource monitoring and performance advisor application note www.hp.com/go/desktops Overview HP Toptools is a modular web-based device management tool that provides dynamic information about HP hardware
More informationA Performance Study of Load Balancing Strategies for Approximate String Matching on an MPI Heterogeneous System Environment
A Performance Study of Load Balancing Strategies for Approximate String Matching on an MPI Heterogeneous System Environment Panagiotis D. Michailidis and Konstantinos G. Margaritis Parallel and Distributed
More informationEnhance student s learning with an aid of simulation software to understand Computer Networking Undergraduate courses.
Enhance student s learning with an aid of simulation software to understand Computer Networking Undergraduate courses. Veeramani Shanmugam s.veeramani@curtin.edu.my Lenin Gopal lenin@curtin.edu.my Zeya
More informationProcessor Capacity Reserves: An Abstraction for Managing Processor Usage
Processor Capacity Reserves: An Abstraction for Managing Processor Usage Clifford W. Mercer, Stefan Savage, and Hideyuki Tokuda School of Computer Science Carnegie Mellon University Pittsburgh, Pennsylvania
More informationWindows Server Performance Monitoring
Spot server problems before they are noticed The system s really slow today! How often have you heard that? Finding the solution isn t so easy. The obvious questions to ask are why is it running slowly
More informationAttunity RepliWeb Event Driven Jobs
Attunity RepliWeb Event Driven Jobs Software Version 5.2 June 25, 2012 RepliWeb, Inc., 6441 Lyons Road, Coconut Creek, FL 33073 Tel: (954) 946-2274, Fax: (954) 337-6424 E-mail: info@repliweb.com, Support:
More informationGPU File System Encryption Kartik Kulkarni and Eugene Linkov
GPU File System Encryption Kartik Kulkarni and Eugene Linkov 5/10/2012 SUMMARY. We implemented a file system that encrypts and decrypts files. The implementation uses the AES algorithm computed through
More informationBuilding an Inexpensive Parallel Computer
Res. Lett. Inf. Math. Sci., (2000) 1, 113-118 Available online at http://www.massey.ac.nz/~wwiims/rlims/ Building an Inexpensive Parallel Computer Lutz Grosz and Andre Barczak I.I.M.S., Massey University
More informationWhite Paper Perceived Performance Tuning a system for what really matters
TMurgent Technologies White Paper Perceived Performance Tuning a system for what really matters September 18, 2003 White Paper: Perceived Performance 1/7 TMurgent Technologies Introduction The purpose
More informationAN AUTOMATED DATA THROUGHPUT MEASUREMENT OVER A 10/100MBPS LAN
AN AUTOMATED DATA THROUGHPUT MEASUREMENT OVER A 10/100MBPS LAN John B. Oladosu* & Abigail B. Adetunji Computer Science & Engineering Department, Ladoke Akintola University of Technology, Ogbomoso, Nigeria
More informationDesign Issues in a Bare PC Web Server
Design Issues in a Bare PC Web Server Long He, Ramesh K. Karne, Alexander L. Wijesinha, Sandeep Girumala, and Gholam H. Khaksari Department of Computer & Information Sciences, Towson University, 78 York
More informationAgenda. Distributed System Structures. Why Distributed Systems? Motivation
Agenda Distributed System Structures CSCI 444/544 Operating Systems Fall 2008 Motivation Network structure Fundamental network services Sockets and ports Client/server model Remote Procedure Call (RPC)
More informationSla Aware Load Balancing Algorithm Using Join-Idle Queue for Virtual Machines in Cloud Computing
Sla Aware Load Balancing Using Join-Idle Queue for Virtual Machines in Cloud Computing Mehak Choudhary M.Tech Student [CSE], Dept. of CSE, SKIET, Kurukshetra University, Haryana, India ABSTRACT: Cloud
More informationLOAD BALANCING STRATEGY BASED ON CLOUD PARTITIONING CONCEPT
Journal homepage: www.mjret.in ISSN:2348-6953 LOAD BALANCING STRATEGY BASED ON CLOUD PARTITIONING CONCEPT Ms. Shilpa D.More 1, Prof. Arti Mohanpurkar 2 1,2 Department of computer Engineering DYPSOET, Pune,India
More informationA user friendly toolbox for exploratory data analysis of underwater sound
061215-064 1 A user friendly toolbox for exploratory data analysis of underwater sound Fernando J. Pires 1 and Victor Lobo 2, Member, IEEE Abstract The underwater acoustics research group at the Portuguese
More informationMuse Server Sizing. 18 June 2012. Document Version 0.0.1.9 Muse 2.7.0.0
Muse Server Sizing 18 June 2012 Document Version 0.0.1.9 Muse 2.7.0.0 Notice No part of this publication may be reproduced stored in a retrieval system, or transmitted, in any form or by any means, without
More informationParallelism and Cloud Computing
Parallelism and Cloud Computing Kai Shen Parallel Computing Parallel computing: Process sub tasks simultaneously so that work can be completed faster. For instances: divide the work of matrix multiplication
More informationElemental functions: Writing data-parallel code in C/C++ using Intel Cilk Plus
Elemental functions: Writing data-parallel code in C/C++ using Intel Cilk Plus A simple C/C++ language extension construct for data parallel operations Robert Geva robert.geva@intel.com Introduction Intel
More informationParallel & Distributed Optimization. Based on Mark Schmidt s slides
Parallel & Distributed Optimization Based on Mark Schmidt s slides Motivation behind using parallel & Distributed optimization Performance Computational throughput have increased exponentially in linear
More informationComputer Network. Interconnected collection of autonomous computers that are able to exchange information
Introduction Computer Network. Interconnected collection of autonomous computers that are able to exchange information No master/slave relationship between the computers in the network Data Communications.
More informationA Review of Customized Dynamic Load Balancing for a Network of Workstations
A Review of Customized Dynamic Load Balancing for a Network of Workstations Taken from work done by: Mohammed Javeed Zaki, Wei Li, Srinivasan Parthasarathy Computer Science Department, University of Rochester
More informationEfficient Load Balancing using VM Migration by QEMU-KVM
International Journal of Computer Science and Telecommunications [Volume 5, Issue 8, August 2014] 49 ISSN 2047-3338 Efficient Load Balancing using VM Migration by QEMU-KVM Sharang Telkikar 1, Shreyas Talele
More informationKeywords: Dynamic Load Balancing, Process Migration, Load Indices, Threshold Level, Response Time, Process Age.
Volume 3, Issue 10, October 2013 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Load Measurement
More informationHow To Test For Performance And Scalability On A Server With A Multi-Core Computer (For A Large Server)
Scalability Results Select the right hardware configuration for your organization to optimize performance Table of Contents Introduction... 1 Scalability... 2 Definition... 2 CPU and Memory Usage... 2
More informationDYNAMIC LOAD BALANCING APPLICATIONS ON A HETEROGENEOUS UNIX/NT CLUSTER
European Congress on Computational Methods in Applied Sciences and Engineering ECCOMAS 2000 Barcelona, 11-14 September 2000 ECCOMAS DYNAMIC LOAD BALANCING APPLICATIONS ON A HETEROGENEOUS UNIX/NT CLUSTER
More informationUnderstanding the Benefits of IBM SPSS Statistics Server
IBM SPSS Statistics Server Understanding the Benefits of IBM SPSS Statistics Server Contents: 1 Introduction 2 Performance 101: Understanding the drivers of better performance 3 Why performance is faster
More informationCHAPTER 1 INTRODUCTION
1 CHAPTER 1 INTRODUCTION 1.1 MOTIVATION OF RESEARCH Multicore processors have two or more execution cores (processors) implemented on a single chip having their own set of execution and architectural recourses.
More informationLCMON Network Traffic Analysis
LCMON Network Traffic Analysis Adam Black Centre for Advanced Internet Architectures, Technical Report 79A Swinburne University of Technology Melbourne, Australia adamblack@swin.edu.au Abstract The Swinburne
More informationControl 2004, University of Bath, UK, September 2004
Control, University of Bath, UK, September ID- IMPACT OF DEPENDENCY AND LOAD BALANCING IN MULTITHREADING REAL-TIME CONTROL ALGORITHMS M A Hossain and M O Tokhi Department of Computing, The University of
More informationNeural network software tool development: exploring programming language options
INEB- PSI Technical Report 2006-1 Neural network software tool development: exploring programming language options Alexandra Oliveira aao@fe.up.pt Supervisor: Professor Joaquim Marques de Sá June 2006
More informationAn Oracle White Paper November 2010. Oracle Real Application Clusters One Node: The Always On Single-Instance Database
An Oracle White Paper November 2010 Oracle Real Application Clusters One Node: The Always On Single-Instance Database Executive Summary... 1 Oracle Real Application Clusters One Node Overview... 1 Always
More informationBootstrapping Big Data
Bootstrapping Big Data Ariel Kleiner Ameet Talwalkar Purnamrita Sarkar Michael I. Jordan Computer Science Division University of California, Berkeley {akleiner, ameet, psarkar, jordan}@eecs.berkeley.edu
More informationEFFICIENT DATA PRE-PROCESSING FOR DATA MINING
EFFICIENT DATA PRE-PROCESSING FOR DATA MINING USING NEURAL NETWORKS JothiKumar.R 1, Sivabalan.R.V 2 1 Research scholar, Noorul Islam University, Nagercoil, India Assistant Professor, Adhiparasakthi College
More informationSERVER CLUSTERING TECHNOLOGY & CONCEPT
SERVER CLUSTERING TECHNOLOGY & CONCEPT M00383937, Computer Network, Middlesex University, E mail: vaibhav.mathur2007@gmail.com Abstract Server Cluster is one of the clustering technologies; it is use for
More informationSelf Organizing Maps: Fundamentals
Self Organizing Maps: Fundamentals Introduction to Neural Networks : Lecture 16 John A. Bullinaria, 2004 1. What is a Self Organizing Map? 2. Topographic Maps 3. Setting up a Self Organizing Map 4. Kohonen
More informationSCHEDULING IN CLOUD COMPUTING
SCHEDULING IN CLOUD COMPUTING Lipsa Tripathy, Rasmi Ranjan Patra CSA,CPGS,OUAT,Bhubaneswar,Odisha Abstract Cloud computing is an emerging technology. It process huge amount of data so scheduling mechanism
More informationEfficient DNS based Load Balancing for Bursty Web Application Traffic
ISSN Volume 1, No.1, September October 2012 International Journal of Science the and Internet. Applied However, Information this trend leads Technology to sudden burst of Available Online at http://warse.org/pdfs/ijmcis01112012.pdf
More informationMANAGING QUEUE STABILITY USING ART2 IN ACTIVE QUEUE MANAGEMENT FOR CONGESTION CONTROL
MANAGING QUEUE STABILITY USING ART2 IN ACTIVE QUEUE MANAGEMENT FOR CONGESTION CONTROL G. Maria Priscilla 1 and C. P. Sumathi 2 1 S.N.R. Sons College (Autonomous), Coimbatore, India 2 SDNB Vaishnav College
More informationUSING SELF-ORGANISING MAPS FOR ANOMALOUS BEHAVIOUR DETECTION IN A COMPUTER FORENSIC INVESTIGATION
USING SELF-ORGANISING MAPS FOR ANOMALOUS BEHAVIOUR DETECTION IN A COMPUTER FORENSIC INVESTIGATION B.K.L. Fei, J.H.P. Eloff, M.S. Olivier, H.M. Tillwick and H.S. Venter Information and Computer Security
More informationXTM Web 2.0 Enterprise Architecture Hardware Implementation Guidelines. A.Zydroń 18 April 2009. Page 1 of 12
XTM Web 2.0 Enterprise Architecture Hardware Implementation Guidelines A.Zydroń 18 April 2009 Page 1 of 12 1. Introduction...3 2. XTM Database...4 3. JVM and Tomcat considerations...5 4. XTM Engine...5
More informationThe software beyond the climatic. Environment simulation
Spirale 2 The software beyond the climatic Environment simulation Spirale 2... Take it easy! Spirale 2 is a new software, based on a reliable system (Windows NT) and if its performances are surprising,
More informationDynamic load balancing of parallel cellular automata
Dynamic load balancing of parallel cellular automata Marc Mazzariol, Benoit A. Gennart, Roger D. Hersch Ecole Polytechnique Fédérale de Lausanne, EPFL * ABSTRACT We are interested in running in parallel
More informationOperating Systems for Parallel Processing Assistent Lecturer Alecu Felician Economic Informatics Department Academy of Economic Studies Bucharest
Operating Systems for Parallel Processing Assistent Lecturer Alecu Felician Economic Informatics Department Academy of Economic Studies Bucharest 1. Introduction Few years ago, parallel computers could
More informationA Comparison of Distributed Systems: ChorusOS and Amoeba
A Comparison of Distributed Systems: ChorusOS and Amoeba Angelo Bertolli Prepared for MSIT 610 on October 27, 2004 University of Maryland University College Adelphi, Maryland United States of America Abstract.
More informationLoad Balance Strategies for DEVS Approximated Parallel and Distributed Discrete-Event Simulations
Load Balance Strategies for DEVS Approximated Parallel and Distributed Discrete-Event Simulations Alonso Inostrosa-Psijas, Roberto Solar, Verónica Gil-Costa and Mauricio Marín Universidad de Santiago,
More informationRemoving Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering
Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays Red Hat Performance Engineering Version 1.0 August 2013 1801 Varsity Drive Raleigh NC
More informationSimplest Scalable Architecture
Simplest Scalable Architecture NOW Network Of Workstations Many types of Clusters (form HP s Dr. Bruce J. Walker) High Performance Clusters Beowulf; 1000 nodes; parallel programs; MPI Load-leveling Clusters
More informationIntroduction 1 Performance on Hosted Server 1. Benchmarks 2. System Requirements 7 Load Balancing 7
Introduction 1 Performance on Hosted Server 1 Figure 1: Real World Performance 1 Benchmarks 2 System configuration used for benchmarks 2 Figure 2a: New tickets per minute on E5440 processors 3 Figure 2b:
More informationASSEMBLY PROGRAMMING ON A VIRTUAL COMPUTER
ASSEMBLY PROGRAMMING ON A VIRTUAL COMPUTER Pierre A. von Kaenel Mathematics and Computer Science Department Skidmore College Saratoga Springs, NY 12866 (518) 580-5292 pvonk@skidmore.edu ABSTRACT This paper
More informationReady Time Observations
VMWARE PERFORMANCE STUDY VMware ESX Server 3 Ready Time Observations VMware ESX Server is a thin software layer designed to multiplex hardware resources efficiently among virtual machines running unmodified
More informationTask Scheduling in Hadoop
Task Scheduling in Hadoop Sagar Mamdapure Munira Ginwala Neha Papat SAE,Kondhwa SAE,Kondhwa SAE,Kondhwa Abstract Hadoop is widely used for storing large datasets and processing them efficiently under distributed
More informationVDI FIT and VDI UX: Composite Metrics Track Good, Fair, Poor Desktop Performance
VDI FIT and VDI UX: Composite Metrics Track Good, Fair, Poor Desktop Performance Key indicators and classification capabilities in Stratusphere FIT and Stratusphere UX Whitepaper INTRODUCTION This whitepaper
More informationScaling from 1 PC to a super computer using Mascot
Scaling from 1 PC to a super computer using Mascot 1 Mascot - options for high throughput When is more processing power required? Options: Multiple Mascot installations Using a large Unix server Clusters
More informationUnderstanding the Performance of an X550 11-User Environment
Understanding the Performance of an X550 11-User Environment Overview NComputing's desktop virtualization technology enables significantly lower computing costs by letting multiple users share a single
More informationMicrosoft Windows Server 2003 with Internet Information Services (IIS) 6.0 vs. Linux Competitive Web Server Performance Comparison
April 23 11 Aviation Parkway, Suite 4 Morrisville, NC 2756 919-38-28 Fax 919-38-2899 32 B Lakeside Drive Foster City, CA 9444 65-513-8 Fax 65-513-899 www.veritest.com info@veritest.com Microsoft Windows
More informationReal Time Network Server Monitoring using Smartphone with Dynamic Load Balancing
www.ijcsi.org 227 Real Time Network Server Monitoring using Smartphone with Dynamic Load Balancing Dhuha Basheer Abdullah 1, Zeena Abdulgafar Thanoon 2, 1 Computer Science Department, Mosul University,
More informationSIP Registration Stress Test
SIP Registration Stress Test Miroslav Voznak and Jan Rozhon Department of Telecommunications VSB Technical University of Ostrava 17. listopadu 15/2172, 708 33 Ostrava Poruba CZECH REPUBLIC miroslav.voznak@vsb.cz,
More informationChapter 4: Artificial Neural Networks
Chapter 4: Artificial Neural Networks CS 536: Machine Learning Littman (Wu, TA) Administration icml-03: instructional Conference on Machine Learning http://www.cs.rutgers.edu/~mlittman/courses/ml03/icml03/
More informationRecommended hardware system configurations for ANSYS users
Recommended hardware system configurations for ANSYS users The purpose of this document is to recommend system configurations that will deliver high performance for ANSYS users across the entire range
More informationPreserving Message Integrity in Dynamic Process Migration
Preserving Message Integrity in Dynamic Process Migration E. Heymann, F. Tinetti, E. Luque Universidad Autónoma de Barcelona Departamento de Informática 8193 - Bellaterra, Barcelona, Spain e-mail: e.heymann@cc.uab.es
More informationExperiments on the local load balancing algorithms; part 1
Experiments on the local load balancing algorithms; part 1 Ştefan Măruşter Institute e-austria Timisoara West University of Timişoara, Romania maruster@info.uvt.ro Abstract. In this paper the influence
More informationInfor Web UI Sizing and Deployment for a Thin Client Solution
Infor Web UI Sizing and Deployment for a Thin Client Solution Copyright 2012 Infor Important Notices The material contained in this publication (including any supplementary information) constitutes and
More informationAnalysis of an Artificial Hormone System (Extended abstract)
c 2013. This is the author s version of the work. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purpose or for creating
More informationVisualization of large data sets using MDS combined with LVQ.
Visualization of large data sets using MDS combined with LVQ. Antoine Naud and Włodzisław Duch Department of Informatics, Nicholas Copernicus University, Grudziądzka 5, 87-100 Toruń, Poland. www.phys.uni.torun.pl/kmk
More informationMAS 200 Supported Platform Matrix
The information in this document applies to MAS 200 Level 3.60 and higher. Best Software only supports the current MAS 200 level and one major level back. As of the Level 3.70 release, MAS 200 Levels 3.51
More informationPerformance analysis of a Linux based FTP server
Performance analysis of a Linux based FTP server A Thesis Submitted in Partial Fulfillment of the Requirements for the Degree of Master of Technology by Anand Srivastava to the Department of Computer Science
More informationParallel Computing for Option Pricing Based on the Backward Stochastic Differential Equation
Parallel Computing for Option Pricing Based on the Backward Stochastic Differential Equation Ying Peng, Bin Gong, Hui Liu, and Yanxin Zhang School of Computer Science and Technology, Shandong University,
More informationLoad Distribution on a Linux Cluster using Load Balancing
Load Distribution on a Linux Cluster using Load Balancing Aravind Elango M. Mohammed Safiq Undergraduate Students of Engg. Dept. of Computer Science and Engg. PSG College of Technology India Abstract:
More informationLVQ Plug-In Algorithm for SQL Server
LVQ Plug-In Algorithm for SQL Server Licínia Pedro Monteiro Instituto Superior Técnico licinia.monteiro@tagus.ist.utl.pt I. Executive Summary In this Resume we describe a new functionality implemented
More informationHigh Performance Cluster Support for NLB on Window
High Performance Cluster Support for NLB on Window [1]Arvind Rathi, [2] Kirti, [3] Neelam [1]M.Tech Student, Department of CSE, GITM, Gurgaon Haryana (India) arvindrathi88@gmail.com [2]Asst. Professor,
More informationChapter 18: Database System Architectures. Centralized Systems
Chapter 18: Database System Architectures! Centralized Systems! Client--Server Systems! Parallel Systems! Distributed Systems! Network Types 18.1 Centralized Systems! Run on a single computer system and
More informationIntroduction to Virtual Machines
Introduction to Virtual Machines Introduction Abstraction and interfaces Virtualization Computer system architecture Process virtual machines System virtual machines 1 Abstraction Mechanism to manage complexity
More informationWeek 1 out-of-class notes, discussions and sample problems
Week 1 out-of-class notes, discussions and sample problems Although we will primarily concentrate on RISC processors as found in some desktop/laptop computers, here we take a look at the varying types
More informationDell Compellent Storage Center SAN & VMware View 1,000 Desktop Reference Architecture. Dell Compellent Product Specialist Team
Dell Compellent Storage Center SAN & VMware View 1,000 Desktop Reference Architecture Dell Compellent Product Specialist Team THIS WHITE PAPER IS FOR INFORMATIONAL PURPOSES ONLY, AND MAY CONTAIN TYPOGRAPHICAL
More informationPerformance Modeling and Analysis of a Database Server with Write-Heavy Workload
Performance Modeling and Analysis of a Database Server with Write-Heavy Workload Manfred Dellkrantz, Maria Kihl 2, and Anders Robertsson Department of Automatic Control, Lund University 2 Department of
More informationMethodology for Emulating Self Organizing Maps for Visualization of Large Datasets
Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets Macario O. Cordel II and Arnulfo P. Azcarraga College of Computer Studies *Corresponding Author: macario.cordel@dlsu.edu.ph
More informationA Guide to Getting Started with Successful Load Testing
Ingenieurbüro David Fischer AG A Company of the Apica Group http://www.proxy-sniffer.com A Guide to Getting Started with Successful Load Testing English Edition 2007 All Rights Reserved Table of Contents
More informationLearning Vector Quantization: generalization ability and dynamics of competing prototypes
Learning Vector Quantization: generalization ability and dynamics of competing prototypes Aree Witoelar 1, Michael Biehl 1, and Barbara Hammer 2 1 University of Groningen, Mathematics and Computing Science
More informationIntel Cloud Builders Guide to Cloud Design and Deployment on Intel Platforms
Intel Cloud Builders Guide Intel Xeon Processor-based Servers RES Virtual Desktop Extender Intel Cloud Builders Guide to Cloud Design and Deployment on Intel Platforms Client Aware Cloud with RES Virtual
More information