A Theory of the Spatial Computational Domain



Similar documents
Dell High-Performance Computing Clusters and Reservoir Simulation Research at UT Austin.

Cluster Computing at HRI

Globus Striped GridFTP Framework and Server. Raj Kettimuthu, ANL and U. Chicago

High Performance. CAEA elearning Series. Jonathan G. Dudley, Ph.D. 06/09/ CAE Associates

PARALLEL & CLUSTER COMPUTING CS 6260 PROFESSOR: ELISE DE DONCKER BY: LINA HUSSEIN

Performance of the NAS Parallel Benchmarks on Grid Enabled Clusters

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

Parallel Visualization for GIS Applications

Storage Virtualization from clusters to grid

Lecture 1: the anatomy of a supercomputer

RevoScaleR Speed and Scalability

CMS Tier-3 cluster at NISER. Dr. Tania Moulik

Load Balancing on a Non-dedicated Heterogeneous Network of Workstations

lesson 1 An Overview of the Computer System

Improved LS-DYNA Performance on Sun Servers

Scalable and High Performance Computing for Big Data Analytics in Understanding the Human Dynamics in the Mobile Age

A Flexible Cluster Infrastructure for Systems Research and Software Development

Scaling Objectivity Database Performance with Panasas Scale-Out NAS Storage

High Performance Computing in CST STUDIO SUITE

Grids Computing and Collaboration

PACE Predictive Analytics Center of San Diego Supercomputer Center, UCSD. Natasha Balac, Ph.D.

MIKE by DHI 2014 e sviluppi futuri

Spatial Data Analysis

THE EXPAND PARALLEL FILE SYSTEM A FILE SYSTEM FOR CLUSTER AND GRID COMPUTING. José Daniel García Sánchez ARCOS Group University Carlos III of Madrid

HADOOP ON ORACLE ZFS STORAGE A TECHNICAL OVERVIEW

Shared Parallel File System

Flash Memory Arrays Enabling the Virtualized Data Center. July 2010

ultra fast SOM using CUDA

Collaborative & Integrated Network & Systems Management: Management Using Grid Technologies

Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software

The Lattice Project: A Multi-Model Grid Computing System. Center for Bioinformatics and Computational Biology University of Maryland

22S:295 Seminar in Applied Statistics High Performance Computing in Statistics

Elastic Cloud Computing in the Open Cirrus Testbed implemented via Eucalyptus

Intel Cloud Builders Guide to Cloud Design and Deployment on Intel Platforms

Benchmark Hadoop and Mars: MapReduce on cluster versus on GPU

P013 INTRODUCING A NEW GENERATION OF RESERVOIR SIMULATION SOFTWARE

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering

High Performance Computing. Course Notes HPC Fundamentals

Parallels Plesk Automation

Hadoop: Embracing future hardware

Parallels Cloud Storage

KNOWLEDGE GRID An Architecture for Distributed Knowledge Discovery

White Paper. Recording Server Virtualization

Client/Server Computing Distributed Processing, Client/Server, and Clusters

Enabling Technologies for Distributed Computing

Clusters: Mainstream Technology for CAE

Fast Multipole Method for particle interactions: an open source parallel library component

Stanford HPC Conference. Panasas Storage System Integration into a Cluster

Virtualization of ArcGIS Pro. An Esri White Paper December 2015

Cellular Computing on a Linux Cluster

Symmetric Multiprocessing

Configuration Maximums VMware Infrastructure 3

wu.cloud: Insights Gained from Operating a Private Cloud System

Cisco Prime Home 5.0 Minimum System Requirements (Standalone and High Availability)

OpenMP Programming on ScaleMP

Processing Large Amounts of Images on Hadoop with OpenCV

Integrated Grid Solutions. and Greenplum

Parallel Programming Survey

Fusionstor NAS Enterprise Server and Microsoft Windows Storage Server 2003 competitive performance comparison

HP reference configuration for entry-level SAS Grid Manager solutions

Distributed Systems LEEC (2005/06 2º Sem.)

A Study on the Scalability of Hybrid LS-DYNA on Multicore Architectures

SAP database backup and restore solutions for HP StorageWorks Enterprise Virtual Array using HP Data Protector 6.1 software

Parallel Simplification of Large Meshes on PC Clusters

USING SELF-ORGANIZING MAPS FOR INFORMATION VISUALIZATION AND KNOWLEDGE DISCOVERY IN COMPLEX GEOSPATIAL DATASETS

High Performance Computing

Supercomputing and Big Data: Where are the Real Boundaries and Opportunities for Synergy?

Parallel Computing with MATLAB

Computational infrastructure for NGS data analysis. José Carbonell Caballero Pablo Escobar

Cluster Scalability of ANSYS FLUENT 12 for a Large Aerodynamics Case on the Darwin Supercomputer

The Green Index: A Metric for Evaluating System-Wide Energy Efficiency in HPC Systems

FPGA-based Multithreading for In-Memory Hash Joins

Enabling Technologies for Distributed and Cloud Computing

A Hybrid Distributed Architecture for Indexing

Colgate-Palmolive selects SAP HANA to improve the speed of business analytics with IBM and SAP

Building a Top500-class Supercomputing Cluster at LNS-BUAP

How To Compare Amazon Ec2 To A Supercomputer For Scientific Applications

Obj: Sec 1.0, to describe the relationship between hardware and software HW: Read p.2 9. Do Now: Name 3 parts of the computer.

Parallels Server 4 Bare Metal

MEGWARE HPC Cluster am LRZ eine mehr als 12-jährige Zusammenarbeit. Prof. Dieter Kranzlmüller (LRZ)

SAS Business Analytics. Base SAS for SAS 9.2

Cluster Computing in a College of Criminal Justice

Panasas at the RCF. Fall 2005 Robert Petkus RHIC/USATLAS Computing Facility Brookhaven National Laboratory. Robert Petkus Panasas at the RCF

Comparing Free Virtualization Products

The team that wrote this redbook Comments welcome Introduction p. 1 Three phases p. 1 Netfinity Performance Lab p. 2 IBM Center for Microsoft

MOSIX: High performance Linux farm

Transcription:

A Theory of the Spatial Computational Domain Shaowen Wang 1 and Marc P. Armstrong 2 1 Academic Technologies Research Services and Department of Geography, The University of Iowa Iowa City, IA 52242 Tel: +1 319-335-6713 FAX: +1 319-335-5505 Email: shaowen-wang@uiowa.edu 2 Department of Geography and Program in Applied Mathematical and Computational Sciences, The University of Iowa Iowa City, IA 52242 Tel: +1 319-335-0153 FAX: +1 319-335-2725 Email: marc-armstrong@uiowa.edu Abstract Most parallel processing methods developed for geographic analyses bind the design of domain decomposition and task scheduling to specific parallel computer architectures. These methods are not adaptable to emerging distributed computing environments that are based on Grid computing and peer-to-peer technologies. This paper presents a theory to support the development of adaptable parallel processing methods for geographic analyses performed on heterogeneous parallel processing environments. This theory of the spatial computational domain represents the computational intensity of geographic data and analysis methods, and transforms it into a common framework based on transformation theories from earlier cartographic research. The application of the theory is illustrated using an inverse distance weighted interpolation method. We describe the underpinnings of this analysis, and then address the latent parallelism of a conventional sequential algorithm based on spatial computational domain theory. Through the application of this theory, the development of domain decomposition methods is decoupled from specific high performance computer architectures and task scheduling implementations, which makes the design of generic parallel processing geographic analysis solutions feasib le. 1. Existing Problems in Parallel Processing of Geographic Information The state-of-the-practice in parallel processing of geographic information binds the design of domain decomposition and task scheduling to specific conventional parallel computer architectures. This tight-coupling approach is problematic to software design for three reasons: 1.) Domain decomposition and task scheduling methods focus on the characteristics of spatial data and operations performed on them.

2.) Any change to spatial data or operations requires a corresponding change in the design of domain decompositio n and task scheduling methods. 3.) Domain decomposition and task scheduling strategies depend upon specific parallel architectures. These three problems are analogous to those observed in graphics programming before the advent of device-independent software. In this case, these problems hinder the development of portable parallel geographic analysis methods based on computational Grids (see endnote). To eliminate these problems, we introduce a new spatial computational domain theory and illustrate an application using the TeraGrid (TeraGrid, 2005). 2. Spatial Computational Domain Theory The spatial computational domain is formally defined to comprise several two-dimensional computational intensity surfaces. Given a spatial domain projected to a two-dimensional (x and y) Euclidian space, each two-dimensional computational surface can be represented as a vector c = (c ij ) in the space R N, where N = x c y c, and where x c is the number of cells in the x dimension, and y c is the number of cells in the y dimension of the computational surfaces. Each component of c corresponds to the computational intensity at cell (i,. The spatial computational domain is similar to an image obtained from an evaluation of the following function: f: I 2 = [0, 1] [0, 1]? R (1) at cells (i/x c, j/y c ) I 2, i = 1,, x c, j = 1,, y c. Thus, cij = f (i/xc, j/yc) (2) The value of cij represents the computational burden derived from a new computational transformation concept that is consistent with the role of transformations in other domains of GIScience (Tobler, 1979): Computing time: the time taken to perform the analysis within cell (i, ; Memory: the memory required to perform the analysis within cell (i, ; and I/O: data input/output and transfers to perform analyses within cell (i,. 3. Computational Transformation A computational transformation elucidates the computational intensity of a particular geographic analysis based on the characteristics of spatial data and operations. Two types of computational transformations are identified: 1.) Data-centric functions transform spatial data characteristics into memory or I/O surfaces. 2.) Operation-centric functions consider the spatial operations that are directly or indirectly related to spatial data characteristics, and transform the characteristics of spatial operations into a computing time surface. 3.1 An Illustrative Example An operation-centric transformation function illustrates the spatial computational domain for

Clarke s inverse distance weighted (IDW) interpolation algorithm (Clarke, 1995). This domain consists of a single computing time surface because both memory and I/O requirements are trivial even if a common desktop PC (e.g., 1G RAM and 2GHz Pentium processor) is used to execute the algorithm. The transformation function - IDW-of on the cell (i, is defined as follows: ne (3) j ) IDW-of? timeunit ( ns 1) dp + densitythreshold + where ne represents the number of empty grids that invoke a search for their k-nearest neighbors; ns represents the number of sampling points; dp represents the local density of sampling points; timeunit is a constant used to convert the value to a computing time unit (e.g., seconds); densitythreshold is a constant that represents the minimum point density in any given region of the domain considered, and prevents the denominator from becoming zero. The value of dp j ) is calculated using the following equation: dp = j ) ns ns + ne + + number neighbor k k= 1 number neighbor k= 1 ns ( ns k + ne ) where nsk represents the number of existing sampling points within the (i, neighbor cell indexed as k; and number represents the number of (i, neighbors being considered. neighbor Figure 1 shows a synthetic dataset with 2000 sampling points. The raster structure (2000 by 2000) is created in the initialization step of Clarke s algorithm. Figure 2 shows a 3-D plot of k (4)

Figure 1. A dataset of two thousand points, one cluster with a uniform random distribution

Figure 2. A 3-D plot of the spatial computational domain for Clarke s algorithm using the dataset in Figure 1; cell size is specified as 62.5 62.5, and thus the computiontal domain dimension is fixed at 32 32 given a 2000 2000 data domain dimension. the spatial computational domain derived from this dataset using the computational transformation function (Equation 3). Based on a decompositio n of the spatial computational domain using a quadtree method (Wang and Armstrong, 2003), a parallel interpolation of this dataset using the TeraGrid (Table 1) has a speedup of more than 20 times compared to a sequential algorithm executed on the fastest single processor of the TeraGrid. Table 2 shows the computing time taken by the parallel interpolation analysis in a set of computational experiments in which the spatial computational domain was decomposed to create different numbers of parallel jobs. These jobs are scheduled to available TeraGrid resources to achieve load balancing based on computational

intensity information derived from their spatial computational domain. This is a preliminary result; we are devising additional experiments to evaluate performance. Table 1. Hardware information about TeraGrid resources used Sites Compute nodes CPU Intranetwork ANL/ UC Dual-processor: 96 RAM memory: 4GB Dual-processor: 16 RAM memory: 4GB Intel Xeon 2.4GHz Intel Itanium 2, 1.3 GHz Myrinet 2000, Gigabit Ethernet, Fib er Channel Disk OS Local resource manager PBS Caltech Dual-processor: 52 RAM memory: 6GB 75 TB Parallel Linux 2.4.19 - Virtual File SMP System NCSA Dual-processor: 128 RAM memory: 12GB Dual-processor: 128 RAM memory: 4 GB 5TB Network File System 4TB Linux Network 2.4.21- File SystemSMP (shared memory multiprocessor) 4TB Linux Network 2.4.21- File SystemSMP Linux 2.4.21- SMP SDSC Dual-processor: 128 RAM memory: 4GB 1.6TB Network Linux 2.4.19 - File SystemSMP 4. Concluding Discussion Our spatial computational domain theory has solved several problems present in previous GIScience parallel processing research because it provides computational intensity representations that are required for efficient parallel processing of geographic analyses. At the same time, this theory has paved the way to establish generality in domain decomposition methods. The development of these functions is independent of any computer architecture, which also indicates the theory can be used to address problems with architectural incompatibilities. Our ongoing research focuses on the application of this theory to develop geo-middleware that supports other types of Grid-based geographic information analyses.

Table 2. Interpolation time in the TeraGrid environment TeraGrid sites Number of computing jobs submitted to each TeraGrid site ANL/UC 230 57 260 68 NCSA 138 44 219 71 SDSC 113 34 161 74 Caltech 62 25 106 38 ANL/UC 1 0 2 5 Computing Time (hh:mm:ss) 00:39:46 00:44:19 00:42:21 00:48:3 2 Note Grid computing has been the recent focus of a substantial amount of research and development activity. Though its basic concepts were stated earlier, the term Grid was coined in the late 90s by Foster and Kesselman (1999) to describe a set of resources distributed over wide-area networks that can support large-scale distributed applications. The analogy likens the Grid to the electrical power grid: access to computation and data should be as easy, pervasive, and as standardized as plugging an appliance into an outlet (Foster and Kesselman, 1999). In the foundational paper The Anatomy of the Grid, Foster et al. (2001) more formally define the Grid as a set of environments for coordinated resource sharing and problem solving in dynamic, multi-institutional, virtual organizations. This coordination is orchestrated using protocols and specialized software referred to as Grid middleware. In that paper, they go on to argue that the Grid is central not only to escience, but also to industry and commerce, where coordination of distributed resources both within and across organizations has become an important business activity. 5. References Clarke, K. C., 1995, Analytical and Computer Cartography, Second Edition (Englewood Cliffs, NJ: Prentice-Hall). Foster, I., Kesselman, C., and Tuecke, S. (2001) The anatomy of the Grid: enabling scalable virtual organizations. International Journal Supercomputer Applications, 15(3), available at: http://www.globus.org/research/papers.html. Foster, I., and Kesselman C., 1999, The Grid: Blue Print for a New Computing Infrastructure (San Francisco, CA: Morgan Kaufmann Publishers, Inc.). TeraGrid, 2005, http://www.teragrid.org/. Tobler, W. R., 1979, A transformational view of cartography. American Cartographer, 6(2), 101-106. Wang, S., and Armstrong, M. P., 2003, A quadtree approach to domain decomposition for spatial interpolation in Grid computing environments. Parallel Computing, 29(10), 1481-1504.